Today’s Research Highlights
AI-enhanced summaries of the latest research papers from arXiv.
Table of Contents
- cs.CL [Total: 133]
- cs.CV [Total: 365]
- cs.AI [Total: 106]
- cs.SD [Total: 12]
- cs.LG [Total: 305]
- cs.MA [Total: 10]
- cs.MM [Total: 6]
- eess.AS [Total: 11]
- eess.IV [Total: 17]
cs.CL
[1] Text Annotation via Inductive Coding: Comparing Human Experts to LLMs in Qualitative Data Analysis
Angelina Parfenova, Andreas Marfurt, Alexander Denzler, Juergen Pfeffer
Main category: cs.CL
TL;DR: LLMs show opposite performance pattern to humans in inductive coding - humans struggle with simple sentences but excel with complex ones, while LLMs do the reverse. LLMs sometimes align better with true labels but receive lower human evaluations.
Details
Motivation: To investigate automation of qualitative data analysis using LLMs for inductive coding (where labels emerge from data rather than using predefined labels), and compare LLM performance to human experts.Method: Evaluated six open-source LLMs compared to human experts on inductive coding tasks. Experts rated perceived difficulty of quotes they coded. Compared both human and LLM generated labels to golden standard test set, and examined systematic deviations.
Result: Found a dichotomy: human coders perform well with complex sentences but struggle with simpler ones, while LLMs show opposite trend. Human annotations sometimes differ from golden standard but are rated more favorably by other humans. Some LLMs demonstrate closer alignment with true labels but receive lower evaluations from experts.
Conclusion: LLMs and humans have complementary strengths in inductive coding - humans excel with complexity while LLMs handle simplicity better. Human evaluation bias exists where human-generated labels are preferred even when less accurate, suggesting need for careful evaluation metrics in automated qualitative analysis.
Abstract: This paper investigates the automation of qualitative data analysis, focusing on inductive coding using large language models (LLMs). Unlike traditional approaches that rely on deductive methods with predefined labels, this research investigates the inductive process where labels emerge from the data. The study evaluates the performance of six open-source LLMs compared to human experts. As part of the evaluation, experts rated the perceived difficulty of the quotes they coded. The results reveal a peculiar dichotomy: human coders consistently perform well when labeling complex sentences but struggle with simpler ones, while LLMs exhibit the opposite trend. Additionally, the study explores systematic deviations in both human and LLM generated labels by comparing them to the golden standard from the test set. While human annotations may sometimes differ from the golden standard, they are often rated more favorably by other humans. In contrast, some LLMs demonstrate closer alignment with the true labels but receive lower evaluations from experts.
[2] Emergent Convergence in Multi-Agent LLM Annotation
Angelina Parfenova, Alexander Denzler, Juergen Pfeffer
Main category: cs.CL
TL;DR: LLM agents coordinate effectively in collaborative coding tasks without explicit prompting, showing convergence, negotiation behaviors, and semantic compression over interactions.
Details
Motivation: LLMs are increasingly used in collaborative settings, but little is known about how they coordinate as black-box agents. The paper aims to understand emergent coordination strategies in multi-agent LLM interactions without explicit role assignments.Method: Simulated 7500 multi-agent, multi-round discussions in inductive coding tasks, generating 125k+ utterances. Introduced process-level metrics (code stability, semantic self-consistency, lexical confidence, sentiment, convergence) and analyzed evolving geometry of output embeddings to track coordination dynamics.
Result: LLM groups converge lexically and semantically, develop asymmetric influence patterns, exhibit negotiation-like behaviors without explicit prompting. Intrinsic dimensionality of output embeddings declines over rounds, suggesting semantic compression. Black-box interaction analysis reveals emergent coordination strategies.
Conclusion: Black-box interaction analysis can surface emergent coordination strategies in LLM agents, offering a scalable complement to internal probe-based interpretability methods. LLMs demonstrate sophisticated coordination capabilities even without explicit role assignments.
Abstract: Large language models (LLMs) are increasingly deployed in collaborative settings, yet little is known about how they coordinate when treated as black-box agents. We simulate 7500 multi-agent, multi-round discussions in an inductive coding task, generating over 125000 utterances that capture both final annotations and their interactional histories. We introduce process-level metrics: code stability, semantic self-consistency, and lexical confidence alongside sentiment and convergence measures, to track coordination dynamics. To probe deeper alignment signals, we analyze the evolving geometry of output embeddings, showing that intrinsic dimensionality declines over rounds, suggesting semantic compression. The results reveal that LLM groups converge lexically and semantically, develop asymmetric influence patterns, and exhibit negotiation-like behaviors despite the absence of explicit role prompting. This work demonstrates how black-box interaction analysis can surface emergent coordination strategies, offering a scalable complement to internal probe-based interpretability methods.
[3] Tree Matching Networks for Natural Language Inference: Parameter-Efficient Semantic Understanding via Dependency Parse Trees
Jason Lunder
Main category: cs.CL
TL;DR: Tree Matching Networks (TMN) using dependency parse trees outperform BERT on NLI tasks with much smaller size and faster training, but struggle on similarity tasks.
Details
Motivation: Transformer models like BERT achieve high NLI accuracy but require massive parameters and learn relationships from scratch. Explicit linguistic structures (dependency parse trees) could leverage prior knowledge and improve learning efficiency.Method: Adapt Graph Matching Networks (GMN) to operate on dependency parse trees, creating Tree Matching Networks (TMN). Compare TMN to BERT on SNLI entailment and SemEval similarity tasks. Propose multi-headed attention aggregation to address scalability limitations.
Result: TMN achieves significantly better results than BERT on SNLI task with reduced memory footprint and less training time. Both models struggle on SemEval task. Explicit structural representations outperform sequence-based models at comparable scales.
Conclusion: Dependency parse tree-based models (TMN) are more efficient and effective than BERT for NLI tasks, but current aggregation methods limit scalability. Multi-headed attention aggregation is proposed to address this limitation.
Abstract: In creating sentence embeddings for Natural Language Inference (NLI) tasks, using transformer-based models like BERT leads to high accuracy, but require hundreds of millions of parameters. These models take in sentences as a sequence of tokens, and learn to encode the meaning of the sequence into embeddings such that those embeddings can be used reliably for NLI tasks. Essentially, every word is considered against every other word in the sequence, and the transformer model is able to determine the relationships between them, entirely from scratch. However, a model that accepts explicit linguistic structures like dependency parse trees may be able to leverage prior encoded information about these relationships, without having to learn them from scratch, thus improving learning efficiency. To investigate this, we adapt Graph Matching Networks (GMN) to operate on dependency parse trees, creating Tree Matching Networks (TMN). We compare TMN to a BERT based model on the SNLI entailment task and on the SemEval similarity task. TMN is able to achieve significantly better results with a significantly reduced memory footprint and much less training time than the BERT based model on the SNLI task, while both models struggled to preform well on the SemEval. Explicit structural representations significantly outperform sequence-based models at comparable scales, but current aggregation methods limit scalability. We propose multi-headed attention aggregation to address this limitation.
[4] Towards Corpus-Grounded Agentic LLMs for Multilingual Grammatical Analysis
Matej Klemen, Tjaša Arčon, Luka Terčon, Marko Robnik-Šikonja, Kaja Dobrovoljc
Main category: cs.CL
TL;DR: LLM agents can automate corpus-based grammatical analysis by reasoning over annotated corpora, tested on UD corpora for 13 word-order features across 170+ languages.
Details
Motivation: Empirical grammar research requires substantial methodological effort despite being data-driven; there's a need to streamline systematic analysis of annotated corpora using modern AI approaches.Method: Agentic LLM framework integrating natural-language task interpretation, code generation, and data-driven reasoning applied to Universal Dependencies corpora for multilingual grammatical tasks inspired by WALS.
Result: Evaluated on 13 word-order features across 170+ languages using three dimensions: dominant-order accuracy, order-coverage completeness, and distributional fidelity, demonstrating feasibility of LLM reasoning with structured linguistic data.
Conclusion: Combining LLM reasoning with structured linguistic data offers a first step toward interpretable, scalable automation of corpus-based grammatical inquiry.
Abstract: Empirical grammar research has become increasingly data-driven, but the systematic analysis of annotated corpora still requires substantial methodological and technical effort. We explore how agentic large language models (LLMs) can streamline this process by reasoning over annotated corpora and producing interpretable, data-grounded answers to linguistic questions. We introduce an agentic framework for corpus-grounded grammatical analysis that integrates concepts such as natural-language task interpretation, code generation, and data-driven reasoning. As a proof of concept, we apply it to Universal Dependencies (UD) corpora, testing it on multilingual grammatical tasks inspired by the World Atlas of Language Structures (WALS). The evaluation spans 13 word-order features and over 170 languages, assessing system performance across three complementary dimensions - dominant-order accuracy, order-coverage completeness, and distributional fidelity - which reflect how well the system generalizes, identifies, and quantifies word-order variations. The results demonstrate the feasibility of combining LLM reasoning with structured linguistic data, offering a first step toward interpretable, scalable automation of corpus-based grammatical inquiry.
[5] Minimal-Edit Instruction Tuning for Low-Resource Indic GEC
Akhil Rajeev P
Main category: cs.CL
TL;DR: Instruction-tuned LLMs with classifier-informed prompts and conservative decoding achieve strong GEC results for Indic languages without data augmentation.
Details
Motivation: Indic languages face challenges in grammatical error correction due to limited supervision, diverse scripts, and rich morphology, creating a need for efficient, augmentation-free approaches.Method: Instruction-tune a 12B GEMMA 3 model with 4-bit precision and PEFT using Alpaca-style formatting. Use deterministic, constraint-aware decoding with lightweight normalizer for minimal edits. Employ classifier-informed prompts based on error taxonomy and distributions.
Result: Achieved 92.41 GLEU on Malayalam (6th overall) and 81.44 GLEU on Hindi (3rd overall) under official untuned evaluation, showing competitive performance.
Conclusion: Classifier-informed prompts, adapter-based instruction tuning, and deterministic decoding provide reproducible, computationally efficient alternative to augmentation pipelines for Indic GEC, motivating future work on stronger constraints and human evaluation.
Abstract: Grammatical error correction for Indic languages faces limited supervision, diverse scripts, and rich morphology. We propose an augmentation-free setup that uses instruction-tuned large language models and conservative decoding. A 12B GEMMA 3 model is instruction-tuned in bnb 4-bit precision with parameter-efficient fine-tuning (PEFT) and Alpaca-style formatting. Decoding follows a deterministic, constraint-aware procedure with a lightweight normaliser that encourages minimal, meaning-preserving edits. We operationalise inference, subsequent to instruction fine-tuning (IFT), via a fixed, language-specific prompt directly synthesised from a deterministic error classifier’s taxonomy, label distributions, and precedence ordering computed on the training corpus. Under the official untuned GLEU evaluation, the system scores 92.41 on Malayalam, sixth overall, and 81.44 on Hindi, third overall. These results indicate that classifier-informed prompt design, adapter-based instruction tuning, and deterministic decoding provide a reproducible and a computationally efficient alternative to augmentation-centred pipelines for Indic GEC. The approach also motivates future work on stronger morphosyntactic constraints and human-centred evaluation of conservative edits.
[6] MAC-SLU: Multi-Intent Automotive Cabin Spoken Language Understanding Benchmark
Yuezhang Peng, Chonghao Cai, Ziang Liu, Shuai Fan, Sheng Jiang, Hua Xu, Yuxin Liu, Qiguang Chen, Kele Xu, Yao Li, Sheng Wang, Libo Qin, Xie Chen
Main category: cs.CL
TL;DR: MAC-SLU is a new multi-intent automotive cabin spoken language understanding dataset that increases SLU difficulty with authentic complex data, used to benchmark LLMs and LALMs across various methods.
Details
Motivation: Existing SLU datasets lack sufficient diversity and complexity, and there's no unified benchmark for evaluating the latest LLMs and LALMs on SLU tasks.Method: Introduces MAC-SLU dataset with authentic multi-intent automotive cabin data, then benchmarks leading open-source LLMs and LALMs using in-context learning, supervised fine-tuning, and both end-to-end and pipeline paradigms.
Result: LLMs and LALMs can complete SLU tasks via in-context learning but performance lags significantly behind supervised fine-tuning. End-to-end LALMs show comparable performance to pipeline approaches while avoiding speech recognition error propagation.
Conclusion: MAC-SLU provides a challenging benchmark for SLU evaluation, demonstrating current limitations of LLMs/LALMs via in-context learning while showing promise of end-to-end approaches that avoid error propagation.
Abstract: Spoken Language Understanding (SLU), which aims to extract user semantics to execute downstream tasks, is a crucial component of task-oriented dialog systems. Existing SLU datasets generally lack sufficient diversity and complexity, and there is an absence of a unified benchmark for the latest Large Language Models (LLMs) and Large Audio Language Models (LALMs). This work introduces MAC-SLU, a novel Multi-Intent Automotive Cabin Spoken Language Understanding Dataset, which increases the difficulty of the SLU task by incorporating authentic and complex multi-intent data. Based on MAC-SLU, we conducted a comprehensive benchmark of leading open-source LLMs and LALMs, covering methods like in-context learning, supervised fine-tuning (SFT), and end-to-end (E2E) and pipeline paradigms. Our experiments show that while LLMs and LALMs have the potential to complete SLU tasks through in-context learning, their performance still lags significantly behind SFT. Meanwhile, E2E LALMs demonstrate performance comparable to pipeline approaches and effectively avoid error propagation from speech recognition. Code\footnote{https://github.com/Gatsby-web/MAC_SLU} and datasets\footnote{huggingface.co/datasets/Gatsby1984/MAC_SLU} are released publicly.
[7] OmniFusion: Simultaneous Multilingual Multimodal Translations via Modular Fusion
Sai Koneru, Matthias Huck, Jan Niehues
Main category: cs.CL
TL;DR: OmniFusion: An end-to-end multimodal translation system that fuses multimodal foundation models with translation LLMs to enable speech-to-text, speech-and-image-to-text, and text-and-image-to-text translation with reduced latency and improved quality.
Details
Motivation: Current open-source text-only translation LLMs require cascaded pipelines for speech translation, introducing latency issues in simultaneous translation and preventing exploitation of multimodal context like images for disambiguation. While multimodal foundation models have strong cross-modal capabilities, they lack the multilingual coverage and specialized translation performance of dedicated translation LLMs.Method: Proposes an end-to-end fusion approach connecting hidden states from multiple layers of a pretrained multimodal foundation model (Omni 2.5-7B) to a translation LLM (SeedX PPO-7B). Uses a novel fusion strategy that enables joint end-to-end training for multimodal translation tasks.
Result: OmniFusion effectively leverages both audio and visual inputs, achieves 1-second latency reduction in simultaneous speech translation compared to cascaded pipelines, and improves overall translation quality across speech-to-text, speech-and-image-to-text, and text-and-image-to-text translation tasks.
Conclusion: The proposed fusion approach successfully combines the strengths of multimodal foundation models and specialized translation LLMs, creating an effective end-to-end multimodal translation system that reduces latency and improves translation quality while enabling multimodal context exploitation.
Abstract: There has been significant progress in open-source text-only translation large language models (LLMs) with better language coverage and quality. However, these models can be only used in cascaded pipelines for speech translation (ST), performing automatic speech recognition first followed by translation. This introduces additional latency, which is particularly critical in simultaneous ST (SimulST), and prevents the model from exploiting multimodal context, such as images, which can aid disambiguation. Pretrained multimodal foundation models (MMFMs) already possess strong perception and reasoning capabilities across multiple modalities, but generally lack the multilingual coverage and specialized translation performance of dedicated translation LLMs. To build an effective multimodal translation system, we propose an end-to-end approach that fuses MMFMs with translation LLMs. We introduce a novel fusion strategy that connects hidden states from multiple layers of a pretrained MMFM to a translation LLM, enabling joint end-to-end training. The resulting model, OmniFusion, built on Omni 2.5-7B as the MMFM and SeedX PPO-7B as the translation LLM, can perform speech-to-text, speech-and-image-to-text, and text-and-image-to-text translation. Experiments demonstrate that OmniFusion effectively leverages both audio and visual inputs, achieves a 1-second latency reduction in SimulST compared to cascaded pipelines and also improves the overall translation quality\footnote{Code is available at https://github.com/saikoneru/OmniFusion}.
[8] MEGConformer: Conformer-Based MEG Decoder for Robust Speech and Phoneme Classification
Xabier de Zuazo, Ibon Saratxaga, Eva Navas
Main category: cs.CL
TL;DR: Conformer-based decoders for MEG tasks in LibriBrain 2025 competition, achieving 88.9% on Speech Detection and 65.8% on Phoneme Classification with task-specific adaptations.
Details
Motivation: To develop effective neural decoders for foundational MEG tasks (Speech Detection and Phoneme Classification) in the LibriBrain 2025 PNPL competition, surpassing baseline performance and achieving competitive rankings.Method: Adapted compact Conformer architecture to raw 306-channel MEG signals with lightweight convolutional projection layer and task-specific heads. Used MEG-oriented SpecAugment for Speech Detection, inverse-square-root class weighting and dynamic grouping loader for Phoneme Classification, plus instance-level normalization to handle distribution shifts.
Result: Achieved 88.9% on Speech Detection and 65.8% on Phoneme Classification on the competition leaderboard, surpassing baselines and ranking within top-10 in both tasks using official Standard track splits and F1-macro for model selection.
Conclusion: Conformer-based decoders with task-specific adaptations effectively handle MEG signal processing for speech-related tasks, with instance normalization proving critical for robustness against distribution shifts. The approach demonstrates competitive performance in the LibriBrain competition.
Abstract: We present Conformer-based decoders for the LibriBrain 2025 PNPL competition, targeting two foundational MEG tasks: Speech Detection and Phoneme Classification. Our approach adapts a compact Conformer to raw 306-channel MEG signals, with a lightweight convolutional projection layer and task-specific heads. For Speech Detection, a MEG-oriented SpecAugment provided a first exploration of MEG-specific augmentation. For Phoneme Classification, we used inverse-square-root class weighting and a dynamic grouping loader to handle 100-sample averaged examples. In addition, a simple instance-level normalization proved critical to mitigate distribution shifts on the holdout split. Using the official Standard track splits and F1-macro for model selection, our best systems achieved 88.9% (Speech) and 65.8% (Phoneme) on the leaderboard, surpassing the competition baselines and ranking within the top-10 in both tasks. For further implementation details, the technical documentation, source code, and checkpoints are available at https://github.com/neural2speech/libribrain-experiments.
[9] Lost without translation – Can transformer (language models) understand mood states?
Prakrithi Shivaprakash, Diptadhi Mukherjee, Lekhansh Shukla, Animesh Mukherjee, Prabhat Chand, Pratima Murthy
Main category: cs.CL
TL;DR: LLMs struggle to understand mood states directly from Indic languages but perform well when phrases are translated to English or Chinese, highlighting a language barrier in psychiatric applications.
Details
Motivation: To assess whether language models can accurately understand and represent mood states expressed in Indian languages, given that current LLMs are English-centric and different languages have unique idioms of distress.Method: Collected 247 unique phrases for four mood states across 11 Indic languages. Tested seven conditions comparing k-means clustering on direct embeddings of native/Romanised scripts vs. embeddings of phrases translated to English/Chinese. Used multilingual and Indic-specific models, measuring performance with composite scores based on multiple metrics.
Result: Direct embedding of Indic languages failed completely (Composite Score = 0.002). Translation-based approaches showed significant improvement, with best performance from human-translated English further translated to Chinese and embedded with Chinese model (Composite = 0.67). Specialized Indic models performed poorly.
Conclusion: Current models cannot meaningfully represent mood states directly from Indic languages, creating a barrier for psychiatric applications in India. While translation bridges this gap, reliance on proprietary models is unsustainable. Models must be built to understand diverse local languages for effective global mental health applications.
Abstract: Background: Large Language Models show promise in psychiatry but are English-centric. Their ability to understand mood states in other languages is unclear, as different languages have their own idioms of distress. Aim: To quantify the ability of language models to faithfully represent phrases (idioms of distress) of four distinct mood states (depression, euthymia, euphoric mania, dysphoric mania) expressed in Indian languages. Methods: We collected 247 unique phrases for the four mood states across 11 Indic languages. We tested seven experimental conditions, comparing k-means clustering performance on: (a) direct embeddings of native and Romanised scripts (using multilingual and Indic-specific models) and (b) embeddings of phrases translated to English and Chinese. Performance was measured using a composite score based on Adjusted Rand Index, Normalised Mutual Information, Homogeneity and Completeness. Results: Direct embedding of Indic languages failed to cluster mood states (Composite Score = 0.002). All translation-based approaches showed significant improvement. High performance was achieved using Gemini-translated English (Composite=0.60) and human-translated English (Composite=0.61) embedded with gemini-001. Surprisingly, human-translated English, further translated into Chinese and embedded with a Chinese model, performed best (Composite = 0.67). Specialised Indic models (IndicBERT and Sarvam-M) performed poorly. Conclusion: Current models cannot meaningfully represent mood states directly from Indic languages, posing a fundamental barrier to their psychiatric application for diagnostic or therapeutic purposes in India. While high-quality translation bridges this gap, reliance on proprietary models or complex translation pipelines is unsustainable. Models must first be built to understand diverse local languages to be effective in global mental health.
[10] EduEval: A Hierarchical Cognitive Benchmark for Evaluating Large Language Models in Chinese Education
Guoqing Ma, Jia Zhu, Hanghui Guo, Weijie Shi, Yue Cui, Jiawei Shen, Zilong Li, Yidan Liang
Main category: cs.CL
TL;DR: EduEval is a hierarchical benchmark for evaluating LLMs in Chinese K-12 education, featuring 24 task types with 11,000+ questions across cognitive dimensions, revealing models struggle with dialogue classification and creative tasks while open-source models sometimes outperform proprietary ones on reasoning.
Details
Motivation: LLMs show educational potential but unscrutinized deployment risks educational standards, requiring rigorous evaluation frameworks specifically for Chinese K-12 education contexts.Method: Created EduEval benchmark with three key contributions: (1) EduAbility Taxonomy unifying Bloom’s Taxonomy and Webb’s Depth of Knowledge across six cognitive dimensions; (2) Integration of real exam questions, classroom conversations, student essays, and expert prompts; (3) 24 task types with 11,000+ questions spanning primary to high school levels. Evaluated 14 LLMs under zero-shot and few-shot settings.
Result: Models perform well on factual tasks but struggle with classroom dialogue classification and show inconsistent creative content generation. Open-source models sometimes outperform proprietary systems on complex educational reasoning. Few-shot prompting effectiveness varies across cognitive dimensions.
Conclusion: Different educational objectives require tailored approaches, and EduEval provides targeted benchmarking metrics for developing LLMs optimized for diverse Chinese educational tasks.
Abstract: Large language models (LLMs) demonstrate significant potential for educational applications. However, their unscrutinized deployment poses risks to educational standards, underscoring the need for rigorous evaluation. We introduce EduEval, a comprehensive hierarchical benchmark for evaluating LLMs in Chinese K-12 education. This benchmark makes three key contributions: (1) Cognitive Framework: We propose the EduAbility Taxonomy, which unifies Bloom’s Taxonomy and Webb’s Depth of Knowledge to organize tasks across six cognitive dimensions including Memorization, Understanding, Application, Reasoning, Creativity, and Ethics. (2) Authenticity: Our benchmark integrates real exam questions, classroom conversation, student essays, and expert-designed prompts to reflect genuine educational challenges; (3) Scale: EduEval comprises 24 distinct task types with over 11,000 questions spanning primary to high school levels. We evaluate 14 leading LLMs under both zero-shot and few-shot settings, revealing that while models perform well on factual tasks, they struggle with classroom dialogue classification and exhibit inconsistent results in creative content generation. Interestingly, several open source models outperform proprietary systems on complex educational reasoning. Few-shot prompting shows varying effectiveness across cognitive dimensions, suggesting that different educational objectives require tailored approaches. These findings provide targeted benchmarking metrics for developing LLMs specifically optimized for diverse Chinese educational tasks.
[11] Comparative Analysis of 47 Context-Based Question Answer Models Across 8 Diverse Datasets
Muhammad Muneeb, David B. Ascher, Ahsan Baidar Bakht
Main category: cs.CL
TL;DR: Benchmark study of 47 context-based QA models from Hugging Face across 8 datasets without fine-tuning, identifying best-performing models and analyzing performance factors.
Details
Motivation: To identify the best-performing CBQA models across diverse datasets without additional fine-tuning, enabling practical applications where retraining models for specific datasets is minimized.Method: Benchmarked 47 CBQA models from Hugging Face on eight different datasets, analyzed performance factors (context length, model size, answer length, context complexity), and used Genetic algorithm to improve accuracy by integrating responses from other models.
Result: ahotrod/electra_large_discriminator_squad2_512 performed best overall (43% accuracy across all datasets), with top performance on bioasq10b-factoid (65.92%), biomedical_cpgQA (96.45%), QuAC (11.13%), and Question Answer Dataset (41.6%). Bert-large-uncased-whole-word-masking-finetuned-squad achieved 82% on IELTS dataset.
Conclusion: Best-performing models were trained on SQuAD datasets, performance depends on context length, model size, answer length, and context complexity, and model integration via Genetic algorithm can improve overall accuracy.
Abstract: Context-based question answering (CBQA) models provide more accurate and relevant answers by considering the contextual information. They effectively extract specific information given a context, making them functional in various applications involving user support, information retrieval, and educational platforms. In this manuscript, we benchmarked the performance of 47 CBQA models from Hugging Face on eight different datasets. This study aims to identify the best-performing model across diverse datasets without additional fine-tuning. It is valuable for practical applications where the need to retrain models for specific datasets is minimized, streamlining the implementation of these models in various contexts. The best-performing models were trained on the SQuAD v2 or SQuAD v1 datasets. The best-performing model was ahotrod/electra_large_discriminator_squad2_512, which yielded 43% accuracy across all datasets. We observed that the computation time of all models depends on the context length and the model size. The model’s performance usually decreases with an increase in the answer length. Moreover, the model’s performance depends on the context complexity. We also used the Genetic algorithm to improve the overall accuracy by integrating responses from other models. ahotrod/electra_large_discriminator_squad2_512 generated the best results for bioasq10b-factoid (65.92%), biomedical_cpgQA (96.45%), QuAC (11.13%), and Question Answer Dataset (41.6%). Bert-large-uncased-whole-word-masking-finetuned-squad achieved an accuracy of 82% on the IELTS dataset.
[12] Evidence-Guided Schema Normalization for Temporal Tabular Reasoning
Ashish Thanga, Vibhu Dixit, Abhilash Shankarampeta, Vivek Gupta
Main category: cs.CL
TL;DR: SQL-based QA system for temporal reasoning on Wikipedia infoboxes shows schema design quality matters more than model capacity for precision.
Details
Motivation: Temporal reasoning over evolving semi-structured tables (like Wikipedia infoboxes) is challenging for current QA systems, requiring better approaches to handle structured temporal data.Method: Three-step SQL-based approach: (1) generate 3NF schema from Wikipedia infoboxes, (2) generate SQL queries, (3) execute queries. Focuses on schema design principles including normalization preserving context, semantic naming reducing ambiguity, and consistent temporal anchoring.
Result: Best configuration (Gemini 2.5 Flash schema + Gemini-2.0-Flash queries) achieves 80.39 EM, a 16.8% improvement over baseline (68.89 EM). Key finding: schema design quality has greater impact on QA precision than model capacity.
Conclusion: For temporal reasoning over semi-structured tables, high-quality schema design (with proper normalization, semantic naming, and temporal anchoring) is more critical than model scaling, establishing evidence-based principles for effective QA systems.
Abstract: Temporal reasoning over evolving semi-structured tables poses a challenge to current QA systems. We propose a SQL-based approach that involves (1) generating a 3NF schema from Wikipedia infoboxes, (2) generating SQL queries, and (3) query execution. Our central finding challenges model scaling assumptions: the quality of schema design has a greater impact on QA precision than model capacity. We establish three evidence-based principles: normalization that preserves context, semantic naming that reduces ambiguity, and consistent temporal anchoring. Our best configuration (Gemini 2.5 Flash schema + Gemini-2.0-Flash queries) achieves 80.39 EM, a 16.8% improvement over the baseline (68.89 EM).
[13] Assertion-Conditioned Compliance: A Provenance-Aware Vulnerability in Multi-Turn Tool-Calling Agents
Daud Waqas, Aaryamaan Golthi, Erika Hayashida, Huanzhi Mao
Main category: cs.CL
TL;DR: The paper introduces Assertion-Conditioned Compliance (A-CC), a new evaluation framework for testing multi-turn tool-calling LLMs’ robustness against misleading assertions from users and system functions.
Details
Motivation: Multi-turn tool-calling LLMs are increasingly used in safety-critical applications, but there's a lack of visibility into their conversation-level robustness. Current benchmarks like BFCL don't adequately test resilience against misleading assertions in real-world multi-turn dialogues.Method: Proposes Assertion-Conditioned Compliance (A-CC) evaluation paradigm with holistic metrics to assess model behavior when confronted with two types of misleading assertions: User-Sourced Assertions (USAs) that test sycophancy toward misinformed user beliefs, and Function-Sourced Assertions (FSAs) that test compliance with contradictory system policies.
Result: Models show high vulnerability to both USA sycophancy (complying with plausible but misinformed user beliefs) and FSA policy conflicts (complying with contradictory system policies), revealing A-CC as a critical latent vulnerability in deployed agents.
Conclusion: A-CC identifies significant robustness gaps in multi-turn function-calling LLMs, highlighting the need for better evaluation of conversation-level vulnerabilities in safety-critical applications where models interact with real-world systems.
Abstract: Multi-turn tool-calling LLMs (models capable of invoking external APIs or tools across several user turns) have emerged as a key feature in modern AI assistants, enabling extended dialogues from benign tasks to critical business, medical, and financial operations. Yet implementing multi-turn pipelines remains difficult for many safety-critical industries due to ongoing concerns regarding model resilience. While standardized benchmarks such as the Berkeley Function-Calling Leaderboard (BFCL) have underpinned confidence concerning advanced function-calling models (like Salesforce’s xLAM V2), there is still a lack of visibility into multi-turn conversation-level robustness, especially given their exposure to real-world systems. In this paper, we introduce Assertion-Conditioned Compliance (A-CC), a novel evaluation paradigm for multi-turn function-calling dialogues. A-CC provides holistic metrics that evaluate a model’s behavior when confronted with misleading assertions originating from two distinct vectors: (1) user-sourced assertions (USAs), which measure sycophancy toward plausible but misinformed user beliefs, and (2) function-sourced assertions (FSAs), which measure compliance with plausible but contradictory system policies (e.g., stale hints from unmaintained tools). Our results show that models are highly vulnerable to both USA sycophancy and FSA policy conflicts, confirming A-CC as a critical, latent vulnerability in deployed agents.
[14] Multilingual DistilWhisper: Efficient Distillation of Multi-task Speech Models via Language-Specific Experts
Thomas Palmeira Ferraz, Marcely Zanon Boito, Caroline Brun, Vassilina Nikoulina
Main category: cs.CL
TL;DR: DistilWhisper improves ASR performance for under-represented languages in Whisper models using lightweight modular fine-tuning with language-specific experts and knowledge distillation from larger models.
Details
Motivation: Whisper underperforms on many under-represented languages, especially in smaller model versions, creating a performance gap that needs to be addressed while maintaining multitask and multilingual capabilities.Method: Two key strategies: 1) Lightweight modular ASR fine-tuning of whisper-small using language-specific experts, and 2) Knowledge distillation from whisper-large-v2 to boost performance while retaining robustness.
Result: More effective than standard fine-tuning or LoRA adapters, improves performance for targeted languages on both in- and out-of-domain test sets with negligible parameter overhead at inference.
Conclusion: DistilWhisper successfully bridges the ASR performance gap for under-represented languages while preserving the multitask and multilingual advantages of the original Whisper architecture.
Abstract: Whisper is a multitask and multilingual speech model covering 99 languages. It yields commendable automatic speech recognition (ASR) results in a subset of its covered languages, but the model still underperforms on a non-negligible number of under-represented languages, a problem exacerbated in smaller model versions. In this work, we propose DistilWhisper, an approach able to bridge the performance gap in ASR for these languages while retaining the advantages of multitask and multilingual capabilities. Our approach involves two key strategies: lightweight modular ASR fine-tuning of whisper-small using language-specific experts, and knowledge distillation from whisper-large-v2. This dual approach allows us to effectively boost ASR performance while keeping the robustness inherited from the multitask and multilingual pre-training. Results demonstrate that our approach is more effective than standard fine-tuning or LoRA adapters, boosting performance in the targeted languages for both in- and out-of-domain test sets, while introducing only a negligible parameter overhead at inference.
[15] NeKo: Cross-Modality Post-Recognition Error Correction with Tasks-Guided Mixture-of-Experts Language Model
Yen-Ting Lin, Zhehuai Chen, Piotr Zelasko, Zhen Wan, Xuesong Yang, Zih-Ching Chen, Krishna C Puvvada, Szu-Wei Fu, Ke Hu, Jun Wei Chiu, Jagadeesh Balam, Boris Ginsburg, Yu-Chiang Frank Wang, Chao-Han Huck Yang
Main category: cs.CL
TL;DR: Mixture-of-Experts approach for multi-domain post-recognition error correction achieves state-of-the-art performance across speech, translation, and vision tasks with significant WER reductions.
Details
Motivation: Need for a general-purpose post-recognition error corrector that can effectively handle multiple domains without parameter explosion from separate models.Method: Multi-Task Correction MoE with experts specialized for speech-to-text, language-to-text, and vision-to-text datasets, using token routing to mapped experts.
Result: Achieved 5.0% average relative WER reduction on Open ASR Leaderboard, outperformed GPT-3.5 and Claude-Opus by 15.5-27.6% in zero-shot evaluation, and performed competitively on grammar and post-OCR correction.
Conclusion: MoEs are effective for multi-domain error correction, enabling single-model expertise across domains without parameter explosion, establishing new SOTA performance.
Abstract: Construction of a general-purpose post-recognition error corrector poses a crucial question: how can we most effectively train a model on a large mixture of domain datasets? The answer would lie in learning dataset-specific features and digesting their knowledge in a single model. Previous methods achieve this by having separate correction language models, resulting in a significant increase in parameters. In this work, we present Mixture-of-Experts as a solution, highlighting that MoEs are much more than a scalability tool. We propose a Multi-Task Correction MoE, where we train the experts to become an ``expert’’ of speech-to-text, language-to-text and vision-to-text datasets by learning to route each dataset’s tokens to its mapped expert. Experiments on the Open ASR Leaderboard show that we explore a new state-of-the-art performance by achieving an average relative 5.0% WER reduction and substantial improvements in BLEU scores for speech and translation tasks. On zero-shot evaluation, NeKo outperforms GPT-3.5 and Claude-Opus with 15.5% to 27.6% relative WER reduction in the Hyporadise benchmark. NeKo performs competitively on grammar and post-OCR correction as a multi-task model.
[16] IndicParam: Benchmark to evaluate LLMs on low-resource Indic Languages
Ayush Maheshwari, Kaushal Sharma, Vivek Patel, Aditya Maheshwari
Main category: cs.CL
TL;DR: IndicParam is a human-curated benchmark of 13,000+ multiple-choice questions covering 11 low- and extremely low-resource Indic languages plus Sanskrit-English code-mixed set, revealing that even top LLMs like GPT-5 only achieve 45% average accuracy.
Details
Motivation: Low- and extremely low-resource Indic languages remain severely under-evaluated despite LLM advancements, creating a need for comprehensive benchmarks to assess model performance on these underrepresented languages.Method: Created a human-curated benchmark with over 13,000 multiple-choice questions covering 11 Indic languages (Nepali, Gujarati, Marathi, Odia as low-resource; Dogri, Maithili, Rajasthani, Sanskrit, Bodo, Santali, Konkani as extremely low-resource) plus Sanskrit-English code-mixed set. Evaluated 19 LLMs (proprietary and open-weights) and analyzed performance across different question types including knowledge-oriented vs. linguistic questions, and diverse formats like list-based matching, assertion-reason pairs, and sequence ordering.
Result: Even top-performing GPT-5 reached only 45.0% average accuracy, followed by DeepSeek-3.2 (43.1%) and Claude-4.5 (42.7%). The benchmark reveals significant limitations in cross-lingual transfer for Indic languages and establishes challenging performance baselines.
Conclusion: IndicParam provides crucial insights into LLM limitations for low-resource Indic languages, establishes a challenging benchmark, and highlights the need for improved cross-lingual transfer capabilities. The dataset and evaluation scripts are publicly available for further research.
Abstract: While large language models excel on high-resource multilingual tasks, low- and extremely low-resource Indic languages remain severely under-evaluated. We present IndicParam, a human-curated benchmark of over 13,000 multiple-choice questions covering 11 such languages (Nepali, Gujarati, Marathi, Odia as low-resource; Dogri, Maithili, Rajasthani, Sanskrit, Bodo, Santali, Konkani as extremely low-resource) plus Sanskrit-English code-mixed set. We evaluated 19 LLMs, both proprietary and open-weights, which reveals that even the top-performing GPT-5 reaches only 45.0% average accuracy, followed by DeepSeek-3.2 (43.1) and Claude-4.5 (42.7). We additionally label each question as knowledge-oriented or purely linguistic to discriminate factual recall from grammatical proficiency. Further, we assess the ability of LLMs to handle diverse question formats-such as list-based matching, assertion-reason pairs, and sequence ordering-alongside conventional multiple-choice questions. IndicParam provides insights into limitations of cross-lingual transfer and establishes a challenging benchmark for Indic languages. The dataset is available at https://huggingface.co/datasets/bharatgenai/IndicParam. Scripts to run benchmark are present at https://github.com/ayushbits/IndicParam.
[17] Comparative Evaluation of Expressive Japanese Character Text-to-Speech with VITS and Style-BERT-VITS2
Zackary Rackauckas, Julia Hirschberg
Main category: cs.CL
TL;DR: SBV2JE outperforms VITS for Japanese character speech synthesis, matching human naturalness and achieving better intelligibility with pitch-accent controls.
Details
Motivation: Japanese character speech synthesis faces challenges with pitch-accent sensitivity and stylistic variability, requiring specialized evaluation of TTS models for character-driven applications.Method: Empirical evaluation of VITS and Style-BERT-VITS2 JP Extra using three character-specific datasets, assessing naturalness (MOS, CMOS), intelligibility (WER), and speaker consistency.
Result: SBV2JE matches human ground truth in naturalness (MOS 4.37 vs 4.38), achieves lower WER, shows slight CMOS preference, and demonstrates effective pitch-accent control despite higher computational demands.
Conclusion: SBV2JE proves effective for Japanese character speech applications like language learning and dialogue generation, outperforming VITS through enhanced pitch-accent controls and WavLM-based discriminator.
Abstract: Synthesizing expressive Japanese character speech poses unique challenges due to pitch-accent sensitivity and stylistic variability. This paper empirically evaluates two open-source text-to-speech models–VITS and Style-BERT-VITS2 JP Extra (SBV2JE)–on in-domain, character-driven Japanese speech. Using three character-specific datasets, we evaluate models across naturalness (mean opinion and comparative mean opinion score), intelligibility (word error rate), and speaker consistency. SBV2JE matches human ground truth in naturalness (MOS 4.37 vs. 4.38), achieves lower WER, and shows slight preference in CMOS. Enhanced by pitch-accent controls and a WavLM-based discriminator, SBV2JE proves effective for applications like language learning and character dialogue generation, despite higher computational demands.
[18] SpeechIQ: Speech-Agentic Intelligence Quotient Across Cognitive Levels in Voice Understanding by Large Language Models
Zhen Wan, Chao-Han Huck Yang, Yahan Yu, Jinchuan Tian, Sheng Li, Ke Hu, Zhehuai Chen, Shinji Watanabe, Fei Cheng, Chenhui Chu, Sadao Kurohashi
Main category: cs.CL
TL;DR: SIQ is a new cognitive evaluation pipeline for voice understanding LLMs that assesses across three Bloom’s Taxonomy levels: remembering (WER), understanding (interpretation similarity), and application (QA accuracy).
Details
Motivation: Current voice understanding metrics like WER are limited and don't capture deeper cognitive abilities. There's a need for a more comprehensive, human cognition-inspired evaluation framework that can assess voice understanding LLMs across multiple cognitive levels and provide unified comparisons between different approaches.Method: SIQ evaluates voice understanding LLMs across three cognitive levels: (1) Remembering - using WER for verbatim accuracy, (2) Understanding - measuring similarity of LLM’s interpretations, and (3) Application - using QA accuracy for simulating downstream tasks. The framework bridges cognitive principles with voice-oriented benchmarks.
Result: SIQ successfully quantifies voice understanding abilities, provides unified comparisons between cascaded (ASR+LLM) and end-to-end models, identifies annotation errors in existing benchmarks, and detects hallucinations in LLM Voice systems.
Conclusion: SIQ represents a first-of-its-kind intelligence examination that bridges cognitive principles with voice understanding evaluation, exposing overlooked challenges in multi-modal training while providing a comprehensive assessment framework for voice understanding LLMs.
Abstract: We introduce Speech-based Intelligence Quotient (SIQ) as a new form of human cognition-inspired evaluation pipeline for voice understanding large language models, LLM Voice, designed to assess their voice understanding ability. Moving beyond popular voice understanding metrics such as word error rate (WER), SIQ examines LLM Voice across three cognitive levels motivated by Bloom’s Taxonomy: (1) Remembering (i.e., WER for verbatim accuracy); (2) Understanding (i.e., similarity of LLM’s interpretations); and (3) Application (i.e., QA accuracy for simulating downstream tasks). We demonstrate that SIQ not only quantifies voice understanding abilities but also provides unified comparisons between cascaded methods (e.g., ASR LLM) and end-to-end models, identifies annotation errors in existing benchmarks, and detects hallucinations in LLM Voice. Our framework represents a first-of-its-kind intelligence examination that bridges cognitive principles with voice-oriented benchmarks, while exposing overlooked challenges in multi-modal training. Our code and data will be open source to encourage future studies.
[19] CourseTimeQA: A Lecture-Video Benchmark and a Latency-Constrained Cross-Modal Fusion Method for Timestamped QA
Vsevolod Kovalev, Parteek Kumar
Main category: cs.CL
TL;DR: A lightweight cross-modal retriever (CrossFusion-RAG) for timestamped QA on lecture videos achieves 1.55s latency on single A100 while improving retrieval metrics over strong baselines.
Details
Motivation: Address timestamped question answering over educational lecture videos under strict single-GPU latency and memory constraints, requiring efficient cross-modal retrieval that combines visual and audio/text information.Method: Propose CrossFusion-RAG: combines frozen encoders, learned vision projection (512→768), shallow query-agnostic cross-attention over ASR and frames with temporal-consistency regularization, and small cross-attentive reranker. Evaluated on CourseTimeQA dataset (52.3h, 902 queries across 6 courses).
Result: Improves nDCG@10 by 0.10 and MRR by 0.08 over strong BLIP-2 retriever, achieves ~1.55s median end-to-end latency on single A100. Outperforms multiple baselines under matched hardware/indexing constraints, shows robustness to ASR noise.
Conclusion: CrossFusion-RAG provides effective, latency-constrained solution for educational video QA, balancing retrieval performance with practical deployment constraints on limited hardware.
Abstract: We study timestamped question answering over educational lecture videos under a single-GPU latency/memory budget. Given a natural-language query, the system retrieves relevant timestamped segments and synthesizes a grounded answer. We present CourseTimeQA (52.3 h, 902 queries across six courses) and a lightweight, latency-constrained cross-modal retriever (CrossFusion-RAG) that combines frozen encoders, a learned 512->768 vision projection, shallow query-agnostic cross-attention over ASR and frames with a temporal-consistency regularizer, and a small cross-attentive reranker. On CourseTimeQA, CrossFusion-RAG improves nDCG@10 by 0.10 and MRR by 0.08 over a strong BLIP-2 retriever while achieving approximately 1.55 s median end-to-end latency on a single A100. Closest comparators (zero-shot CLIP multi-frame pooling; CLIP + cross-encoder reranker + MMR; learned late-fusion gating; text-only hybrid with cross-encoder reranking and its MMR variant; caption-augmented text retrieval; non-learned temporal smoothing) are evaluated under matched hardware and indexing. We report robustness across ASR noise (WER quartiles), diagnostics for temporal localization, and full training/tuning details to support reproducible comparison.
[20] Mitigating the Threshold Priming Effect in Large Language Model-Based Relevance Judgments via Personality Infusing
Nuo Chen, Hanpei Fang, Jiqun Liu, Wilson Wei, Tetsuya Sakai, Xiao-Ming Wu
Main category: cs.CL
TL;DR: LLMs with specific Big Five personality profiles (High Openness, Low Neuroticism) show reduced susceptibility to priming effects in relevance labeling tasks.
Details
Motivation: While LLMs are used for scalable relevance labeling, they suffer from priming effects where prior judgments influence later ones. Psychological theories link personality traits to such biases, but it's unclear if simulated personalities in LLMs exhibit similar effects.Method: Investigated how Big Five personality profiles in LLMs influence priming in relevance labeling using multiple LLMs on TREC 2021 and 2022 Deep Learning Track datasets.
Result: Certain personality profiles (High Openness and Low Neuroticism) consistently reduce priming susceptibility. The most effective personality for mitigating priming varies across models and task types.
Conclusion: Propose personality prompting as a method to mitigate threshold priming, connecting psychological evidence with LLM-based evaluation practices.
Abstract: Recent research has explored LLMs as scalable tools for relevance labeling, but studies indicate they are susceptible to priming effects, where prior relevance judgments influence later ones. Although psychological theories link personality traits to such biases, it is unclear whether simulated personalities in LLMs exhibit similar effects. We investigate how Big Five personality profiles in LLMs influence priming in relevance labeling, using multiple LLMs on TREC 2021 and 2022 Deep Learning Track datasets. Our results show that certain profiles, such as High Openness and Low Neuroticism, consistently reduce priming susceptibility. Additionally, the most effective personality in mitigating priming may vary across models and task types. Based on these findings, we propose personality prompting as a method to mitigate threshold priming, connecting psychological evidence with LLM-based evaluation practices.
[21] InnoGym: Benchmarking the Innovation Potential of AI Agents
Jintian Zhang, Kewei Xu, Jingsheng Zheng, Zhuoyun Yu, Yuqi Zhu, Yujie Luo, Lanning Wei, Shuofei Qiao, Lun Du, Da Zheng, Shumin Deng, Huajun Chen, Ningyu Zhang
Main category: cs.CL
TL;DR: InnoGym is the first benchmark and framework to evaluate AI agents’ innovation potential using performance gain and novelty metrics across 18 real-world engineering/scientific tasks.
Details
Motivation: Existing benchmarks focus only on correctness, overlooking solution diversity and originality. True innovation requires both correct answers and novel approaches, but current evaluation lacks systematic assessment of innovation potential.Method: InnoGym introduces two complementary metrics: performance gain (improvement over best-known solutions) and novelty (methodological differences from prior approaches). The benchmark includes 18 curated real-world tasks with standardized resource filtering, evaluator validation, and solution collection. Also provides iGym, a unified execution environment for reproducible, long-horizon evaluations.
Result: Experiments show some agents produce novel approaches but lack robustness, limiting performance gains. This reveals a gap between creativity and effectiveness in current AI agents.
Conclusion: There’s a need for benchmarks that evaluate both correctness and innovation. InnoGym addresses this gap by systematically assessing innovation potential through performance gain and novelty metrics, highlighting the importance of balancing creativity with effectiveness.
Abstract: LLMs and Agents have achieved impressive progress in code generation, mathematical reasoning, and scientific discovery. However, existing benchmarks primarily measure correctness, overlooking the diversity of methods behind solutions. True innovation depends not only on producing correct answers but also on the originality of the approach. We present InnoGym, the first benchmark and framework designed to systematically evaluate the innovation potential of AI agents. InnoGym introduces two complementary metrics: performance gain, which measures improvement over the best-known solutions, and novelty, which captures methodological differences from prior approaches. The benchmark includes 18 carefully curated tasks from real-world engineering and scientific domains, each standardized through resource filtering, evaluator validation, and solution collection. In addition, we provide iGym, a unified execution environment for reproducible and long-horizon evaluations. Extensive experiments show that while some agents produce novel approaches, their lack of robustness limits performance gains. These results highlight a key gap between creativity and effectiveness, underscoring the need for benchmarks that evaluate both.
[22] A Taxonomy of Errors in English as she is spoke: Toward an AI-Based Method of Error Analysis for EFL Writing Instruction
Damian Heywood, Joseph Andrew Carrier, Kyu-Hong Hwang
Main category: cs.CL
TL;DR: AI system using LLMs (Claude 3.5 Sonnet & DeepSeek R1) analyzes English writing errors with linguistic taxonomy, providing detailed feedback beyond traditional rubrics.
Details
Motivation: To transform EFL instruction by automating detailed error analysis and feedback, moving beyond traditional rubric-based assessments to provide more granular linguistic analysis.Method: Uses LLMs with Python-coded API calls, implements error taxonomy based on Corder (1967), Richards (1971), and James (1998), classifies errors at word/sentence levels (spelling, grammar, punctuation), tested on isolated errors and authentic text “English as she is spoke”.
Result: AI successfully identified diverse error types but showed limitations in contextual understanding and occasionally generated new error categories for uncoded errors.
Conclusion: Demonstrates AI’s potential to transform EFL instruction through automated error analysis, but needs further development for contextual accuracy and expansion to stylistic/discourse-level errors.
Abstract: This study describes the development of an AI-assisted error analysis system designed to identify, categorize, and correct writing errors in English. Utilizing Large Language Models (LLMs) like Claude 3.5 Sonnet and DeepSeek R1, the system employs a detailed taxonomy grounded in linguistic theories from Corder (1967), Richards (1971), and James (1998). Errors are classified at both word and sentence levels, covering spelling, grammar, and punctuation. Implemented through Python-coded API calls, the system provides granular feedback beyond traditional rubric-based assessments. Initial testing on isolated errors refined the taxonomy, addressing challenges like overlapping categories. Final testing used “English as she is spoke” by Jose da Fonseca (1855), a text rich with authentic linguistic errors, to evaluate the system’s capacity for handling complex, multi-layered analysis. The AI successfully identified diverse error types but showed limitations in contextual understanding and occasionally generated new error categories when encountering uncoded errors. This research demonstrates AI’s potential to transform EFL instruction by automating detailed error analysis and feedback. While promising, further development is needed to improve contextual accuracy and expand the taxonomy to stylistic and discourse-level errors.
[23] CryptoBench: A Dynamic Benchmark for Expert-Level Evaluation of LLM Agents in Cryptocurrency
Jiacheng Guo, Suozhi Huang, Zixin Yao, Yifan Zhang, Yifu Lu, Jiashuo Liu, Zihao Li, Yanyan Deng, Qixin Xiao, Jia Tian, Kanghong Zhan, Tianyi Li, Xiaochen Liu, Jason Ge, Chaoyang He, Kaixuan Huang, Lin Yang, Wenhao Huang, Mengdi Wang
Main category: cs.CL
TL;DR: CryptoBench is the first expert-curated dynamic benchmark for evaluating LLM agents in cryptocurrency analysis, featuring time-sensitive, adversarial tasks across retrieval and prediction categories, revealing a retrieval-prediction imbalance in current models.
Details
Motivation: Existing benchmarks don't capture the unique challenges of cryptocurrency analysis: extreme time-sensitivity, adversarial information environments, and need to synthesize data from diverse specialized sources like on-chain intelligence platforms and real-time DeFi dashboards.Method: Created a live dynamic benchmark with 50 questions per month designed by crypto-native professionals, categorized into four quadrants: Simple Retrieval, Complex Retrieval, Simple Prediction, and Complex Prediction. Evaluated 10 LLMs both directly and within agentic frameworks.
Result: Revealed a performance hierarchy and uncovered a “retrieval-prediction imbalance” - many leading models are proficient at data retrieval but show pronounced weakness in predictive analysis tasks, highlighting models that appear factually grounded but lack deeper analytical synthesis capabilities.
Conclusion: CryptoBench provides a more challenging and valuable scenario for LLM agent assessment in cryptocurrency domains, exposing critical gaps in current models’ analytical capabilities despite their retrieval strengths, and establishes a framework for more rigorous evaluation of real-world crypto analysis performance.
Abstract: This paper introduces CryptoBench, the first expert-curated, dynamic benchmark designed to rigorously evaluate the real-world capabilities of Large Language Model (LLM) agents in the uniquely demanding and fast-paced cryptocurrency domain. Unlike general-purpose agent benchmarks for search and prediction, professional crypto analysis presents specific challenges: \emph{extreme time-sensitivity}, \emph{a highly adversarial information environment}, and the critical need to synthesize data from \emph{diverse, specialized sources}, such as on-chain intelligence platforms and real-time Decentralized Finance (DeFi) dashboards. CryptoBench thus serves as a much more challenging and valuable scenario for LLM agent assessment. To address these challenges, we constructed a live, dynamic benchmark featuring 50 questions per month, expertly designed by crypto-native professionals to mirror actual analyst workflows. These tasks are rigorously categorized within a four-quadrant system: Simple Retrieval, Complex Retrieval, Simple Prediction, and Complex Prediction. This granular categorization enables a precise assessment of an LLM agent’s foundational data-gathering capabilities alongside its advanced analytical and forecasting skills. Our evaluation of ten LLMs, both directly and within an agentic framework, reveals a performance hierarchy and uncovers a failure mode. We observe a \textit{retrieval-prediction imbalance}, where many leading models, despite being proficient at data retrieval, demonstrate a pronounced weakness in tasks requiring predictive analysis. This highlights a problematic tendency for agents to appear factually grounded while lacking the deeper analytical capabilities to synthesize information.
[24] SCALE: Selective Resource Allocation for Overcoming Performance Bottlenecks in Mathematical Test-time Scaling
Yang Xiao, Chunpu Xu, Ruifeng Yuan, Jiashuo Wang, Wenjie Li, Pengfei Liu
Main category: cs.CL
TL;DR: SCALE is a selective resource allocation framework for LLM mathematical reasoning that dynamically allocates compute based on sub-problem difficulty, outperforming uniform scaling methods while reducing computational costs.
Details
Motivation: Current test-time compute scaling methods use uniform resource distribution across all reasoning sub-problems, creating bottlenecks where challenging sub-problems get insufficient attention while routine operations waste resources, leading to diminishing returns from additional compute.Method: Four-stage framework: (1) decompose problem into sequential reasoning sub-problems, (2) assess difficulty of each sub-problem, (3) assign selective processing modes (System 1 for simple, System 2 for complex), (4) execute sequentially with context propagation.
Result: Achieves up to 13.75 percentage point accuracy improvement (57.50% to 71.25% on AIME25) while reducing computational costs by 33%-53%, significantly outperforming uniform scaling baselines.
Conclusion: SCALE represents a major advance in test-time scaling by addressing fundamental limitations of uniform resource allocation, enabling superior performance with better resource utilization through selective compute allocation based on sub-problem difficulty.
Abstract: Test-time compute scaling has emerged as a powerful paradigm for enhancing mathematical reasoning in large language models (LLMs) by allocating additional computational resources during inference. However, current methods employ uniform resource distribution across all reasoning sub-problems, creating fundamental bottlenecks where challenging sub-problems receive insufficient attention while routine operations consume disproportionate resources. This uniform allocation creates performance bottlenecks where additional computational resources yield diminishing returns. Inspired by dual-process theory, we propose \textbf{SCALE} (Selective Resource Allocation), a framework that selectively allocates computational resources based on sub-problem difficulty. SCALE operates through four stages: (1) problem decomposition into sequential reasoning sub-problems, (2) difficulty assessment of each sub-problem to distinguish between routine operations and computationally challenging sub-problems, (3) selective processing mode assignment between System 1 for simple sub-problems and System 2 for complex ones, and (4) sequential execution with context propagation. By concentrating resources on challenging sub-problems while processing routine operations efficiently, SCALE achieves substantial performance improvements with superior resource utilization. Extensive experiments demonstrate that SCALE significantly outperforms uniform scaling baselines, achieving accuracy improvements of up to 13.75 percentage points (57.50% to 71.25% on AIME25) while reducing computational costs by 33%-53%, representing a major advance in test-time scaling that addresses fundamental limitations of current approaches.
[25] CACARA: Cross-Modal Alignment Leveraging a Text-Centric Approach for Cost-Effective Multimodal and Multilingual Learning
Diego A. B. Moreira, Alef I. Ferreira, Jhessica Silva, Gabriel O. dos Santos, Gustavo Bonil, João Gondim, Marina dos Santos, Helena Maia, Simone Hashiguti, Nádia da Silva, Carolina Scarton, Helio Pedrini, Sandra Avila
Main category: cs.CL
TL;DR: CACARA is a multimodal and multilingual architecture using emergent alignment learning that enables adding new modalities to existing models without full retraining, and surprisingly develops multilingual capabilities from monolingual training.
Details
Motivation: Current multimodal models require resource-intensive training across all modalities, and extending them to new languages follows the same heavy training strategy. There's a need for more efficient ways to integrate new modalities and languages without full retraining.Method: Proposes CACARA architecture trained through emergent alignment learning, which allows seamless integration of new modalities into existing bimodal/multimodal models. Only fine-tunes newly incorporated modalities on English-aligned data, without explicit multilingual pretraining or text encoder tuning.
Result: Achieves up to 14.24 percentage points improvement in R@1 audio-to-text retrieval, outperforming state-of-the-art multimodal models. Gains support for over 100 languages without explicit multilingual training, with training cost comparable to monolingual models.
Conclusion: Emergent alignment learning enables efficient multimodal and multilingual capabilities, preserving previously learned knowledge while avoiding heavy computational costs of retraining across every modality and language.
Abstract: As deep learning models evolve, new applications and challenges are rapidly emerging. Tasks that once relied on a single modality, such as text, images, or audio, are now enriched by seamless interactions between multimodal data. These connections bridge information gaps: an image can visually materialize a text, while audio can add context to an image. Researchers have developed numerous multimodal models, but most rely on resource-intensive training across multiple modalities. Similarly, extending these models to new languages often follows the same resource-heavy training strategy. In this work, we propose a multimodal and multilingual architecture, CACARA, trained through emergent alignment learning, enabling the seamless integration of new modalities into an existing bimodal/multimodal model without requiring full retraining. This work breaks new ground by demonstrating that this emergent alignment paradigm can unlock multilingual capabilities from monolingual training. By fine-tuning the newly incorporated modality only on data aligned with the English language, our model develops support for over 100 languages without explicit multilingual pretraining or tuning of the text encoder. Such emergent multimodal and multilingual properties are gained efficiently, preserving previously learned knowledge at a training cost comparable to that of a monolingual model. Our strategy achieves up to a 14.24 percentage points improvement in R@1 audio-to-text retrieval, outperforming state-of-the-art multimodal models – all without the heavy computational cost of retraining across every modality and language.
[26] G-KV: Decoding-Time KV Cache Eviction with Global Attention
Mengqi Liao, Lu Wang, Chaoyun Zhang, Zekai Shen, Xiaowei Mao, Si Qin, Qingwei Lin, Saravan Rajmohan, Dongmei Zhang, Huaiyu Wan
Main category: cs.CL
TL;DR: G-KV is a KV cache eviction method that uses global scoring (combining local and historical attention scores) to better assess token importance, with post-training techniques to optimize models for compressed KV cache settings.
Details
Motivation: Large language models face computational and memory challenges with long sequences. Existing KV cache compression methods focus on prompt compression or local attention scores, overlooking long-term token importance.Method: Proposes G-KV with global scoring mechanism combining local and historical attention scores to assess token importance. Uses post-training techniques including reinforcement learning and distillation to optimize models for compressed KV cache.
Result: The method is available as open source code on GitHub. The abstract suggests it addresses limitations of existing approaches by considering long-term token importance.
Conclusion: G-KV provides a more accurate KV cache eviction method through global scoring and post-training optimization, improving efficiency for reasoning LLMs with long sequences.
Abstract: Recent reasoning large language models (LLMs) excel in complex tasks but encounter significant computational and memory challenges due to long sequence lengths. KV cache compression has emerged as an effective approach to greatly enhance the efficiency of reasoning. However, existing methods often focus on prompt compression or token eviction with local attention score, overlooking the long-term importance of tokens. We propose G-KV, a KV cache eviction method that employs a global scoring mechanism, combining local and historical attention scores to more accurately assess token importance. Additionally, we introduce post-training techniques, including reinforcement learning and distillation, to optimize models for compressed KV cache settings. The code of this paper is available on: https://github.com/microsoft/G-KV.
[27] Developing a Comprehensive Framework for Sentiment Analysis in Turkish
Cem Rifki Aydin
Main category: cs.CL
TL;DR: A comprehensive sentiment analysis framework with novel feature sets, neural architectures, and language-specific approaches for Turkish and English, achieving state-of-the-art results.
Details
Motivation: To develop a comprehensive sentiment analysis framework that addresses multiple aspects of the problem, with particular focus on Turkish (which lacks extensive research) while also contributing to English sentiment analysis.Method: Combined unsupervised, semi-supervised, and supervised metrics to create novel feature sets; used classical ML methods; built novel neural architectures combining recurrent and recursive networks; created domain-specific polarity lexicons; performed fine morphological analysis for Turkish; developed sentiment-aware word embeddings; redefined context windows as subclauses.
Result: Achieved state-of-the-art results for both Turkish and English sentiment analysis; outperformed neural network models with classical ML approaches; created the first domain-specific polarity lexicon for Turkish; developed novel neural architectures and word embeddings.
Conclusion: This thesis represents the most detailed and comprehensive study on Turkish sentiment analysis as of July 2020, with significant contributions to both Turkish and English sentiment analysis through novel approaches that can be adapted to other languages and NLP tasks.
Abstract: In this thesis, we developed a comprehensive framework for sentiment analysis that takes its many aspects into account mainly for Turkish. We have also proposed several approaches specific to sentiment analysis in English only. We have accordingly made five major and three minor contributions. We generated a novel and effective feature set by combining unsupervised, semi-supervised, and supervised metrics. We then fed them as input into classical machine learning methods, and outperformed neural network models for datasets of different genres in both Turkish and English. We created a polarity lexicon with a semi-supervised domain-specific method, which has been the first approach applied for corpora in Turkish. We performed a fine morphological analysis for the sentiment classification task in Turkish by determining the polarities of morphemes. This can be adapted to other morphologically-rich or agglutinative languages as well. We have built a novel neural network architecture, which combines recurrent and recursive neural network models for English. We built novel word embeddings that exploit sentiment, syntactic, semantic, and lexical characteristics for both Turkish and English. We also redefined context windows as subclauses in modelling word representations in English. This can also be applied to other linguistic fields and natural language processing tasks. We have achieved state-of-the-art and significant results for all these original approaches. Our minor contributions include methods related to aspect-based sentiment in Turkish, parameter redefinition in the semi-supervised approach, and aspect term extraction techniques for English. This thesis can be considered the most detailed and comprehensive study made on sentiment analysis in Turkish as of July, 2020. Our work has also contributed to the opinion classification problem in English.
[28] Catch Me If You Can: How Smaller Reasoning Models Pretend to Reason with Mathematical Fidelity
Subramanyam Sahoo, Vinija Jain, Saanidhya Vats, Siddharth Mohapatra, Rui Min, Aman Chadha, Divya Chaudhary
Main category: cs.CL
TL;DR: The paper introduces a diagnostic framework to evaluate genuine mathematical reasoning vs. pattern matching in language models, revealing that Qwen3-0.6B shows poor reasoning fidelity despite decent accuracy.
Details
Motivation: Current evaluation metrics (answer accuracy) may mask fundamental failures in logical computation, creating a need for more comprehensive diagnostic tools to assess genuine mathematical reasoning capabilities.Method: Developed a diagnostic framework with four axes: forward-backward consistency, transitivity coverage, counterfactual sensitivity, and perturbation robustness. Applied this to Qwen3-0.6B on the MenatQA dataset.
Result: Qwen3-0.6B showed reasonable answer accuracy (70%+) but poor backward consistency (15%), limited transitivity coverage (32.2%), and brittle sensitivity to perturbations, indicating reliance on pattern matching rather than genuine reasoning.
Conclusion: The diagnostic framework exposes reasoning failures invisible to traditional accuracy metrics, suggesting current models may rely on superficial pattern matching. The framework is model-agnostic and released to enable broader assessment of reasoning fidelity.
Abstract: Current evaluation of mathematical reasoning in language models relies primarily on answer accuracy, potentially masking fundamental failures in logical computation. We introduce a diagnostic framework that distinguishes genuine mathematical reasoning from superficial pattern matching through four complementary axes: forward-backward consistency, transitivity coverage, counterfactual sensitivity, and perturbation robustness. Through a case study applying this framework to Qwen3-0.6B on the MenatQA dataset, we reveal a striking disconnect between surface performance and reasoning fidelity. While the model achieves reasonable answer accuracy (70%+), it demonstrates poor backward consistency (15%), limited transitivity coverage (32.2%), and brittle sensitivity to perturbations. Our diagnostics expose reasoning failures invisible to traditional accuracy metrics, suggesting that this small model relies heavily on pattern matching rather than genuine logical computation. While our empirical findings are based on a single 600M-parameter model, the diagnostic framework itself is model-agnostic and generalizable. We release our evaluation protocols to enable the research community to assess reasoning fidelity across different model scales and architectures, moving beyond surface-level accuracy toward verifiable mathematical reasoning.
[29] Slovak Conceptual Dictionary
Miroslav Blšták
Main category: cs.CL
TL;DR: Introduces first conceptual dictionary for Slovak language to address NLP resource limitations
Details
Motivation: Slovak language lacks machine-readable linguistic resources, leading to weaker NLP results compared to other languages. Many NLP tasks require dictionary tools but these are unavailable for low-resource languages like Slovak.Method: Creation of a new conceptual dictionary specifically for Slovak language - the first linguistic tool of this kind for Slovak.
Result: Development of the first conceptual dictionary for Slovak language, addressing the gap in machine-readable linguistic resources.
Conclusion: This new conceptual dictionary represents a significant step forward for Slovak NLP, enabling better automated text processing and addressing the resource limitations that have hindered Slovak language processing tasks.
Abstract: When solving tasks in the field of natural language processing, we sometimes need dictionary tools, such as lexicons, word form dictionaries or knowledge bases. However, the availability of dictionary data is insufficient in many languages, especially in the case of low resourced languages. In this article, we introduce a new conceptual dictionary for the Slovak language as the first linguistic tool of this kind. Since Slovak language is a language with limited linguistic resources and there are currently not available any machine-readable linguistic data sources with a sufficiently large volume of data, many tasks which require automated processing of Slovak text achieve weaker results compared to other languages and are almost impossible to solve.
[30] Wikontic: Constructing Wikidata-Aligned, Ontology-Aware Knowledge Graphs with Large Language Models
Alla Chepurova, Aydar Bulatov, Yuri Kuratov, Mikhail Burtsev
Main category: cs.CL
TL;DR: Wikontic is a multi-stage pipeline that constructs high-quality, ontology-consistent knowledge graphs from open-domain text, achieving strong performance on QA benchmarks while being more efficient than existing methods.
Details
Motivation: Current LLM-based systems use knowledge graphs mainly as auxiliary structures for text retrieval, leaving their intrinsic quality underexplored. There's a need for better methods to construct high-quality KGs that can provide structured, verifiable grounding for LLMs.Method: A multi-stage pipeline that: 1) extracts candidate triplets with qualifiers from text, 2) enforces Wikidata-based type and relation constraints for consistency, and 3) normalizes entities to reduce duplication. The approach focuses on creating compact, well-connected KGs.
Result: Achieves 96% correct answer entity coverage on MuSiQue, 76.0 F1 on HotpotQA, and 59.8 F1 on MuSiQue - matching or surpassing retrieval-augmented baselines. Sets SOTA on MINE-1 benchmark (86%). Highly efficient: uses <1,000 output tokens (3× fewer than AriGraph, <1/20 of GraphRAG).
Conclusion: Wikontic enhances KG quality and offers a scalable solution for leveraging structured knowledge in LLMs, demonstrating that high-quality KGs alone can match or exceed performance of systems requiring textual context.
Abstract: Knowledge graphs (KGs) provide structured, verifiable grounding for large language models (LLMs), but current LLM-based systems commonly use KGs as auxiliary structures for text retrieval, leaving their intrinsic quality underexplored. In this work, we propose Wikontic, a multi-stage pipeline that constructs KGs from open-domain text by extracting candidate triplets with qualifiers, enforcing Wikidata-based type and relation constraints, and normalizing entities to reduce duplication. The resulting KGs are compact, ontology-consistent, and well-connected; on MuSiQue, the correct answer entity appears in 96% of generated triplets. On HotpotQA, our triplets-only setup achieves 76.0 F1, and on MuSiQue 59.8 F1, matching or surpassing several retrieval-augmented generation baselines that still require textual context. In addition, Wikontic attains state-of-the-art information-retention performance on the MINE-1 benchmark (86%), outperforming prior KG construction methods. Wikontic is also efficient at build time: KG construction uses less than 1,000 output tokens, about 3$\times$ fewer than AriGraph and $<$1/20 of GraphRAG. The proposed pipeline enhances the quality of the generated KG and offers a scalable solution for leveraging structured knowledge in LLMs.
[31] Prism: A Minimal Compositional Metalanguage for Specifying Agent Behavior
Franck Binard, Vanja Kljajevic
Main category: cs.CL
TL;DR: Prism is a compositional metalanguage for specifying tool-using agent behavior using a core grammar with abstract combinators, allowing domain-specific extensions while maintaining inspectable, executable policies.
Details
Motivation: The paper addresses the need for a systematic, compositional approach to specifying tool-using software agent behavior, moving away from ad hoc control constructs. It aims to provide a clear separation between reusable core grammar and domain-specific extensions while making agent policies inspectable, executable, and amenable to analysis.Method: Prism is built around Core1, a minimal background grammar with categories (numbers, strings, user prompts, tools) and abstract combinators (booleans, predicates, pairs, lists). Agent policies are written as ordinary expressions using a single abstraction operator, with conditionals as selections between alternatives. Domains extend the core by defining context-mini-grammars that introduce new categories, predicates, and external tools while reusing the same compositional machinery.
Result: The authors demonstrate Prism through worked examples from thermostat control, home security, e-commerce recommendation, and medical monitoring, showing how natural language decision rules can be mapped to inspectable, executable policies. The approach provides a compact interface language for agent control that makes action spaces explicit.
Conclusion: Prism offers both linguistic and engineering benefits: linguistically, it enforces separation between reusable grammar-like core and domain-specific lexicons, treating tools as bridges between internal representations and external world; engineering-wise, it provides a compact interface language amenable to analysis, verification, and safety constraints for tool-using agents.
Abstract: Prism is a small, compositional metalanguage for specifying the behaviour of tool-using software agents. Rather than introducing ad hoc control constructs, Prism is built around a fixed core context, Core1, which provides a minimal background grammar of categories numbers, strings, user prompts, tools together with abstract combinators for booleans, predicates, pairs, and lists. Agent policies are written as ordinary expressions using a single abstraction operator so that conditionals appear as selections between alternatives instead of imperative if-else blocks. Domains extend the core by defining their own context-mini-grammars that introduce new categories, predicates, and external tools while reusing the same compositional machinery. We illustrate this with worked examples from thermostat control, home security, e-commerce recommendation, and medical monitoring, showing how natural language decision rules can be mapped to inspectable, executable policies. From a linguistic perspective, Prism enforces a clear separation between a reusable grammar-like core and domain specific lexicons and treats tools as bridges between internal policy representations and the external world. From an engineering perspective, it offers a compact interface language for agent control, making the space of possible actions explicit and amenable to analysis, verification, and safety constraints.
[32] ART: Adaptive Response Tuning Framework – A Multi-Agent Tournament-Based Approach to LLM Response Optimization
Omer Jauhar Khan
Main category: cs.CL
TL;DR: ART framework uses tournament-style ELO ranking and multi-agent reasoning to optimize LLM outputs through competition and collaboration, achieving 8.4% quality improvement.
Details
Motivation: Single LLM responses suffer from inconsistencies, hallucinations, and varying quality across domains, requiring a systematic approach to produce more reliable outputs.Method: Tournament-style ELO ranking with multi-agent reasoning where multiple LLM agents compete, critique, and collaborate through structured workflows with configurable parameters, dynamic agent selection, and consensus fusion strategies.
Result: Significant improvements in response accuracy, coherence, and reliability with 8.4% overall quality improvement and R² values exceeding 0.96 in ELO rating convergence.
Conclusion: ART provides a scalable, production-ready solution for high-quality, vetted LLM responses by leveraging multi-agent competition and collaboration to outperform individual models.
Abstract: Large Language Models (LLMs) have demonstrated remarkable capabilities in natural language understanding and generation. However, single-model responses often exhibit inconsistencies, hallucinations, and varying quality across different query domains. This paper presents ART (Adaptive Response Tuning), a novel framework that employs tournament-style ELO ranking and multi-agent reasoning to systematically optimize LLM outputs. By enabling multiple LLM agents to compete, critique, and collaborate through structured tournament workflows, ART produces consensus responses that outperform individual model outputs. Our framework introduces configurable tournament parameters, dynamic agent selection, and multiple consensus fusion strategies. Experimental evaluations demonstrate significant improvements in response accuracy, coherence, and reliability compared to baseline single-model approaches. The ART framework provides a scalable, production-ready solution for applications requiring high-quality, vetted LLM responses, achieving an 8.4% improvement in overall quality metrics and R22 values exceeding 0.96 in ELO rating convergence.
[33] Sycophancy Claims about Language Models: The Missing Human-in-the-Loop
Jan Batzner, Volker Stocker, Stefan Schmid, Gjergji Kasneci
Main category: cs.CL
TL;DR: This paper reviews methodological challenges in measuring sycophancy in LLMs, identifies five core operationalizations, notes the lack of human perception evaluation, and provides recommendations for future research.
Details
Motivation: The motivation is to address the increasing claims about sycophantic response patterns in LLMs by systematically reviewing methodological challenges in measuring this phenomenon, as current research lacks evaluation of human perception despite sycophancy being inherently human-centric.Method: The method involves reviewing existing literature on LLM sycophancy, identifying five core operationalizations of sycophancy measurement, analyzing difficulties in distinguishing sycophantic responses from related AI alignment concepts, and conducting methodological analysis.
Result: The paper identifies five core operationalizations for measuring LLM sycophancy and highlights that current research fails to evaluate human perception, despite sycophancy being human-centric. It also reveals difficulties in distinguishing sycophantic responses from related AI alignment concepts.
Conclusion: The conclusion emphasizes the need for better methodological approaches to measure LLM sycophancy, particularly incorporating human perception evaluation, and provides actionable recommendations for future research in this area.
Abstract: Sycophantic response patterns in Large Language Models (LLMs) have been increasingly claimed in the literature. We review methodological challenges in measuring LLM sycophancy and identify five core operationalizations. Despite sycophancy being inherently human-centric, current research does not evaluate human perception. Our analysis highlights the difficulties in distinguishing sycophantic responses from related concepts in AI alignment and offers actionable recommendations for future research.
[34] Graphing the Truth: Structured Visualizations for Automated Hallucination Detection in LLMs
Tanmay Agrawal
Main category: cs.CL
TL;DR: A framework using interactive visual knowledge graphs to help users identify and correct LLM hallucinations by linking model assertions to underlying sources and showing confidence levels.
Details
Motivation: LLMs in enterprise settings often produce credible-looking hallucinations due to limited context windows and inconsistencies between pre-training data and domain knowledge. Current mitigation strategies (gold-standard Q&A curation or secondary model verification) are costly and don't provide deterministic assurance.Method: Organizes proprietary knowledge and model-generated content into interactive visual knowledge graphs that link model assertions to underlying sources of truth and indicate confidence levels. Provides a visual interface for users to diagnose inconsistencies, identify weak reasoning chains, and supply corrective feedback.
Result: Creates a human-in-the-loop workflow with a structured feedback loop that enhances model reliability and continuously improves response quality by enabling users to visually identify hallucination zones.
Conclusion: The visual knowledge graph framework offers a practical solution for mitigating LLM hallucinations in enterprise settings by providing transparent, interactive visualization of model reasoning and enabling structured human feedback to improve reliability.
Abstract: Large Language Models have rapidly advanced in their ability to interpret and generate natural language. In enterprise settings, they are frequently augmented with closed-source domain knowledge to deliver more contextually informed responses. However, operational constraints such as limited context windows and inconsistencies between pre-training data and supplied knowledge often lead to hallucinations, some of which appear highly credible and escape routine human review. Current mitigation strategies either depend on costly, large-scale gold-standard Q&A curation or rely on secondary model verification, neither of which offers deterministic assurance. This paper introduces a framework that organizes proprietary knowledge and model-generated content into interactive visual knowledge graphs. The objective is to provide end users with a clear, intuitive view of potential hallucination zones by linking model assertions to underlying sources of truth and indicating confidence levels. Through this visual interface, users can diagnose inconsistencies, identify weak reasoning chains, and supply corrective feedback. The resulting human-in-the-loop workflow creates a structured feedback loop that can enhance model reliability and continuously improve response quality.
[35] A Comparison of Human and ChatGPT Classification Performance on Complex Social Media Data
Breanna E. Green, Ashley L. Shea, Pengfei Zhao, Drew B. Margolin
Main category: cs.CL
TL;DR: GPT-4 struggles with classifying nuanced language in computational social science tasks, performing worse than human annotators despite prompt engineering improvements.
Details
Motivation: While ChatGPT is increasingly used in computational social science, there's limited understanding of its performance on complex tasks involving nuanced language classification and annotation.Method: Tested GPT-3.5, GPT-4, and GPT-4o on classification tasks with nuanced language. Used four different prompt styles, evaluated precision, recall, and F1 scores, and compared results to human annotators through both quantitative and qualitative analysis.
Result: GPT-4 has difficulty classifying nuanced language despite prompt engineering. Including label definitions in prompts helps but doesn’t solve the problem. Four specific qualitative findings emerged. Overall performance was inferior to human annotators.
Conclusion: ChatGPT should be used with caution for classification tasks involving nuanced language in computational social science, as current models struggle with such complexity despite rapid technological advancements.
Abstract: Generative artificial intelligence tools, like ChatGPT, are an increasingly utilized resource among computational social scientists. Nevertheless, there remains space for improved understanding of the performance of ChatGPT in complex tasks such as classifying and annotating datasets containing nuanced language. Method. In this paper, we measure the performance of GPT-4 on one such task and compare results to human annotators. We investigate ChatGPT versions 3.5, 4, and 4o to examine performance given rapid changes in technological advancement of large language models. We craft four prompt styles as input and evaluate precision, recall, and F1 scores. Both quantitative and qualitative evaluations of results demonstrate that while including label definitions in prompts may help performance, overall GPT-4 has difficulty classifying nuanced language. Qualitative analysis reveals four specific findings. Our results suggest the use of ChatGPT in classification tasks involving nuanced language should be conducted with prudence.
[36] FastPOS: Language-Agnostic Scalable POS Tagging Framework Low-Resource Use Case
Md Abdullah Al Kafi, Sumit Kumar Banshal
Main category: cs.CL
TL;DR: A language-agnostic transformer-based POS tagging framework for low-resource languages, demonstrated with Bangla and Hindi, achieving high accuracy with minimal code adaptation.
Details
Motivation: To address the challenge of developing NLP tools for low-resource languages by creating a portable, modular framework that reduces development overhead and allows researchers to focus on linguistic preprocessing and dataset refinement.Method: A transformer-based POS tagging framework designed for language-agnostic application, requiring only three lines of framework-specific code to adapt from Bangla to Hindi. The modular architecture allows easy replacement of transformer models with minimal code adjustments.
Result: Achieved 96.85% token-level accuracy for Bangla and 97% for Hindi across POS categories, with strong F1 scores despite dataset imbalance and linguistic overlap. Performance discrepancy in one POS category highlighted dataset curation challenges.
Conclusion: The framework successfully enables rapid cross-lingual adaptation for low-resource languages with minimal code changes, reducing model design overhead and allowing researchers to focus on essential linguistic preprocessing and dataset refinement for underrepresented languages.
Abstract: This study proposes a language-agnostic transformer-based POS tagging framework designed for low-resource languages, using Bangla and Hindi as case studies. With only three lines of framework-specific code, the model was adapted from Bangla to Hindi, demonstrating effective portability with minimal modification. The framework achieves 96.85 percent and 97 percent token-level accuracy across POS categories in Bangla and Hindi while sustaining strong F1 scores despite dataset imbalance and linguistic overlap. A performance discrepancy in a specific POS category underscores ongoing challenges in dataset curation. The strong results stem from the underlying transformer architecture, which can be replaced with limited code adjustments. Its modular and open-source design enables rapid cross-lingual adaptation while reducing model design and tuning overhead, allowing researchers to focus on linguistic preprocessing and dataset refinement, which are essential for advancing NLP in underrepresented languages.
[37] Auxiliary-Hyperparameter-Free Sampling: Entropy Equilibrium for Text Generation
Xiaodong Cai, Hai Lin, Shaoxiong Zhan, Weiqi Luo, Hong-Gee Kim, Hongyan Hao, Yu Yang, Hai-Tao Zheng
Main category: cs.CL
TL;DR: Entropy Equilibrium Sampling (EES) is a hyperparameter-free token sampling method that balances normalized entropy with probability mass to dynamically adjust candidate sets, improving text generation quality without requiring tuning.
Details
Motivation: Existing token sampling strategies introduce additional hyperparameters that require extensive tuning and complicate deployment of large language models.Method: EES is an auxiliary hyperparameter-free approach inspired by information theory that dynamically adjusts candidate sets by balancing normalized entropy with probability mass.
Result: EES consistently performs well across temperature settings on both reasoning and generation tasks, delivering competitive accuracy and coherence while maintaining diversity across various model architectures.
Conclusion: By eliminating the need for hyperparameter tuning, EES greatly simplifies deployment while improving performance in text generation tasks.
Abstract: Token sampling strategies critically influence text generation quality in large language models (LLMs). However, existing methods introduce additional hyperparameters, requiring extensive tuning and complicating deployment. We present Entropy Equilibrium Sampling (EES), an auxiliary hyperparameter-free approach inspired by information theory that can dynamically adjust candidate sets by balancing normalized entropy with probability mass. We evaluate EES on both reasoning and generation tasks across a range of model architectures. Our results show that EES consistently performs well across temperature settings, delivering competitive accuracy and coherence while maintaining diversity. By eliminating the need for hyperparameter tuning, EES greatly simplifies deployment while improving performance. Code is available at https://github.com/shuanncai/EES
[38] Accelerating Bangla NLP Tasks with Automatic Mixed Precision: Resource-Efficient Training Preserving Model Efficacy
Md Mehrab Hossain Opi, Sumaiya Khan, Moshammad Farzana Rahman
Main category: cs.CL
TL;DR: AMP training reduces computational costs for Bangla NLP by 44.5% faster training and 17.6% less memory while maintaining 99.7% of full-precision performance.
Details
Motivation: NLP training requires substantial computational resources, posing challenges for Bangla NLP development where access to high-end hardware is limited. There's a need to improve computational efficiency without sacrificing model performance.Method: Used automatic mixed precision (AMP) training with dynamic mix of 16-bit and 32-bit floating-point computations. Evaluated across four Bangla NLP tasks (sentiment analysis, NER, error classification, QA) using four transformer models (BanglaBERT, BanglishBERT, XLM-R, mBERT).
Result: AMP accelerates training by 44.5% and reduces memory consumption by 17.6%, while maintaining F-1 score within 99.7% of full-precision baselines across all tasks and models.
Conclusion: AMP has significant potential to democratize access to state-of-the-art NLP capabilities in hardware-constrained settings by lowering computational barriers for Bangla NLP development.
Abstract: Training models for Natural Language Processing (NLP) requires substantial computational resources and time, posing significant challenges, especially for NLP development in Bangla, where access to high-end hardware is often limited. In this work, we explore automatic mixed precision (AMP) training as a means to improve computational efficiency without sacrificing model performance. By leveraging a dynamic mix of 16-bit and 32-bit floating-point computations, AMP lowers GPU memory requirements and speeds up training without degrading model performance. We evaluate AMP across four standard Bangla NLP tasks, namely sentiment analysis, named entity recognition, error classification, and question answering, using four transformer-based models: BanglaBERT, BanglishBERT, XLM-R, and mBERT. Our results demonstrate that AMP accelerates training by 44.5% and reduces memory consumption by 17.6%, while maintaining F-1 score within 99.7% of the full-precision baselines. This empirical study highlights AMP’s potential to democratize access to state-of-the-art NLP capabilities in hardware-constrained settings by lowering computational barriers.
[39] WaterSearch: A Quality-Aware Search-based Watermarking Framework for Large Language Models
Yukang Lin, Jiahao Shao, Shuoran Jiang, Wentao Zhu, Bingjie Lu, Xiangping Wu, Joanna Siebert, Qingcai Chen
Main category: cs.CL
TL;DR: WaterSearch: A sentence-level search-based watermarking framework that improves text quality while maintaining strong watermark detectability by optimizing distribution fidelity and signal characteristics.
Details
Motivation: Existing LLM watermarking methods face a trade-off between detectability and text quality - stronger watermark signals degrade downstream task performance. There's a need for watermarking that maintains both high text quality and robust detectability.Method: Proposes WaterSearch, a sentence-level search-based framework that controls seed pools for diverse parallel generation and jointly optimizes distribution fidelity and watermark signal characteristics. Includes a sentence-level detection method with attack robustness.
Result: Achieves 51.01% average performance improvement over SOTA baselines at 95% detectability strength. Gains of 47.78% for short text and 36.47% for low-entropy outputs. Maintains high detectability under insertion, synonym substitution, and paraphrase attacks.
Conclusion: WaterSearch effectively addresses the detectability-text quality trade-off in LLM watermarking, providing robust watermarking with minimal performance degradation across diverse tasks and attack scenarios.
Abstract: Watermarking acts as a critical safeguard in text generated by Large Language Models (LLMs). By embedding identifiable signals into model outputs, watermarking enables reliable attribution and enhances the security of machine-generated content. Existing approaches typically embed signals by manipulating token generation probabilities. Despite their effectiveness, these methods inherently face a trade-off between detectability and text quality: the signal strength and randomness required for robust watermarking tend to degrade the performance of downstream tasks. In this paper, we design a novel embedding scheme that controls seed pools to facilitate diverse parallel generation of watermarked text. Based on that scheme, we propose WaterSearch, a sentence-level, search-based watermarking framework adaptable to a wide range of existing methods. WaterSearch enhances text quality by jointly optimizing two key aspects: 1) distribution fidelity and 2) watermark signal characteristics. Furthermore, WaterSearch is complemented by a sentence-level detection method with strong attack robustness. We evaluate our method on three popular LLMs across ten diverse tasks. Extensive experiments demonstrate that our method achieves an average performance improvement of 51.01% over state-of-the-art baselines at a watermark detectability strength of 95%. In challenging scenarios such as short text generation and low-entropy output generation, our method yields performance gains of 47.78% and 36.47%, respectively. Moreover, under different attack senarios including insertion, synonym substitution and paraphrase attasks, WaterSearch maintains high detectability, further validating its robust anti-attack capabilities. Our code is available at \href{https://github.com/Yukang-Lin/WaterSearch}{https://github.com/Yukang-Lin/WaterSearch}.
[40] Less is More: Resource-Efficient Low-Rank Adaptation
Chunlin Tian, Xuyang Wei, Huanrong Liu, Zhijiang Guo, Li Li
Main category: cs.CL
TL;DR: EffiLoRA improves LoRA by reducing parameter redundancy through shared A matrices across layers and selective B matrix updates, achieving better efficiency and performance across language, multimodal, and diffusion models.
Details
Motivation: LoRA has parameter overhead and interference issues in complex datasets, and existing decoupling approaches still have high training costs. The paper aims to address inter-matrix and intra-layer parameter redundancy in LoRA.Method: EffiLoRA uses a unified A matrix across all transformer layers and introduces runtime selective B matrices update that dynamically trades off resource budget and model performance.
Result: EffiLoRA consistently outperforms LoRA across diverse modalities including commonsense reasoning, visual instruction tuning, and image generation, demonstrating improved efficiency and robustness.
Conclusion: EffiLoRA provides a lightweight and generalizable approach for efficient fine-tuning that works across language, multimodal, and diffusion models while maintaining or improving performance.
Abstract: Low-Rank Adaptation (LoRA) is a widely adopted parameter-efficient fine-tuning (PEFT) method for Large Language Models (LLMs), but it still incurs notable overhead and suffers from parameter interference in complex datasets. While re- cent works decouple LoRA update matrices to exploit matrix-wise asymmetry, training costs remain high. We revisit LoRA from the perspective of inter-matrix and intra-layer parameter redundancy and propose Resource-Efficient Low-Rank Adaptation, EffiLoRA, a lightweight and generalizable approach for language, multimodal, and diffusion models. EffiLoRA employs a unified A matrix across all transformer layers and introduces a runtime selective B matrices up- date to dynamically trade-off the system resource budget and model performance. EffiLoRA consistently outperforms LoRA across diverse modalities, including commonsense reasoning, visual instruction tuning, and image generation, demon- strating improved efficiency and robustness.
[41] Reward Auditor: Inference on Reward Modeling Suitability in Real-World Perturbed Scenarios
Jianxiang Zang, Yongda Wei, Ruxue Bai, Shiyu Jiang, Nijia Mo, Binhong Li, Qiang Sun, Hui Liu
Main category: cs.CL
TL;DR: The paper introduces Reward Auditor, a hypothesis-testing framework for evaluating reward model suitability - the conditional reliability under real-world perturbations - rather than just preference accuracy in static scenarios.
Details
Motivation: Current reward model evaluation methods only measure preference perception accuracy in given scenarios, missing critical vulnerabilities that emerge in real-world conditions with perturbations. There's a need to assess whether RMs exhibit systematic vulnerabilities in specific real-world scenarios.Method: Reward Auditor uses a hypothesis-testing framework to infer RM suitability. It quantifies statistical significance and effect size by auditing distribution degradation of RM preference perception confidence under real-world perturbed scenarios, enabling systematic vulnerability assessment.
Result: The framework enables inference of both the certainty and severity of RM vulnerabilities across diverse real-world scenarios, moving beyond simple accuracy metrics to systematic suitability assessment.
Conclusion: Reward Auditor lays a foundation for building next-generation LLM alignment systems that are verifiably safe, more robust, and trustworthy by providing a scientific auditing approach to assess reward model suitability in real-world conditions.
Abstract: Reliable reward models (RMs) are critical for ensuring the safe alignment of large language models (LLMs). However, current evaluation methods focus solely on preference perception accuracies in given specific scenarios, obscuring the critical vulnerabilities of RMs in real-world scenarios. We identify the true challenge lies in assessing a novel dimension: Suitability, defined as conditional reliability under specific real-world perturbations. To this end, we introduce Reward Auditor, a hypothesis-testing framework specifically designed for RM suitability inference. Rather than answering “How accurate is the RM’s preference perception for given samples?”, it employs scientific auditing to answer: “Can we infer RMs exhibit systematic vulnerabilities in specific real-world scenarios?”. Under real-world perturbed scenarios, Reward Auditor quantifies statistical significance and effect size by auditing distribution degradation of RM preference perception confidence. This enables inference of both the certainty and severity of RM vulnerabilities across diverse real-world scenarios. This lays a solid foundation for building next-generation LLM alignment systems that are verifiably safe, more robust, and trustworthy.
[42] Mitigating Hallucinations in Zero-Shot Scientific Summarisation: A Pilot Study
Imane Jaaouine, Ross D. King
Main category: cs.CL
TL;DR: Prompt engineering methods, particularly context repetition and random addition, can reduce context inconsistency hallucinations in zero-shot LLM summarization of scientific texts by improving lexical alignment between generated summaries and source abstracts.
Details
Motivation: LLMs produce context inconsistency hallucinations where generated outputs misalign with user prompts, especially in zero-shot scientific summarization tasks where models rely solely on pre-training data without task-specific fine-tuning.Method: Tested 7 prompt engineering methods on 8 yeast biotechnology abstracts using 6 instruction-tuned LLMs: baseline prompt, two instruction complexity levels (PE-1, PE-2), two context repetition levels (CR-K1, CR-K2), and two random addition levels (RA-K1, RA-K2). Evaluated 336 summaries using 6 metrics (ROUGE-1, ROUGE-2, ROUGE-L, BERTScore, METEOR, cosine similarity) and performed statistical analysis on 3744 datapoints with BCa bootstrap confidence intervals and Wilcoxon signed-rank tests.
Result: Context repetition (CR) and random addition (RA) significantly improved lexical alignment of LLM-generated summaries with abstracts, indicating these prompt engineering methods can mitigate context inconsistency hallucinations.
Conclusion: Prompt engineering has potential to impact hallucinations in zero-shot scientific summarization tasks, with context repetition and random addition showing particular effectiveness in improving alignment between LLM outputs and source texts.
Abstract: Large language models (LLMs) produce context inconsistency hallucinations, which are LLM generated outputs that are misaligned with the user prompt. This research project investigates whether prompt engineering (PE) methods can mitigate context inconsistency hallucinations in zero-shot LLM summarisation of scientific texts, where zero-shot indicates that the LLM relies purely on its pre-training data. Across eight yeast biotechnology research paper abstracts, six instruction-tuned LLMs were prompted with seven methods: a base- line prompt, two levels of increasing instruction complexity (PE-1 and PE-2), two levels of context repetition (CR-K1 and CR-K2), and two levels of random addition (RA-K1 and RA-K2). Context repetition involved the identification and repetition of K key sentences from the abstract, whereas random addition involved the repetition of K randomly selected sentences from the abstract, where K is 1 or 2. A total of 336 LLM-generated summaries were evaluated using six metrics: ROUGE-1, ROUGE-2, ROUGE-L, BERTScore, METEOR, and cosine similarity, which were used to compute the lexical and semantic alignment be- tween the summaries and the abstracts. Four hypotheses on the effects of prompt methods on summary alignment with the reference text were tested. Statistical analysis on 3744 collected datapoints was performed using bias-corrected and accelerated (BCa) bootstrap confidence intervals and Wilcoxon signed-rank tests with Bonferroni-Holm correction. The results demonstrated that CR and RA significantly improve the lexical alignment of LLM-generated summaries with the abstracts. These findings indicate that prompt engineering has the potential to impact hallucinations in zero-shot scientific summarisation tasks.
[43] DeformAr: Rethinking NER Evaluation through Component Analysis and Visual Analytics
Ahmed Mustafa Younes
Main category: cs.CL
TL;DR: DeformAr is a framework for debugging Arabic NER systems by analyzing performance gaps between Arabic and English through cross-component and behavioral analysis.
Details
Motivation: Transformer models perform well in English but struggle with Arabic NER due to tokenization issues, dataset quality problems, and annotation inconsistencies. Existing studies analyze these factors in isolation rather than examining their joint effects.Method: DeformAr framework with data extraction library and interactive dashboard. Two evaluation modes: 1) Cross-component analysis dividing language into dataset/model components, 2) Behavioral analysis combining interpretability techniques, token-level metrics, visualizations, and representation space analysis.
Result: First Arabic-specific, component-based interpretability tool that enables component-aware diagnostics to detect model behaviors and explain them through underlying representational patterns and data factors.
Conclusion: DeformAr provides a crucial resource for advancing model analysis in under-resourced languages by systematically investigating performance discrepancies between Arabic and English NER systems.
Abstract: Transformer models have significantly advanced Natural Language Processing (NLP), demonstrating strong performance in English. However, their effectiveness in Arabic, particularly for Named Entity Recognition (NER), remains limited, even with larger pre-trained models. This performance gap stems from multiple factors, including tokenisation, dataset quality, and annotation inconsistencies. Existing studies often analyze these issues in isolation, failing to capture their joint effect on system behaviour and performance. We introduce DeformAr (Debugging and Evaluation Framework for Transformer-based NER Systems), a novel framework designed to investigate and explain the performance discrepancy between Arabic and English NER systems. DeformAr integrates a data extraction library and an interactive dashboard, supporting two modes of evaluation: cross-component analysis and behavioural analysis. The framework divides each language into dataset and model components to examine their interactions. The analysis proceeds in two stages. First, cross-component analysis provides systematic diagnostic measures across data and model subcomponents, addressing the “what,” “how,” and “why” behind observed discrepancies. The second stage applies behavioural analysis by combining interpretability techniques with token-level metrics, interactive visualisations, and representation space analysis. These stages enable a component-aware diagnostic process that detects model behaviours and explains them by linking them to underlying representational patterns and data factors. DeformAr is the first Arabic-specific, component-based interpretability tool, offering a crucial resource for advancing model analysis in under-resourced languages.
[44] Fine-tuning of lightweight large language models for sentiment classification on heterogeneous financial textual data
Alvaro Paredes Amorin, Andre Python, Christoph Weisser
Main category: cs.CL
TL;DR: Lightweight open-source LLMs (Qwen3 8B, Llama3 8B) achieve competitive financial sentiment analysis performance with minimal training data, offering cost-effective alternatives to resource-intensive models.
Details
Motivation: Proprietary LLMs require extensive computational resources and proprietary datasets that are inaccessible to many researchers. The paper investigates whether lightweight open-source LLMs can provide viable alternatives for financial sentiment analysis across diverse datasets.Method: Compared FinBERT benchmark with three lightweight open-source LLMs (DeepSeek-LLM 7B, Llama3 8B Instruct, Qwen3 8B) on five public financial datasets spanning different sizes, sources, formats, and languages. Evaluated performance with limited training data (as low as 5%) in zero-shot and few-shot learning scenarios.
Result: Lightweight LLMs, particularly Qwen3 8B and Llama3 8B, performed best in most scenarios, achieving competitive results even with only 5% of training data. Performance held across zero-shot and few-shot learning settings.
Conclusion: Lightweight open-source LLMs offer cost-effective alternatives for financial sentiment analysis, achieving competitive performance with minimal training data across heterogeneous textual sources, making them accessible to researchers with limited resources.
Abstract: Large language models (LLMs) play an increasingly important role in finan- cial markets analysis by capturing signals from complex and heterogeneous textual data sources, such as tweets, news articles, reports, and microblogs. However, their performance is dependent on large computational resources and proprietary datasets, which are costly, restricted, and therefore inacces- sible to many researchers and practitioners. To reflect realistic situations we investigate the ability of lightweight open-source LLMs - smaller and publicly available models designed to operate with limited computational resources - to generalize sentiment understanding from financial datasets of varying sizes, sources, formats, and languages. We compare the benchmark finance natural language processing (NLP) model, FinBERT, and three open-source lightweight LLMs, DeepSeek-LLM 7B, Llama3 8B Instruct, and Qwen3 8B on five publicly available datasets: FinancialPhraseBank, Financial Question Answering, Gold News Sentiment, Twitter Sentiment and Chinese Finance Sentiment. We find that LLMs, specially Qwen3 8B and Llama3 8B, perform best in most scenarios, even from using only 5% of the available training data. These results hold in zero-shot and few-shot learning scenarios. Our findings indicate that lightweight, open-source large language models (LLMs) consti- tute a cost-effective option, as they can achieve competitive performance on heterogeneous textual data even when trained on only a limited subset of the extensive annotated corpora that are typically deemed necessary.
[45] Table as a Modality for Large Language Models
Liyao Li, Chao Ye, Wentao Ye, Yifei Sun, Zhe Jiang, Haobo Wang, Jiaming Tian, Yiming Zhang, Ningtao Wang, Xing Fu, Gang Chen, Junbo Zhao
Main category: cs.CL
TL;DR: TAMO is a multimodal framework that treats tables as an independent modality integrated with text tokens, using a hypergraph neural network as global table encoder with LLMs, achieving 42.65% average relative gain on table reasoning tasks.
Details
Motivation: Current LLMs struggle with tabular data reasoning because they serialize tables, losing structural information. Even advanced LLMs like GPTs fall short on table reasoning tasks, as shown by probing experiments on the StructQA benchmark.Method: Proposes TAMO (Tables As an independent MOdality) - a multimodal framework with a hypergraph neural network as global table encoder integrated with mainstream LLMs. Treats tables as independent modality alongside text tokens.
Result: Significant improvements on generalization across multiple benchmarking datasets (HiTab, WikiTQ, WikiSQL, FeTaQA, StructQA) with average relative gain of 42.65%.
Conclusion: Treating tables as an independent modality with structural preservation through hypergraph neural networks integrated with LLMs effectively addresses the limitations of current table serialization approaches, leading to substantial performance gains in table reasoning tasks.
Abstract: To migrate the remarkable successes of Large Language Models (LLMs), the community has made numerous efforts to generalize them to the table reasoning tasks for the widely deployed tabular data. Despite that, in this work, by showing a probing experiment on our proposed StructQA benchmark, we postulate that even the most advanced LLMs (such as GPTs) may still fall short of coping with tabular data. More specifically, the current scheme often simply relies on serializing the tabular data, together with the meta information, then inputting them through the LLMs. We argue that the loss of structural information is the root of this shortcoming. In this work, we further propose TAMO, which bears an ideology to treat the tables as an independent modality integrated with the text tokens. The resulting model in TAMO is a multimodal framework consisting of a hypergraph neural network as the global table encoder seamlessly integrated with the mainstream LLM. Empirical results on various benchmarking datasets, including HiTab, WikiTQ, WikiSQL, FeTaQA, and StructQA, have demonstrated significant improvements on generalization with an average relative gain of 42.65%.
[46] Dr.Mi-Bench: A Modular-integrated Benchmark for Scientific Deep Research Agent
Zhihan Guo, Feiyang Xu, Yifan Li, Muzhi Li, Shuai Zou, Jiele Wu, Han Shi, Haoli Bai, Ho-fung Leung, Irwin King
Main category: cs.CL
TL;DR: Dr.Mi-Bench is a new benchmark for evaluating scientific deep research agents, addressing gaps in existing benchmarks by focusing on planning, retrieval, and reasoning across 10 scientific domains using 200 human-annotated instances from academic literature.
Details
Motivation: Existing benchmarks for deep research agents have two main limitations: 1) they focus narrowly on retrieval while neglecting high-level planning and reasoning, and 2) they favor general domains over scientific domains which are core applications for DR agents. There's a need for better evaluation tools for automated academic research assistants.Method: The authors introduce Dr.Mi-Bench, a modular-integrated benchmark grounded in academic literature with 200 human-annotated instances across 10 scientific domains (including both research and review papers). They also propose Dr.Mi-Eval, a novel evaluation paradigm that leverages academic paper structure to assess planning, retrieval, and reasoning through two modes: end-to-end evaluation for DR agents and isolated evaluation for foundational LLMs as potential backbones.
Result: Experimental results show a fragmented performance landscape: agents exhibit specialized strengths but share critical weaknesses, particularly in multi-source retrieval for review-style tasks and performing consistently across diverse scientific fields. Improving high-level planning capability is crucial for unlocking the reasoning potential of foundational LLMs as backbones.
Conclusion: Dr.Mi-Bench provides a diagnostic tool to guide development of more reliable academic research assistants by exposing actionable failure modes. The benchmark addresses critical gaps in evaluating deep research agents and reveals that planning capability is key to unlocking LLM reasoning potential for scientific research tasks.
Abstract: The explosive growth in academic literature necessitates automated deep research (DR) agents, yet their evaluation remains a significant challenge. First, existing benchmarks often focus narrowly on retrieval while neglecting high-level planning and reasoning. Second, existing benchmarks favor general domains over the scientific domains that are the core application for DR agents. To address these gaps, we introduce Dr.Mi-Bench, a Modular-integrated benchmark for scientific DR agents. Grounded in academic literature, our benchmark uses a human-annotated dataset of 200 instances across 10 scientific domains, including both research and review papers. Besides, we also propose a Modular-integrated Evaluation Paradigm for DR Agents (Dr.Mi-Eval), a novel modular-integrated evaluation paradigm, which leverages the rich structure of academic papers to assess the core competencies of planning, retrieval, and reasoning through two complementary modes: an end-to-end evaluation for DR agents and an isolated evaluation for foundational LLMs as potential backbones. Experimental results reveal a fragmented performance landscape: agents exhibit specialized strengths but share critical weaknesses, most notably in performing the multi-source retrieval required for review-style tasks and performing consistently across diverse scientific fields. Moreover, improving high-level planning capability is the crucial factor for unlocking the reasoning potential of foundational LLMs as backbones. By exposing these actionable failure modes, Dr.Mi-Bench provides a diagnostic tool to guide the development of more reliable academic research assistants.
[47] Advancing Academic Chatbots: Evaluation of Non Traditional Outputs
Nicole Favero, Francesca Salute, Daniel Hardt
Main category: cs.CL
TL;DR: This paper evaluates LLMs on non-traditional academic outputs (slide decks, podcast scripts) and compares Graph RAG vs Advanced RAG retrieval strategies for QA tasks.
Details
Motivation: Most LLM evaluations focus on standard tasks like factual QA or short summarization, but this research expands scope to non-traditional academic outputs and compares different retrieval strategies for QA.Method: Implemented prototype combining Meta’s LLaMA 3 70B and OpenAI’s GPT 4o mini. Compared Graph RAG (structured knowledge-graph based) vs Advanced RAG (hybrid keyword-semantic search) for QA. Evaluated slide deck and podcast script generation with document-grounded retrieval. Used both human ratings across 11 quality dimensions and LLM judges for cross-validation.
Result: GPT 4o mini with Advanced RAG produced most accurate QA responses. Graph RAG offered limited improvements and caused more hallucinations due to structural complexity. For slide/podcast generation, GPT 4o mini performed best, though LLaMA 3 showed promise in narrative coherence. Human reviewers were crucial for detecting layout/stylistic flaws.
Conclusion: Combined human-LLM evaluation is essential for assessing emerging academic outputs. Advanced RAG outperformed Graph RAG for QA, and GPT 4o mini generally performed better than LLaMA 3, though each had strengths in different areas.
Abstract: Most evaluations of large language models focus on standard tasks such as factual question answering or short summarization. This research expands that scope in two directions: first, by comparing two retrieval strategies, Graph RAG, structured knowledge-graph based, and Advanced RAG, hybrid keyword-semantic search, for QA; and second, by evaluating whether LLMs can generate high quality non-traditional academic outputs, specifically slide decks and podcast scripts. We implemented a prototype combining Meta’s LLaMA 3 70B open weight and OpenAI’s GPT 4o mini API based. QA performance was evaluated using both human ratings across eleven quality dimensions and large language model judges for scalable cross validation. GPT 4o mini with Advanced RAG produced the most accurate responses. Graph RAG offered limited improvements and led to more hallucinations, partly due to its structural complexity and manual setup. Slide and podcast generation was tested with document grounded retrieval. GPT 4o mini again performed best, though LLaMA 3 showed promise in narrative coherence. Human reviewers were crucial for detecting layout and stylistic flaws, highlighting the need for combined human LLM evaluation in assessing emerging academic outputs.
[48] When Safety Blocks Sense: Measuring Semantic Confusion in LLM Refusals
Riad Ahmed Anonto, Md Labid Al Nahiyan, Md Tanvir Hassan, Ch. Md. Rakin Haider
Main category: cs.CL
TL;DR: Paper introduces “semantic confusion” to measure local inconsistency in safety-aligned LLMs where models accept one phrasing but reject close paraphrases, proposes ParaGuard dataset and token-level metrics to diagnose this issue.
Details
Motivation: Current safety evaluation metrics (false rejection, compliance) treat each prompt in isolation and miss local inconsistency where models accept one phrasing but reject semantically equivalent paraphrases, limiting diagnosis and tuning capabilities.Method: 1) Introduce “semantic confusion” failure mode; 2) Build ParaGuard - 10k-prompt corpus with controlled paraphrase clusters holding intent fixed while varying surface form; 3) Propose three token-level metrics: Confusion Index, Confusion Rate, and Confusion Depth using token embeddings, next-token probabilities, and perplexity signals.
Result: Experiments show global false-rejection rate hides critical structure: reveals globally unstable boundaries in some settings, localized pockets of inconsistency in others, and cases where stricter refusal doesn’t increase inconsistency. Confusion-aware auditing separates refusal frequency from refusal sensibility.
Conclusion: Provides practical signal for developers to reduce false refusals while preserving safety by measuring semantic confusion, enabling better diagnosis and tuning of safety-aligned language models beyond global metrics.
Abstract: Safety-aligned language models often refuse prompts that are actually harmless. Current evaluations mostly report global rates such as false rejection or compliance. These scores treat each prompt alone and miss local inconsistency, where a model accepts one phrasing of an intent but rejects a close paraphrase. This gap limits diagnosis and tuning. We introduce “semantic confusion,” a failure mode that captures such local inconsistency, and a framework to measure it. We build ParaGuard, a 10k-prompt corpus of controlled paraphrase clusters that hold intent fixed while varying surface form. We then propose three model-agnostic metrics at the token level: Confusion Index, Confusion Rate, and Confusion Depth. These metrics compare each refusal to its nearest accepted neighbors and use token embeddings, next-token probabilities, and perplexity signals. Experiments across diverse model families and deployment guards show that global false-rejection rate hides critical structure. Our metrics reveal globally unstable boundaries in some settings, localized pockets of inconsistency in others, and cases where stricter refusal does not increase inconsistency. We also show how confusion-aware auditing separates how often a system refuses from how sensibly it refuses. This gives developers a practical signal to reduce false refusals while preserving safety.
[49] ELR-1000: A Community-Generated Dataset for Endangered Indic Indigenous Languages
Neha Joshi, Pamir Gogoi, Aasim Mirza, Aayush Jansari, Aditya Yadavalli, Ayushi Pandey, Arunima Shukla, Deepthi Sudharsan, Kalika Bali, Vivek Seshadri
Main category: cs.CL
TL;DR: Researchers created ELR-1000, a multimodal dataset of 1,060 traditional recipes from 10 endangered languages in Eastern India, collected via mobile interface for low-digital-literacy users. They found LLMs struggle with low-resource cultural translation but improve with targeted context.
Details
Motivation: To address the lack of benchmarks for underrepresented languages and domains, and to advance equitable and culturally-aware language technologies by focusing on endangered languages and their embedded cultural knowledge.Method: Created ELR-1000 dataset through mobile interface crowdsourcing from rural communities, then evaluated multiple state-of-the-art LLMs on translating recipes with and without targeted cultural context (background info, examples, preservation guidelines).
Result: LLMs struggle with low-resource, culturally-specific language translation, but providing targeted context significantly improves translation quality. The dataset reveals gaps in current language technology for underrepresented languages.
Conclusion: Need for benchmarks catering to underrepresented languages and domains to advance equitable language technologies. The released ELR-1000 dataset aims to motivate development of technologies for endangered languages and cultural preservation.
Abstract: We present a culturally-grounded multimodal dataset of 1,060 traditional recipes crowdsourced from rural communities across remote regions of Eastern India, spanning 10 endangered languages. These recipes, rich in linguistic and cultural nuance, were collected using a mobile interface designed for contributors with low digital literacy. Endangered Language Recipes (ELR)-1000 – captures not only culinary practices but also the socio-cultural context embedded in indigenous food traditions. We evaluate the performance of several state-of-the-art large language models (LLMs) on translating these recipes into English and find the following: despite the models’ capabilities, they struggle with low-resource, culturally-specific language. However, we observe that providing targeted context – including background information about the languages, translation examples, and guidelines for cultural preservation – leads to significant improvements in translation quality. Our results underscore the need for benchmarks that cater to underrepresented languages and domains to advance equitable and culturally-aware language technologies. As part of this work, we release the ELR-1000 dataset to the NLP community, hoping it motivates the development of language technologies for endangered languages.
[50] How do we measure privacy in text? A survey of text anonymization metrics
Yaxuan Ren, Krithika Ramesh, Yaxing Yao, Anjalie Field
Main category: cs.CL
TL;DR: Systematic survey of privacy evaluation metrics for text anonymization, analyzing 47 papers to identify six privacy notions and assess their alignment with legal standards and user expectations.
Details
Motivation: Text anonymization is crucial for NLP research with sensitive data, but evaluating whether anonymization methods sufficiently protect privacy remains challenging. There's a need to clarify and reconcile different privacy evaluation metrics used in the field.Method: Conducted a systematic survey of 47 papers reporting privacy metrics, identified and compared six distinct privacy notions, analyzed how associated metrics capture different aspects of privacy risk, and assessed alignment with legal standards (HIPAA and GDPR) and user-centered expectations from HCI studies.
Result: Identified six distinct privacy notions used in text anonymization evaluation, analyzed their strengths and limitations in capturing privacy risk, and assessed how well these notions align with legal standards and user expectations. The analysis reveals gaps in current evaluation practices.
Conclusion: The survey provides practical guidance for navigating privacy evaluation approaches and highlights the need for more robust, comparable, and legally aware privacy evaluations in text anonymization to facilitate better privacy protection in NLP research.
Abstract: In this work, we aim to clarify and reconcile metrics for evaluating privacy protection in text through a systematic survey. Although text anonymization is essential for enabling NLP research and model development in domains with sensitive data, evaluating whether anonymization methods sufficiently protect privacy remains an open challenge. In manually reviewing 47 papers that report privacy metrics, we identify and compare six distinct privacy notions, and analyze how the associated metrics capture different aspects of privacy risk. We then assess how well these notions align with legal privacy standards (HIPAA and GDPR), as well as user-centered expectations grounded in HCI studies. Our analysis offers practical guidance on navigating the landscape of privacy evaluation approaches further and highlights gaps in current practices. Ultimately, we aim to facilitate more robust, comparable, and legally aware privacy evaluations in text anonymization.
[51] DrawingBench: Evaluating Spatial Reasoning and UI Interaction Capabilities of Large Language Models through Mouse-Based Drawing Tasks
Hyunjun Kim, Sooyoung Ryu
Main category: cs.CL
TL;DR: DrawingBench is a transparent verification framework for evaluating trustworthiness of agentic LLMs through spatial reasoning tasks, featuring objective scoring, action-level auditability, and external oversight mechanisms.
Details
Motivation: Existing benchmarks lack transparency and auditability needed to assess whether autonomous AI agents behave reliably, making it difficult to establish trust in agentic systems operating independently.Method: Framework uses spatial reasoning tasks requiring low-level GUI action sequences, with 250 diverse prompts across 20 categories and 4 difficulty levels. Features 8 objective criteria for reproducible scoring, action-level inspection, and multi-turn feedback for external oversight.
Result: Evaluated 4 state-of-the-art LLMs across 1,000 tests: models achieved 92.8% perfect performance with structured external feedback (average +3.2% improvement, up to +32.8% for complex scenes). Specification clarity proved more important than task complexity - 100% perfect performance with explicit, verifiable criteria.
Conclusion: Transparent evaluation frameworks can establish trust in agentic systems, with external oversight proving more reliable than self-correction. The open-source framework provides a template for trustworthy agent assessment.
Abstract: As agentic AI systems increasingly operate autonomously, establishing trust through verifiable evaluation becomes critical. Yet existing benchmarks lack the transparency and auditability needed to assess whether agents behave reliably. We present DrawingBench, a verification framework for evaluating the trustworthiness of agentic LLMs through spatial reasoning tasks that require generating sequences of low-level GUI actions. Unlike opaque evaluations, DrawingBench provides transparent, rule-based assessment: 8 objective criteria enable reproducible scoring, while action-level inspection allows stakeholders to audit agent behavior. Our framework comprises 250 diverse prompts across 20 categories and 4 difficulty levels, deterministic evaluation metrics, and an external oversight mechanism through multi-turn feedback that enables human control over agent refinement. Evaluating four state-of-the-art LLMs (Claude-4 Sonnet, GPT-4.1, GPT-4.1-mini, Gemini-2.5 Flash) across 1,000 tests, we establish both capabilities and limitations: models achieved 92.8% perfect performance with structured external feedback driving significant improvements (average +3.2%, up to +32.8% for complex scenes), but systematic error patterns emerged in tool state management and long-horizon planning. Notably, specification clarity proved more important than task complexity – models achieved 100% perfect performance when given explicit, verifiable criteria. These findings demonstrate that transparent evaluation frameworks can establish trust in agentic systems, with external oversight proving more reliable than self-correction for guiding agent behavior. Our open-source framework provides a template for trustworthy agent assessment. Code and data: https://github.com/hyunjun1121/DrawingBench
[52] TempPerturb-Eval: On the Joint Effects of Internal Temperature and External Perturbations in RAG Robustness
Yongxin Zhou, Philippe Mulhem, Didier Schwab
Main category: cs.CL
TL;DR: This paper investigates how text perturbations (simulating noisy retrieval) interact with temperature settings in RAG systems, showing that high temperatures amplify vulnerability to perturbations and proposing a framework for analyzing these interactions.
Details
Motivation: Current RAG evaluation typically examines retrieval quality and generation parameters like temperature in isolation, overlooking their interaction effects. There's a need to understand how noisy retrieval interacts with temperature settings across multiple LLM runs.Method: Proposes a comprehensive RAG Perturbation-Temperature Analysis Framework that subjects retrieved documents to three distinct perturbation types across varying temperature settings. Conducts extensive experiments on HotpotQA dataset using both open-source and proprietary LLMs.
Result: Performance degradation follows distinct patterns: high-temperature settings consistently amplify vulnerability to perturbations, while certain perturbation types exhibit non-linear sensitivity across the temperature range.
Conclusion: Provides three key contributions: (1) diagnostic benchmark for assessing RAG robustness, (2) analytical framework for quantifying perturbation-temperature interactions, and (3) practical guidelines for model selection and parameter tuning under noisy retrieval conditions.
Abstract: The evaluation of Retrieval-Augmented Generation (RAG) systems typically examines retrieval quality and generation parameters like temperature in isolation, overlooking their interaction. This work presents a systematic investigation of how text perturbations (simulating noisy retrieval) interact with temperature settings across multiple LLM runs. We propose a comprehensive RAG Perturbation-Temperature Analysis Framework that subjects retrieved documents to three distinct perturbation types across varying temperature settings. Through extensive experiments on HotpotQA with both open-source and proprietary LLMs, we demonstrate that performance degradation follows distinct patterns: high-temperature settings consistently amplify vulnerability to perturbations, while certain perturbation types exhibit non-linear sensitivity across the temperature range. Our work yields three key contributions: (1) a diagnostic benchmark for assessing RAG robustness, (2) an analytical framework for quantifying perturbation-temperature interactions, and (3) practical guidelines for model selection and parameter tuning under noisy retrieval conditions.
[53] Generalist Large Language Models Outperform Clinical Tools on Medical Benchmarks
Krithik Vishwanath, Mrigayu Ghosh, Anton Alyakin, Daniel Alexander Alber, Yindalon Aphinyanaphongs, Eric Karl Oermann
Main category: cs.CL
TL;DR: Clinical AI assistants marketed for medical use often underperform compared to general-purpose frontier LLMs like GPT-5, despite being framed as safer alternatives.
Details
Motivation: Specialized clinical AI tools are being deployed in medical practice with claims of superior safety and reliability compared to general-purpose LLMs, but they lack independent quantitative evaluation, creating an evidence gap despite their growing influence on healthcare decisions.Method: Evaluated two widely deployed clinical AI systems (OpenEvidence and UpToDate Expert AI) against three state-of-the-art generalist LLMs (GPT-5, Gemini 3 Pro, and Claude Sonnet 4.5) using a 1,000-item benchmark combining MedQA (medical knowledge) and HealthBench (clinician-alignment) tasks.
Result: Generalist models consistently outperformed clinical tools, with GPT-5 achieving the highest scores. Clinical tools showed deficits in completeness, communication quality, context awareness, and systems-based safety reasoning.
Conclusion: Clinical decision support tools marketed as specialized may lag behind frontier LLMs, highlighting the urgent need for transparent, independent evaluation before deployment in patient-facing workflows.
Abstract: Specialized clinical AI assistants are rapidly entering medical practice, often framed as safer or more reliable than general-purpose large language models (LLMs). Yet, unlike frontier models, these clinical tools are rarely subjected to independent, quantitative evaluation, creating a critical evidence gap despite their growing influence on diagnosis, triage, and guideline interpretation. We assessed two widely deployed clinical AI systems (OpenEvidence and UpToDate Expert AI) against three state-of-the-art generalist LLMs (GPT-5, Gemini 3 Pro, and Claude Sonnet 4.5) using a 1,000-item mini-benchmark combining MedQA (medical knowledge) and HealthBench (clinician-alignment) tasks. Generalist models consistently outperformed clinical tools, with GPT-5 achieving the highest scores, while OpenEvidence and UpToDate demonstrated deficits in completeness, communication quality, context awareness, and systems-based safety reasoning. These findings reveal that tools marketed for clinical decision support may often lag behind frontier LLMs, underscoring the urgent need for transparent, independent evaluation before deployment in patient-facing workflows.
[54] Conveying Imagistic Thinking in Traditional Chinese Medicine Translation: A Prompt Engineering and LLM-Based Evaluation Framework
Jiatong Han
Main category: cs.CL
TL;DR: A study using human-in-the-loop framework with DeepSeek V3.1 to improve TCM translation by identifying metaphor/metonymy, showing prompt-adjusted LLM translations outperform human and baseline translations across cognitive dimensions.
Details
Motivation: Existing English translations of Traditional Chinese Medicine texts rely too much on literal rendering, making it difficult for target-language readers to understand the underlying conceptual networks and apply TCM theory in clinical practice.Method: Used human-in-the-loop framework with DeepSeek V3.1 guided by prompt-based cognitive scaffolding to identify metaphor and metonymy in Huangdi Neijing passages. Evaluated translations using ChatGPT 5 Pro and Gemini 2.5 Pro simulating three types of real-world readers, scoring across five cognitive dimensions with structured interviews and Interpretative Phenomenological Analysis.
Result: Prompt-adjusted LLM translations performed best across all five cognitive dimensions with high cross-model and cross-role consistency. Interviews revealed differences between human and machine translation, effective strategies for metaphor/metonymy transfer, and readers’ cognitive preferences.
Conclusion: The study provides a cognitive, efficient, and replicable human-in-the-loop methodological pathway for translating ancient, concept-dense texts like TCM, demonstrating the effectiveness of prompt-guided LLM approaches for metaphor and metonymy transfer.
Abstract: Traditional Chinese Medicine (TCM) theory is built on imagistic thinking, in which medical principles and diagnostic and therapeutic logic are structured through metaphor and metonymy. However, existing English translations largely rely on literal rendering, making it difficult for target-language readers to reconstruct the underlying conceptual networks and apply them in clinical practice. This study adopted a human-in-the-loop (HITL) framework and selected four passages from the medical canon Huangdi Neijing that are fundamental in theory. Through prompt-based cognitive scaffolding, DeepSeek V3.1 was guided to identify metaphor and metonymy in the source text and convey the theory in translation. In the evaluation stage, ChatGPT 5 Pro and Gemini 2.5 Pro were instructed by prompts to simulate three types of real-world readers. Human translations, baseline model translations, and prompt-adjusted translations were scored by the simulated readers across five cognitive dimensions, followed by structured interviews and Interpretative Phenomenological Analysis (IPA). Results show that the prompt-adjusted LLM translations perform best across all five dimensions, with high cross-model and cross-role consistency. The interview themes reveal differences between human and machine translation, effective strategies for metaphor and metonymy transfer, and readers’ cognitive preferences. This study provides a cognitive, efficient, and replicable HITL methodological pathway for the translation of ancient, concept-dense texts such as TCM.
[55] Sentiment Analysis and Emotion Classification using Machine Learning Techniques for Nagamese Language - A Low-resource Language
Ekha Morang, Surhoni A. Ngullie, Sashienla Longkumer, Teisovi Angami
Main category: cs.CL
TL;DR: First sentiment analysis and emotion classification study for Nagamese language, building lexicon and using ML models.
Details
Motivation: Nagamese is an important creole language in northeast India used for trade, but no sentiment analysis work exists for it despite substantial work on resource-rich languages like English and Hindi.Method: Built sentiment polarity lexicon of 1,195 Nagamese words, used these as features along with additional features for supervised ML using Naive Bayes and Support Vector Machines.
Result: First attempt at sentiment analysis and emotion classification for Nagamese language, creating foundational lexicon and applying ML techniques.
Conclusion: Pioneering work establishing baseline for Nagamese sentiment analysis, demonstrating feasibility of applying NLP techniques to low-resource creole languages.
Abstract: The Nagamese language, a.k.a Naga Pidgin, is an Assamese-lexified creole language developed primarily as a means of communication in trade between the people from Nagaland and people from Assam in the north-east India. Substantial amount of work in sentiment analysis has been done for resource-rich languages like English, Hindi, etc. However, no work has been done in Nagamese language. To the best of our knowledge, this is the first attempt on sentiment analysis and emotion classification for the Nagamese Language. The aim of this work is to detect sentiments in terms of polarity (positive, negative and neutral) and basic emotions contained in textual content of Nagamese language. We build sentiment polarity lexicon of 1,195 nagamese words and use these to build features along with additional features for supervised machine learning techniques using Na"ive Bayes and Support Vector Machines. Keywords: Nagamese, NLP, sentiment analysis, machine learning
[56] SUPERChem: A Multimodal Reasoning Benchmark in Chemistry
Zehua Zhao, Zhixian Huang, Junren Li, Siyu Lin, Junting Zhou, Fengqi Cao, Kun Zhou, Rui Ge, Tingting Long, Yuexiang Zhu, Yan Liu, Jie Zheng, Junnian Wei, Rong Zhu, Peng Zou, Wenyu Li, Zekai Cheng, Tian Ding, Yaxuan Wang, Yizhao Yan, Tingru Wei, Haowei Ming, Weijie Mao, Chen Sun, Yiming Liu, Zichen Wang, Zuo Zhang, Tong Yang, Hao Ma, Zhen Gao, Jian Pei
Main category: cs.CL
TL;DR: SUPERChem is a new benchmark of 500 expert-curated chemistry problems with multimodal/text formats and solution paths, showing top LLMs still underperform human experts (38.5% vs 40.3%).
Details
Motivation: Current chemistry benchmarks for LLMs have oversimplified tasks, lack process-level evaluation, and misalign with expert-level chemistry skills, creating a need for more rigorous assessment.Method: Created 500 expert-curated reasoning-intensive chemistry problems with iterative curation to eliminate flaws and prevent data contamination. Each problem includes expert-authored solution paths and supports both multimodal and text-only formats. Introduces Reasoning Path Fidelity (RPF) scoring to evaluate reasoning quality beyond final answers.
Result: Best-performing model GPT-5 (High) achieved only 38.5% accuracy, below human baseline of 40.3%. Gemini 2.5 Pro (37.9%) and DeepSeek-V3.1-Think (37.3%) followed closely. Benchmark reveals model-dependent effects of visual information and distinguishes high-fidelity reasoners from heuristic ones.
Conclusion: SUPERChem provides a challenging benchmark and reliable evaluation framework to advance LLMs toward expert-level chemical intelligence, with the dataset publicly available for research.
Abstract: Current benchmarks for evaluating the chemical reasoning capabilities of Large Language Models (LLMs) are limited by oversimplified tasks, lack of process-level evaluation, and misalignment with expert-level chemistry skills. To address these issues, we introduce SUPERChem, a benchmark of 500 expert-curated reasoning-intensive chemistry problems, covering diverse subfields and provided in both multimodal and text-only formats. Original content and an iterative curation pipeline eliminate flawed items and mitigate data contamination. Each problem is paired with an expert-authored solution path, enabling Reasoning Path Fidelity (RPF) scoring to evaluate reasoning quality beyond final-answer accuracy. Evaluations against a human baseline of 40.3% accuracy show that even the best-performing model, GPT-5 (High), reaches only 38.5%, followed closely by Gemini 2.5 Pro (37.9%) and DeepSeek-V3.1-Think (37.3%). SUPERChem elicits multi-step, multimodal reasoning, reveals model-dependent effects of visual information, and distinguishes high-fidelity reasoners from heuristic ones. By providing a challenging benchmark and a reliable evaluation framework, SUPERChem aims to facilitate the advancement of LLMs toward expert-level chemical intelligence. The dataset of the benchmark is available at https://huggingface.co/datasets/ZehuaZhao/SUPERChem.
[57] Kardia-R1: Unleashing LLMs to Reason toward Understanding and Empathy for Emotional Support via Rubric-as-Judge Reinforcement Learning
Jiahao Yuan, Zhiqing Cui, Hanqing Wang, Yuansheng Gao, Yucheng Zhou, Usman Naseem
Main category: cs.CL
TL;DR: KardiaBench is a large-scale user-grounded benchmark with 178K QA pairs across 22K conversations anchored to 671 real-world profiles, enabling identity-aware emotional reasoning. Kardia-R1 framework uses Rubric-ERL for interpretable empathetic cognition, outperforming existing methods across multiple metrics.
Details
Motivation: Current conversational agents lack identity-aware emotional reasoning due to: (1) reliance on situation-centric datasets without persistent user identity, missing personalized affective nuances; and (2) opaque, coarse reward signals that hinder verifiable empathetic reasoning development.Method: Created KardiaBench dataset via model-in-the-loop pipeline with iterative rubric-guided refinement for psychological plausibility and persona consistency. Proposed Kardia-R1 framework using Rubric-as-Judge Empathetic Reinforcement Learning (Rubric-ERL), a GRPO-based method with explainable, human-aligned rubric rewards for stepwise empathetic cognition.
Result: Kardia-R1 consistently outperforms other methods across four LLM backbones in emotion accuracy, empathy, relevance, persona consistency, and safety. The dataset comprises 178,080 QA pairs across 22,080 multi-turn conversations anchored to 671 real-world profiles.
Conclusion: The work addresses critical gaps in empathetic AI by providing a user-grounded benchmark and interpretable framework for identity-aware emotional reasoning, enabling more psychologically plausible and persona-consistent conversational agents.
Abstract: As web platforms evolve towards greater personalization and emotional complexity, conversational agents must transcend superficial empathy to demonstrate identity-aware emotional reasoning. However, existing systems face two limitations: (1) reliance on situation-centric datasets lacking persistent user identity, which hampers the capture of personalized affective nuances; and (2) dependence on opaque, coarse reward signals that hinder development of verifiable empathetic reasoning. To address these gaps, we introduce KardiaBench, a large-scale user-grounded benchmark comprising 178,080 QA pairs across 22,080 multi-turn conversations anchored to 671 real-world profiles. The dataset is constructed via a model-in-the-loop pipeline with iterative rubric-guided refinement to ensure psychological plausibility and persona consistency. This progressive empathy pipeline that integrates user comprehension, contextual reasoning, and emotion perception into conversations, followed by iterative critique and rubric-based refinement to ensure psychological plausibility, emotional fidelity, and persona consistency. Building on this, we propose Kardia-R1, a framework that trains models for interpretable, stepwise empathetic cognition. Kardia-R1 leverages Rubric-as-Judge Empathetic Reinforcement Learning (Rubric-ERL), a GRPO-based method that uses explainable, human-aligned rubric rewards to tightly couple user understanding, emotional inference, and supportive response generation. Extensive experiments across four LLM backbones demonstrate that Kardia-R1 consistently outperforms othet methods in emotion accuracy, empathy, relevance, persona consistency, and safety. Our dataset and model will be released at https://github.com/JhCircle/Kardia-R1.
[58] Agreement-Constrained Probabilistic Minimum Bayes Risk Decoding
Koki Natsumi, Hiroyuki Deguchi, Yusuke Sakai, Hidetaka Kamigaito, Taro Watanabe
Main category: cs.CL
TL;DR: AC-PMBR decoding improves the quality-cost trade-off in machine translation by using a knowledge-distilled model to guide matrix completion for probabilistic MBR decoding.
Details
Motivation: MBR decoding produces high-quality translations but requires quadratic time complexity due to pairwise scoring. PMBR reduces computational cost via sampling and matrix completion but degrades translation quality as utility function calls decrease.Method: Proposed Agreement-Constrained PMBR (AC-PMBR) decoding that leverages a knowledge-distilled model to guide the matrix completion process, improving approximation accuracy of missing quality scores.
Result: AC-PMBR improved matrix completion approximation errors by up to 3 times and achieved higher translation quality than PMBR at comparable computational cost on WMT'23 En↔De translation tasks.
Conclusion: AC-PMBR provides a better quality-cost trade-off for MBR decoding by incorporating knowledge distillation to enhance matrix completion, making high-quality translation more computationally efficient.
Abstract: Minimum Bayes risk (MBR) decoding generates high-quality translations by maximizing the expected utility of output candidates, but it evaluates all pairwise scores over the candidate set; hence, it takes quadratic time with respect to the number of candidates. To reduce the number of utility function calls, probabilistic MBR (PMBR) decoding partially evaluates quality scores using sampled pairs of candidates and completes the missing scores with a matrix completion algorithm. Nevertheless, it degrades the translation quality as the number of utility function calls is reduced. Therefore, to improve the trade-off between quality and cost, we propose agreement-constrained PMBR (AC-PMBR) decoding, which leverages a knowledge distilled model to guide the completion of the score matrix. Our AC-PMBR decoding improved approximation errors of matrix completion by up to 3 times and achieved higher translation quality compared with PMBR decoding at a comparable computational cost on the WMT'23 En$\leftrightarrow$De translation tasks.
[59] MARSAD: A Multi-Functional Tool for Real-Time Social Media Analysis
Md. Rafiul Biswas, Firoj Alam, Wajdi Zaghouani
Main category: cs.CL
TL;DR: MARSAD is an Arabic-focused NLP platform for real-time social media monitoring with sentiment analysis, propaganda detection, and other analytical features.
Details
Motivation: There is a need for comprehensive social media monitoring tools specifically designed for the Arabic-speaking world that can handle real-time analysis and provide various analytical dimensions for both researchers and non-technical users.Method: The platform integrates flexible document storage with structured data management in its backend architecture, supports secure data-scraping via API keys, and provides a user-friendly frontend for data upload and interaction.
Result: MARSAD enables examination of both live and archived social media content, producing detailed visualizations and reports across multiple dimensions including sentiment analysis, emotion analysis, propaganda detection, fact-checking, and hate speech detection.
Conclusion: MARSAD serves as a multifunctional NLP platform that efficiently processes large multimodal datasets while making advanced social media analysis accessible to both technical and non-technical users in the Arabic-speaking context.
Abstract: MARSAD is a multifunctional natural language processing (NLP) platform designed for real-time social media monitoring and analysis, with a particular focus on the Arabic-speaking world. It enables researchers and non-technical users alike to examine both live and archived social media content, producing detailed visualizations and reports across various dimensions, including sentiment analysis, emotion analysis, propaganda detection, fact-checking, and hate speech detection. The platform also provides secure data-scraping capabilities through API keys for accessing public social media data. MARSAD’s backend architecture integrates flexible document storage with structured data management, ensuring efficient processing of large and multimodal datasets. Its user-friendly frontend supports seamless data upload and interaction.
[60] DyFuLM: An Advanced Multimodal Framework for Sentiment Analysis
Ruohan Zhou, Jiachen Yuan, Churui Yang, Wenzheng Huang, Guoyan Zhang, Shiyao Wei, Jiazhen Hu, Ning Xin, Md Maruf Hasan
Main category: cs.CL
TL;DR: DyFuLM is a multimodal framework for sentiment analysis that uses hierarchical dynamic fusion and gated feature aggregation to capture both coarse and fine-grained emotional nuances, achieving state-of-the-art performance on multi-task sentiment datasets.
Details
Motivation: Understanding sentiment in complex textual expressions remains challenging, requiring models that can capture both hierarchical semantic representations and fine-grained emotional nuances simultaneously.Method: Proposes Dynamic Fusion Learning Model (DyFuLM) with two key modules: 1) Hierarchical Dynamic Fusion module for adaptive multi-level feature integration, and 2) Gated Feature Aggregation module for regulating cross-layer information flow to achieve balanced representation learning.
Result: Achieves 82.64% coarse-grained and 68.48% fine-grained accuracy, with lowest regression errors (MAE=0.0674, MSE=0.0082) and highest R²=0.6903. Ablation studies show each module contributes significantly - removing all modules causes accuracy drops of 0.91% (coarse) and 0.68% (fine).
Conclusion: DyFuLM enhances sentiment representation and overall performance through effective hierarchical feature fusion, with each module playing a crucial role in feature interaction and task balance.
Abstract: Understanding sentiment in complex textual expressions remains a fundamental challenge in affective computing. To address this, we propose a Dynamic Fusion Learning Model (DyFuLM), a multimodal framework designed to capture both hierarchical semantic representations and fine-grained emotional nuances. DyFuLM introduces two key moodules: a Hierarchical Dynamic Fusion module that adaptively integrates multi-level features, and a Gated Feature Aggregation module that regulates cross-layer information ffow to achieve balanced representation learning. Comprehensive experiments on multi-task sentiment datasets demonstrate that DyFuLM achieves 82.64% coarse-grained and 68.48% fine-grained accuracy, yielding the lowest regression errors (MAE = 0.0674, MSE = 0.0082) and the highest R^2 coefficient of determination (R^2= 0.6903). Furthermore, the ablation study validates the effectiveness of each module in DyFuLM. When all modules are removed, the accuracy drops by 0.91% for coarse-grained and 0.68% for fine-grained tasks. Keeping only the gated fusion module causes decreases of 0.75% and 0.55%, while removing the dynamic loss mechanism results in drops of 0.78% and 0.26% for coarse-grained and fine-grained sentiment classification, respectively. These results demonstrate that each module contributes significantly to feature interaction and task balance. Overall, the experimental findings further validate that DyFuLM enhances sentiment representation and overall performance through effective hierarchical feature fusion.
[61] PromptBridge: Cross-Model Prompt Transfer for Large Language Models
Yaxuan Wang, Quan Liu, Zhenting Wang, Zichao Li, Wei Wei, Yang Liu, Yujia Bao
Main category: cs.CL
TL;DR: PromptBridge enables cross-model prompt transfer without costly re-optimization by learning mappings between model-specific prompts, addressing the severe performance degradation when prompts are reused across different LLMs.
Details
Motivation: LLMs are accessed via various commercial APIs and open-source deployments with rapidly evolving model landscapes. Frequent model switches are driven by capability, cost, deployment constraints, and privacy needs. However, prompts are highly model-sensitive - reusing prompts across models causes severe performance degradation (Model Drifting), requiring costly per-model prompt re-optimization.Method: PromptBridge is a training-free framework that preserves prompt effectiveness under model switches. It requires only a small set of alignment tasks for calibration. First, Model-Adaptive Reflective Prompt Evolution (MAP-RPE) obtains task- and model-specific optimal prompts via iterative reflective refinement and quantitative evaluation. Using calibrated prompt pairs for source and target models, PromptBridge learns a cross-model prompt mapping. At test time, given a source-model prompt for an unseen task, this mapping directly produces an optimized prompt for the target model.
Result: Experiments in single-agent and multi-agent settings show that PromptBridge consistently improves downstream accuracy while reducing migration effort. The framework effectively addresses Model Drifting without requiring costly per-task or per-model re-optimization.
Conclusion: PromptBridge provides an effective solution to the Model Drifting problem by enabling cross-model prompt transfer through learned mappings between model-specific prompts, reducing the need for expensive prompt re-engineering when switching between different LLMs.
Abstract: Large language models (LLMs) underpin applications in code generation, mathematical reasoning, and agent-based workflows. In practice, systems access LLMs via commercial APIs or open-source deployments, and the model landscape (e.g., GPT, Claude, Llama) evolves rapidly. This rapid evolution forces frequent model switches driven by capability, cost, deployment constraints, and privacy. Yet prompts are highly model-sensitive: reusing a prompt engineered for one model on another often yields substantially worse performance than a prompt optimized for the target model. We term this phenomenon Model Drifting. Through extensive empirical analysis across diverse LLM configurations, we show that model drifting is both common and severe. To address this challenge, we introduce PromptBridge, a training-free framework that preserves prompt effectiveness under model switches, enabling cross-model prompt transfer without costly per-task or per-model re-optimization. PromptBridge requires only a small set of alignment tasks for calibration. It first applies Model-Adaptive Reflective Prompt Evolution (MAP-RPE) to obtain task- and model-specific optimal prompts via iterative reflective refinement and quantitative evaluation. Using the resulting calibrated prompt pairs for the source and target models, PromptBridge learns a cross-model prompt mapping. At test time, i.e., for an unseen task, given a source-model prompt, this mapping directly produces an optimized prompt for the target model. Experiments in single-agent and multi-agent settings show that PromptBridge consistently improves downstream accuracy while reducing migration effort. The code will be available soon.
[62] Multilingual Conversational AI for Financial Assistance: Bridging Language Barriers in Indian FinTech
Bharatdeep Hazarika, Arya Suneesh, Prasanna Devadiga, Pawan Kumar Rajpoot, Anshuman B Suresh, Ahmed Ifthaquar Hussain
Main category: cs.CL
TL;DR: Multilingual conversational AI system for Indian fintech that supports code-mixed languages like Hinglish to overcome English-language barriers and improve financial inclusion.
Details
Motivation: India's linguistic diversity creates barriers to financial inclusion - with 31 major languages and over 100 minor ones, but only 10% of the population understands English. This language gap hinders access to digital financial services for the majority of Indians.Method: Multi-agent architecture with language classification, function management, and multilingual response generation. The system supports code-mixed languages like Hinglish and employs comparative analysis of multiple language models.
Result: Significant improvements in user engagement while maintaining low latency overhead (4-8%). Demonstrated effectiveness through real-world deployment in financial assistance use cases.
Conclusion: The work contributes to bridging the language gap in digital financial services for emerging markets, enabling natural interactions for India’s diverse user base and advancing financial inclusion through multilingual conversational AI.
Abstract: India’s linguistic diversity presents both opportunities and challenges for fintech platforms. While the country has 31 major languages and over 100 minor ones, only 10% of the population understands English, creating barriers to financial inclusion. We present a multilingual conversational AI system for a financial assistance use case that supports code-mixed languages like Hinglish, enabling natural interactions for India’s diverse user base. Our system employs a multi-agent architecture with language classification, function management, and multilingual response generation. Through comparative analysis of multiple language models and real-world deployment, we demonstrate significant improvements in user engagement while maintaining low latency overhead (4-8%). This work contributes to bridging the language gap in digital financial services for emerging markets.
[63] Enhancing BERT Fine-Tuning for Sentiment Analysis in Lower-Resourced Languages
Jozef Kubík, Marek Šuppa, Martin Takáč
Main category: cs.CL
TL;DR: Active Learning with clustering-based schedulers boosts fine-tuning for low-resource languages, achieving 30% annotation savings and up to 4 F1 improvement.
Details
Motivation: Low-resource languages have limited training data, leading to weaker language models. Since pre-training is compute-intensive, improving fine-tuning efficiency is more practical.Method: Proposes Active Learning (AL) methods augmented with structured data selection strategies called ‘AL schedulers’, connecting AL to data clustering. Creates an integrated fine-tuning pipeline combining AL, clustering, and dynamic data selection schedulers.
Result: Experiments on Slovak, Maltese, Icelandic, and Turkish show clustering during fine-tuning with AL scheduling yields: 30% annotation savings, up to 4 F1 score improvement, and better fine-tuning stability.
Conclusion: The integrated AL-clustering-scheduler pipeline effectively enhances fine-tuning for low-resource languages, providing significant efficiency gains and performance improvements simultaneously.
Abstract: Limited data for low-resource languages typically yield weaker language models (LMs). Since pre-training is compute-intensive, it is more pragmatic to target improvements during fine-tuning. In this work, we examine the use of Active Learning (AL) methods augmented by structured data selection strategies which we term ‘Active Learning schedulers’, to boost the fine-tuning process with a limited amount of training data. We connect the AL to data clustering and propose an integrated fine-tuning pipeline that systematically combines AL, clustering, and dynamic data selection schedulers to enhance model’s performance. Experiments in the Slovak, Maltese, Icelandic and Turkish languages show that the use of clustering during the fine-tuning phase together with AL scheduling can simultaneously produce annotation savings up to 30% and performance improvements up to four F1 score points, while also providing better fine-tuning stability.
[64] MCAT: Scaling Many-to-Many Speech-to-Text Translation with MLLMs to 70 Languages
Yexing Du, Kaiyuan Liu, Youcheng Pan, Bo Yang, Keqi Deng, Xie Chen, Yang Xiang, Ming Liu, Bin Qin, YaoWei Wang
Main category: cs.CL
TL;DR: MCAT is a Multilingual Cost-effective Accelerated Speech-to-Text Translator framework that extends MLLMs to support 70 languages with efficient speech sequence compression, achieving state-of-the-art performance with minimal training data.
Details
Motivation: Current MLLMs for Speech-to-Text Translation face two major limitations: 1) English-centric datasets restrict many-to-many translation capabilities across diverse languages, and 2) long speech sequences (e.g., 750 tokens) dramatically slow down inference speed.Method: Two innovations: 1) Language scaling method using curriculum learning and data balancing to extend MLLMs to 70 languages, 2) Optimized speech adapter module that compresses speech sequences to only 30 tokens for efficiency.
Result: MCAT surpasses state-of-the-art end-to-end models on FLEURS dataset across 70x69 directions, enhances batch inference efficiency, achieves this with only ~100M trainable parameters and 10 hours of S2TT data per language.
Conclusion: MCAT effectively addresses language coverage and efficiency challenges in MLLMs for S2TT, providing a scalable and efficient framework for multilingual speech translation that has been released as open-source to advance the field.
Abstract: Multimodal Large Language Models (MLLMs) have achieved great success in Speech-to-Text Translation (S2TT) tasks. However, current research is constrained by two key challenges: language coverage and efficiency. Most of the popular S2TT datasets are substantially English-centric, which restricts the scaling-up of MLLMs’ many-to-many translation capabilities. Moreover, the inference speed of MLLMs degrades dramatically when the speech is converted into long sequences (e.g., 750 tokens). To address these limitations, we propose a Multilingual Cost-effective Accelerated Speech-to-Text Translator (MCAT) framework, which includes two innovations. First, a language scaling method that leverages curriculum learning and a data balancing strategy is introduced to extend the language coverage supported by MLLMs to 70 languages and achieve mutual translation among these languages. Second, an optimized speech adapter module is designed to reduce the length of the speech sequence to only 30 tokens. Extensive experiments were conducted on MLLMs of different scales (9B and 27B). The experimental results demonstrate that MCAT not only surpasses state-of-the-art end-to-end models on the FLEURS dataset across 70x69 directions but also enhances batch inference efficiency. This is achieved with only ~100M trainable parameters and by using only 10 hours of S2TT data per language. Furthermore, we have released MCAT as open-source to promote the development of MLLMs for robust S2TT capabilities. The code and models are released at https://github.com/yxduir/m2m-70.
[65] Language Diversity: Evaluating Language Usage and AI Performance on African Languages in Digital Spaces
Edward Ajayi, Eudoxie Umwari, Mawuli Deku, Prosper Singadi, Jules Udahemuka, Bekalu Tadele, Chukuemeka Edeh
Main category: cs.CL
TL;DR: Current language detection tools struggle with African languages due to data scarcity and code-switching in conversational platforms, but perform well on clean news data.
Details
Motivation: African languages face digital representation challenges with sparse online usage, heavy English influence, and code-switching on conversational platforms, creating data scarcity for training language models.Method: Collected data from subreddits and local news sources for Yoruba, Kinyarwanda, and Amharic; evaluated performance of language detection models (AfroLID and general LLM) on both data types.
Result: Reddit data was minimal with heavy code-switching, while local news media provided robust clean monolingual data. Language detection models performed near-perfect on news data but struggled with code-switched Reddit posts.
Conclusion: Professionally curated news content is more reliable for training AI models for African languages than conversational platform data, and future models need to handle both clean and code-switched text.
Abstract: This study examines the digital representation of African languages and the challenges this presents for current language detection tools. We evaluate their performance on Yoruba, Kinyarwanda, and Amharic. While these languages are spoken by millions, their online usage on conversational platforms is often sparse, heavily influenced by English, and not representative of the authentic, monolingual conversations prevalent among native speakers. This lack of readily available authentic data online creates a challenge of scarcity of conversational data for training language models. To investigate this, data was collected from subreddits and local news sources for each language. The analysis showed a stark contrast between the two sources. Reddit data was minimal and characterized by heavy code-switching. Conversely, local news media offered a robust source of clean, monolingual language data, which also prompted more user engagement in the local language on the news publishers social media pages. Language detection models, including the specialized AfroLID and a general LLM, performed with near-perfect accuracy on the clean news data but struggled with the code-switched Reddit posts. The study concludes that professionally curated news content is a more reliable and effective source for training context-rich AI models for African languages than data from conversational platforms. It also highlights the need for future models that can process clean and code-switched text to improve the detection accuracy for African languages.
[66] Learning the Boundary of Solvability: Aligning LLMs to Detect Unsolvable Problems
Dengyun Peng, Qiguang Chen, Bofei Liu, Jiannan Guan, Libo Qin, Zheng Yan, Jinhao Liu, Jianshu Zhang, Wanxiang Che
Main category: cs.CL
TL;DR: The paper introduces UnsolvableQA and UnsolvableRL to help LLMs distinguish between objectively unsolvable problems and those beyond their capability, improving reliability by preventing hallucinations and overconfidence.
Details
Motivation: Current LLMs struggle to differentiate between objective unsolvability (inherent contradictions) and subjective capability limitations, leading to hallucinations and overconfidence. This reliability gap needs addressing for practical deployment.Method: 1) Construct UnsolvableQA dataset using dual-track methodology: programmatic generation for logic puzzles and “Reverse Construction” for mathematics. 2) Develop UnsolvableRL reinforcement learning framework with three reward components for accuracy, unsolvability detection, and difficulty.
Result: The approach achieves near-perfect unsolvability detection while improving accuracy on solvable tasks. Key finding: “Capability Collapse” shows explicit exposure to unsolvable data is essential to prevent systematic overconfidence.
Conclusion: The proposed framework successfully enables LLMs to solve feasible problems, detect inherent contradictions, and prudently refuse tasks beyond their capability, addressing a critical reliability challenge in LLM deployment.
Abstract: Ensuring LLM reliability requires not only solving complex problems but also recognizing when a problem is unsolvable. Current models often struggle to distinguish objective unsolvability (inherent contradictions in the problem) from subjective capability limitations (problems beyond the model’s competence), which leads to hallucinations and overconfidence. To address this, we propose UnsolvableQA and UnsolvableRL to solve feasible problems, detect inherent contradictions, and prudently refuse tasks beyond capability. Specifically, we construct UnsolvableQA, a dataset of paired solvable and unsolvable instances derived via a dual-track methodology: programmatic generation for logic puzzles and a novel “Reverse Construction” method that injects contradictions into valid reasoning chains for mathematics. Building on this dataset, we introduce UnsolvableRL, a reinforcement learning framework with three reward components jointly accounting for accuracy, unsolvability, and difficulty. Empirical results show that our approach achieves near-perfect unsolvability detection while also improving accuracy on solvable tasks. Crucially, we identify Capability Collapse, demonstrating that explicit exposure to unsolvable data is indispensable for preventing models from becoming systematically overconfident. Our code and data are available at https://github.com/sfasfaffa/unsolvableQA.
[67] MMAG: Mixed Memory-Augmented Generation for Large Language Models Applications
Stefano Zeppieri
Main category: cs.CL
TL;DR: MMAG framework organizes memory for LLM agents into five cognitive layers to improve relevance, personalization, and continuity in extended conversations.
Details
Motivation: LLMs struggle with sustaining relevance, personalization, and continuity across extended interactions, while human communication relies on multiple forms of memory including past conversations, personal traits, and situational context.Method: Introduces Mixed Memory-Augmented Generation (MMAG) pattern with five interacting memory layers: conversational, long-term user, episodic and event-linked, sensory and context-aware, and short-term working memory. Draws from cognitive psychology and includes coordination, prioritization, and conflict resolution strategies.
Result: Implemented in Heero conversational agent where encrypted long-term bios and conversational history already improve engagement and retention. Framework addresses storage, retrieval, privacy, and latency concerns.
Conclusion: MMAG provides foundation for building memory-rich language agents that are more coherent, proactive, and aligned with human needs, though open challenges remain.
Abstract: Large Language Models (LLMs) excel at generating coherent text within a single prompt but fall short in sustaining relevance, personalization, and continuity across extended interactions. Human communication, however, relies on multiple forms of memory, from recalling past conversations to adapting to personal traits and situational context. This paper introduces the Mixed Memory-Augmented Generation (MMAG) pattern, a framework that organizes memory for LLM-based agents into five interacting layers: conversational, long-term user, episodic and event-linked, sensory and context-aware, and short-term working memory. Drawing inspiration from cognitive psychology, we map these layers to technical components and outline strategies for coordination, prioritization, and conflict resolution. We demonstrate the approach through its implementation in the Heero conversational agent, where encrypted long-term bios and conversational history already improve engagement and retention. We further discuss implementation concerns around storage, retrieval, privacy, and latency, and highlight open challenges. MMAG provides a foundation for building memory-rich language agents that are more coherent, proactive, and aligned with human needs.
[68] Self-Supervised Borrowing Detection on Multilingual Wordlists
Tim Wientzek
Main category: cs.CL
TL;DR: Self-supervised approach for detecting borrowed words in multilingual lists using PMI similarities and contrastive learning on phonetic features, with automatic threshold selection and performance matching supervised methods.
Details
Motivation: Need for automated borrowing detection in multilingual wordlists without requiring labeled training data, overcoming limitations of existing string similarity measures and supervised approaches that need manual annotation.Method: Combines PMI similarities from global correspondence model with lightweight contrastive component trained on phonetic feature vectors, plus automatic threshold selection procedure without labeled data.
Result: PMI alone outperforms existing string similarity measures (NED, SCA), and combined approach performs on par with or better than supervised baselines. Ablation shows importance of character encoding, temperature settings, and augmentation strategies.
Conclusion: Scalable self-supervised approach works without manual supervision, provides command-line tool for researchers, and achieves competitive performance with supervised methods while eliminating need for labeled data.
Abstract: This paper presents a fully self-supervised approach to borrowing detection in multilingual wordlists. The method combines two sources of information: PMI similarities based on a global correspondence model and a lightweight contrastive component trained on phonetic feature vectors. It further includes an automatic procedure for selecting decision thresholds without requiring labeled data. Experiments on benchmark datasets show that PMI alone already improves over existing string similarity measures such as NED and SCA, and that the combined similarity performs on par with or better than supervised baselines. An ablation study highlights the importance of character encoding, temperature settings and augmentation strategies. The approach scales to datasets of different sizes, works without manual supervision and is provided with a command-line tool that allows researchers to conduct their own studies.
[69] Beware of Reasoning Overconfidence: Pitfalls in the Reasoning Process for Multi-solution Tasks
Jiannan Guan, Qiguang Chen, Libo Qin, Dengyun Peng, Jinhao Liu, Liangyu Huo, Jian Xie, Wanxiang Che
Main category: cs.CL
TL;DR: LLMs show reasoning overconfidence in multi-solution tasks, performing poorly despite excelling at single-answer problems. Long chain-of-thought prompting helps mitigate this issue through iterative exploration.
Details
Motivation: LLMs perform well on single-answer reasoning tasks but struggle with multi-solution problems requiring comprehensive and diverse answers. The authors identify "reasoning overconfidence" as a key limitation where LLMs express undue certainty in incomplete solution sets.Method: Introduced MuSoBench benchmark for multi-solution problems. Compared conventional short chain-of-thought (Short-CoT) with emerging long chain-of-thought (Long-CoT) prompting. Proposed cognitive-rigidity hypothesis and conducted attention-entropy analysis to examine underlying causes.
Result: Short-CoT exhibits pronounced overconfidence, while Long-CoT mitigates it through iterative exploration and self-reflection. Attention-entropy analysis provides preliminary support for the cognitive-rigidity hypothesis that overconfidence arises from premature convergence on narrow thought paths.
Conclusion: The findings provide tools for assessing LLM reasoning completeness and highlight the need to move beyond single-answer accuracy evaluation toward comprehensive exploration in multi-solution tasks.
Abstract: Large Language Models (LLMs) excel in reasoning tasks requiring a single correct answer, but they perform poorly in multi-solution tasks that require generating comprehensive and diverse answers. We attribute this limitation to \textbf{reasoning overconfidence}: a tendency to express undue certainty in an incomplete solution set. To examine the effect, we introduce \textit{MuSoBench}, a benchmark of multi-solution problems. Experiments show that the conventional short chain-of-thought (Short-CoT) prompting paradigm exhibits pronounced overconfidence, whereas the emerging long chain-of-thought (Long-CoT) approach mitigates it through iterative exploration and self-reflection. We further characterise observable behaviours and influential factors. To probe the underlying cause, we propose the \textbf{cognitive-rigidity hypothesis}, which posits that overconfidence arises when the reasoning process prematurely converges on a narrow set of thought paths. An attention-entropy analysis offers preliminary support for this view. These findings provide tools for assessing the completeness of LLM reasoning and highlight the need to move evaluation beyond single-answer accuracy toward comprehensive exploration.
[70] Reasoning About the Unsaid: Misinformation Detection with Omission-Aware Graph Inference
Zhengjia Wang, Danding Wang, Qiang Sheng, Jiaying Wu, Juan Cao
Main category: cs.CL
TL;DR: OmiGraph is the first omission-aware framework for misinformation detection that identifies deception through both explicit fabrication and implicit omission of important information.
Details
Motivation: Current misinformation detection focuses mainly on explicit fabrication, overlooking omission-based deception where important information is omitted to subtly guide readers toward false conclusions under the illusion of completeness.Method: OmiGraph constructs omission-aware graphs using contextual environments to capture complementary perspectives, performs omission-oriented relation modeling to identify contextual dependencies and omission intents, and uses omission-aware message-passing and aggregation to extract deception patterns.
Result: The approach achieves average improvements of +5.4% F1 and +5.3% ACC on two large-scale benchmarks, demonstrating remarkable performance by considering omission perspectives.
Conclusion: Omission-based deception is a significant but overlooked aspect of misinformation, and OmiGraph successfully addresses this gap by providing the first comprehensive framework for omission-aware misinformation detection with substantial performance improvements.
Abstract: This paper investigates the detection of misinformation, which deceives readers by explicitly fabricating misleading content or implicitly omitting important information necessary for informed judgment. While the former has been extensively studied, omission-based deception remains largely overlooked, even though it can subtly guide readers toward false conclusions under the illusion of completeness. To pioneer in this direction, this paper presents OmiGraph, the first omission-aware framework for misinformation detection. Specifically, OmiGraph constructs an omission-aware graph for the target news by utilizing a contextual environment that captures complementary perspectives of the same event, thereby surfacing potentially omitted contents. Based on this graph, omission-oriented relation modeling is then proposed to identify the internal contextual dependencies, as well as the dynamic omission intents, formulating a comprehensive omission relation representation. Finally, to extract omission patterns for detection, OmiGraph introduces omission-aware message-passing and aggregation that establishes holistic deception perception by integrating the omission contents and relations. Experiments show that, by considering the omission perspective, our approach attains remarkable performance, achieving average improvements of +5.4% F1 and +5.3% ACC on two large-scale benchmarks.
[71] Beyond SFT: Reinforcement Learning for Safer Large Reasoning Models with Better Reasoning Ability
Jinghan Jia, Nathalie Baracaldo, Sijia Liu
Main category: cs.CL
TL;DR: RL-based safety alignment for large reasoning models outperforms supervised fine-tuning, achieving stronger safety gains while maintaining reasoning ability by suppressing unsafe intermediate reasoning.
Details
Motivation: Chain-of-thought reasoning in large reasoning models introduces new safety risks as unsafe behaviors can emerge in intermediate reasoning steps even when final answers appear harmless. Existing supervised fine-tuning approaches produce inconsistent safety improvements, degrade reasoning ability, and generalize poorly across model families.Method: Investigates reinforcement learning (RL) as a complementary optimization framework for LRM safety training. Unlike supervised fine-tuning, RL directly optimizes model policies with reward feedback, enabling more adaptive and stable alignment.
Result: Extensive experiments across multiple model families and benchmarks show that RL achieves stronger and more consistent safety gains while maintaining reasoning competence. Analysis reveals RL suppresses unsafe exploratory reasoning while preserving reflective depth.
Conclusion: RL provides a more effective approach for safety alignment in large reasoning models compared to purely supervised methods, leading to safer and more reliable reasoning processes without compromising reasoning ability.
Abstract: Large reasoning models (LRMs) extend large language models by generating explicit chain-of-thought (CoT) reasoning, significantly improving mathematical and logical problem solving. However, this explicit reasoning process also introduces new safety risks, as unsafe behaviors often emerge within intermediate reasoning trajectories, even when final answers appear harmless. Existing safety alignment approaches primarily rely on supervised fine-tuning (SFT) over safety-oriented long CoT datasets. While intuitive, we find that SFT produces inconsistent safety improvements, degrades reasoning ability, and generalizes poorly across model families. These limitations suggest that purely supervised approaches are insufficient for robust safety alignment in LRMs. To address this, we investigate reinforcement learning (RL) as a complementary optimization framework for LRM safety training. Unlike SFT, RL directly optimizes model policies with reward feedback, enabling more adaptive and stable alignment. Extensive experiments across multiple model families and benchmarks show that RL achieves stronger and more consistent safety gains while maintaining reasoning competence. Further analysis of reflection dynamics and token-level entropy reveals that RL suppresses unsafe exploratory reasoning while preserving reflective depth, leading to safer and more reliable reasoning processes.
[72] BHRAM-IL: A Benchmark for Hallucination Recognition and Assessment in Multiple Indian Languages
Hrishikesh Terdalkar, Kirtan Bhojani, Aryan Dongare, Omm Aditya Behera
Main category: cs.CL
TL;DR: BHRAM-IL is a new benchmark for evaluating hallucination detection in Indian languages (Hindi, Gujarati, Marathi, Odia, plus English) with 36K+ questions across 9 categories, showing current multilingual LLMs perform poorly with primary score of 0.23.
Details
Motivation: Hallucination detection has been extensively studied in English but remains largely unexplored for under-resourced Indian languages, despite LLMs being increasingly deployed in multilingual applications where they generate plausible but incorrect outputs.Method: Created BHRAM-IL benchmark with 36,047 curated questions across 9 categories (factual, numerical, reasoning, linguistic tasks). Evaluated 14 state-of-the-art multilingual LLMs on subset of 10,265 questions using category-specific metrics normalized to (0,1) range, analyzing cross-lingual and factual hallucinations across languages, models, scales, categories, and domains.
Result: Aggregation over all categories and models yields primary score of 0.23 and language-corrected fuzzy score of 0.385, demonstrating poor hallucination detection performance. Dataset and code are publicly available on GitHub and HuggingFace.
Conclusion: BHRAM-IL provides a valuable resource for hallucination-focused evaluation in Indian languages, highlighting significant performance gaps and supporting future research in multilingual hallucination detection and mitigation.
Abstract: Large language models (LLMs) are increasingly deployed in multilingual applications but often generate plausible yet incorrect or misleading outputs, known as hallucinations. While hallucination detection has been studied extensively in English, under-resourced Indian languages remain largely unexplored. We present BHRAM-IL, a benchmark for hallucination recognition and assessment in multiple Indian languages, covering Hindi, Gujarati, Marathi, Odia, along with English. The benchmark comprises 36,047 curated questions across nine categories spanning factual, numerical, reasoning, and linguistic tasks. We evaluate 14 state-of-the-art multilingual LLMs on a benchmark subset of 10,265 questions, analyzing cross-lingual and factual hallucinations across languages, models, scales, categories, and domains using category-specific metrics normalized to (0,1) range. Aggregation over all categories and models yields a primary score of 0.23 and a language-corrected fuzzy score of 0.385, demonstrating the usefulness of BHRAM-IL for hallucination-focused evaluation. The dataset, and the code for generation and evaluation are available on GitHub (https://github.com/sambhashana/BHRAM-IL/) and HuggingFace (https://huggingface.co/datasets/sambhashana/BHRAM-IL/) to support future research in multilingual hallucination detection and mitigation.
[73] Cross-Lingual Interleaving for Speech Language Models
Adel Moumen, Guangzhi Sun, Philip C. Woodland
Main category: cs.CL
TL;DR: Cross-lingual interleaving method improves multilingual spoken language models without textual supervision, with new EN-FR dataset and benchmarks.
Details
Motivation: Current SLMs are English-centric due to scarce spoken evaluation benchmarks and training data, limiting cross-lingual learning and access to NLP for languages with limited written resources.Method: Cross-lingual interleaving that mixes speech tokens across languages without textual supervision, plus release of EN-FR training dataset (TinyStories, ~42k hours) and synthetic benchmarks (StoryCloze, TopicCloze) generated using GPT-4.
Result: Interleaving improves monolingual semantic accuracy, enables robust cross-lingual continuation, and strengthens cross-lingual hidden-state alignment in 360M and 1B SLMs under matched training-token budgets.
Conclusion: Cross-lingual interleaving is a simple, scalable route to building multilingual SLMs that understand and converse across languages, with all resources made open-source for reproducibility.
Abstract: Spoken Language Models (SLMs) aim to learn linguistic competence directly from speech using discrete units, widening access to Natural Language Processing (NLP) technologies for languages with limited written resources. However, progress has been largely English-centric due to scarce spoken evaluation benchmarks and training data, making cross-lingual learning difficult. We present a cross-lingual interleaving method that mixes speech tokens across languages without textual supervision. We also release an EN-FR training dataset, TinyStories (~42k hours), together with EN-FR spoken StoryCloze and TopicCloze benchmarks for cross-lingual semantic evaluation, both synthetically generated using GPT-4. On 360M and 1B SLMs under matched training-token budgets, interleaving improves monolingual semantic accuracy, enables robust cross-lingual continuation, and strengthens cross-lingual hidden-state alignment. Taken together, these results indicate that cross-lingual interleaving is a simple, scalable route to building multilingual SLMs that understand and converse across languages. All resources will be made open-source to support reproducibility.
[74] Exploring Human Perceptions of AI Responses: Insights from a Mixed-Methods Study on Risk Mitigation in Generative Models
Heloisa Candello, Muneeza Azmat, Uma Sushmitha Gunturi, Raya Horesh, Rogerio Abreu de Paula, Heloisa Pimentel, Marcelo Carpinette Grave, Aminat Adebiyi, Tiago Machado, Maysa Malfiza Garcia de Macedo
Main category: cs.CL
TL;DR: Study examines human perceptions of AI mitigation strategies for harmful content, finding that language background, AI experience, and annotation familiarity significantly influence evaluations, with participants showing high sensitivity to linguistic quality and context preservation.
Details
Motivation: With generative AI's rapid adoption, understanding human perceptions of AI-generated responses is crucial, especially regarding their tendency to hallucinate and produce harmful content. Despite guardrail implementation efforts, how humans perceive these mitigation strategies remains largely unknown.Method: Mixed-method experiment evaluating mitigation strategy responses across faithfulness, fairness, harm-removal capacity, and relevance. Within-subject study with 57 participants assessing responses under two conditions: harmful response plus its mitigation vs. solely mitigated response.
Result: Participants’ native language, AI work experience, and annotation familiarity significantly influenced evaluations. Participants showed high sensitivity to linguistic and contextual attributes, penalizing minor grammar errors while rewarding preserved semantic contexts. This contrasts with typical quantitative LLM evaluation approaches.
Conclusion: The study provides new metrics for training and evaluating mitigation strategies and offers insights for human-AI evaluation studies, highlighting the importance of considering human factors and linguistic sensitivity in AI safety evaluations.
Abstract: With the rapid uptake of generative AI, investigating human perceptions of generated responses has become crucial. A major challenge is their `aptitude’ for hallucinating and generating harmful contents. Despite major efforts for implementing guardrails, human perceptions of these mitigation strategies are largely unknown. We conducted a mixed-method experiment for evaluating the responses of a mitigation strategy across multiple-dimensions: faithfulness, fairness, harm-removal capacity, and relevance. In a within-subject study design, 57 participants assessed the responses under two conditions: harmful response plus its mitigation and solely mitigated response. Results revealed that participants’ native language, AI work experience, and annotation familiarity significantly influenced evaluations. Participants showed high sensitivity to linguistic and contextual attributes, penalizing minor grammar errors while rewarding preserved semantic contexts. This contrasts with how language is often treated in the quantitative evaluation of LLMs. We also introduced new metrics for training and evaluating mitigation strategies and insights for human-AI evaluation studies.
[75] OPOR-Bench: Evaluating Large Language Models on Online Public Opinion Report Generation
Jinzheng Yu, Yang Xu, Haozhen Li, Junqi Li, Yifan Feng, Ligu Zhu, Hao Shen, Lei Shi
Main category: cs.CL
TL;DR: The paper defines Automated Online Public Opinion Report Generation (OPOR-GEN) task, creates OPOR-BENCH dataset with 463 crisis events, and proposes OPOR-EVAL agent-based evaluation framework that correlates well with human judgments.
Details
Motivation: Online public opinion reports are crucial for crisis management but lack systematic research, formal task definitions, and benchmarks despite technical feasibility with large language models.Method: 1) Define OPOR-GEN task for automated report generation; 2) Construct OPOR-BENCH dataset covering 463 crisis events with news, social media, and reference summaries; 3) Propose OPOR-EVAL agent-based evaluation framework simulating human expert assessment.
Result: Experiments show OPOR-EVAL achieves high correlation with human judgments. The comprehensive framework provides solid foundation for future research in automated public opinion report generation.
Conclusion: The paper successfully bridges the research gap by providing formal task definition, benchmark dataset, and evaluation framework for automated online public opinion report generation, enabling systematic future research in this critical domain.
Abstract: Online Public Opinion Reports consolidate news and social media for timely crisis management by governments and enterprises. While large language models have made automated report generation technically feasible, systematic research in this specific area remains notably absent, particularly lacking formal task definitions and corresponding benchmarks. To bridge this gap, we define the Automated Online Public Opinion Report Generation (OPOR-GEN) task and construct OPOR-BENCH, an event-centric dataset covering 463 crisis events with their corresponding news articles, social media posts, and a reference summary. To evaluate report quality, we propose OPOR-EVAL, a novel agent-based framework that simulates human expert evaluation by analyzing generated reports in context. Experiments with frontier models demonstrate that our framework achieves high correlation with human judgments. Our comprehensive task definition, benchmark dataset, and evaluation framework provide a solid foundation for future research in this critical domain.
[76] Latent Debate: A Surrogate Framework for Interpreting LLM Thinking
Lihu Chen, Xiang Yin, Francesca Toni
Main category: cs.CL
TL;DR: Latent debate is a framework for interpreting LLM predictions by capturing hidden supporting/attacking signals within a single model during inference, providing interpretability and hallucination detection.
Details
Motivation: To understand the internal thinking process of LLMs and the causes of hallucinations, which remain key challenges in AI interpretability.Method: Introduces latent debate framework that captures hidden supporting/attacking signals within a single model during single inference (unlike explicit multi-agent debates). Presents model- and task-agnostic conceptual framework, then instantiates it symbolically for True/False prediction tasks.
Result: Latent debate serves as a faithful structured surrogate model with highly consistent predictions with original LLM. Provides strong baseline for hallucination detection. Analysis reveals strong correlations between hallucinations and debate patterns (e.g., high degree of latent debates in middle layers linked to higher hallucination risk).
Conclusion: Latent debate positions as a potential framework for understanding internal mechanisms of LLMs, especially for scenarios where internal (dis)agreements appear during inference steps.
Abstract: Understanding the internal thinking process of Large Language Models (LLMs) and the cause of hallucinations remains a key challenge. To this end, we introduce latent debate, a novel framework for interpreting model predictions through the lens of implicit internal arguments. Unlike the current work of self-consistency and multi-agent debate, which relies on explicit debates among multiple answers or multiple models, latent debate captures the hidden supporting and attacking signals that arise within a single model during a single inference. We first present a model- and task-agnostic conceptual framework, and then instantiate it symbolically to approximate the thinking process of LLMs on True/False prediction tasks. Empirical studies demonstrate that latent debate is a faithful structured surrogate model that has highly consistent predictions with the original LLM. Beyond interpretability, we demonstrate that latent debate provides a strong baseline for hallucination detection. Further analysis reveals strong correlations between hallucinations and debate patterns, such as a high degree of latent debates in the middle layers is linked to a higher risk of hallucinations. These findings position latent debate as a potential framework for understanding internal mechanisms of LLMs, especially for scenarios where internal (dis)agreements appear during the inference steps.
[77] Rectifying LLM Thought from Lens of Optimization
Junnan Liu, Hongwei Liu, Songyang Zhang, Kai Chen
Main category: cs.CL
TL;DR: RePro (Rectifying Process-level Reward) is a novel method that refines LLM reasoning by framing chain-of-thought as gradient descent and optimizing reasoning processes during post-training using process-level rewards integrated into RLVR pipelines.
Details
Motivation: Long chain-of-thought LLMs often exhibit suboptimal reasoning behaviors like overthinking and excessively protracted reasoning chains that impair performance, despite their emergent reasoning capabilities.Method: Frames CoT as gradient descent, defines surrogate objective to assess optimization process, uses dual scoring mechanism (intensity and stability), aggregates into composite process-level reward, integrates into RLVR pipelines.
Result: Extensive experiments across multiple RL algorithms and diverse LLMs on math, science, and coding benchmarks show RePro consistently enhances reasoning performance and mitigates suboptimal reasoning behaviors.
Conclusion: RePro effectively optimizes LLM reasoning processes by treating CoT as optimization and using process-level rewards, leading to improved performance and reduced suboptimal reasoning patterns.
Abstract: Recent advancements in large language models (LLMs) have been driven by their emergent reasoning capabilities, particularly through long chain-of-thought (CoT) prompting, which enables thorough exploration and deliberation. Despite these advances, long-CoT LLMs often exhibit suboptimal reasoning behaviors, such as overthinking and excessively protracted reasoning chains, which can impair performance. In this paper, we analyze reasoning processes through an optimization lens, framing CoT as a gradient descent procedure where each reasoning step constitutes an update toward problem resolution. Building on this perspective, we introduce RePro (Rectifying Process-level Reward), a novel approach to refine LLM reasoning during post-training. RePro defines a surrogate objective function to assess the optimization process underlying CoT, utilizing a dual scoring mechanism to quantify its intensity and stability. These scores are aggregated into a composite process-level reward, seamlessly integrated into reinforcement learning with verifiable rewards (RLVR) pipelines to optimize LLMs. Extensive experiments across multiple reinforcement learning algorithms and diverse LLMs, evaluated on benchmarks spanning mathematics, science, and coding, demonstrate that RePro consistently enhances reasoning performance and mitigates suboptimal reasoning behaviors.
[78] How Far Are We from Genuinely Useful Deep Research Agents?
Dingling Zhang, He Zhu, Jincheng Ren, Kangqi Song, Xinran Zhou, Boyu Feng, Shudong Liu, Jiabin Luo, Weihao Xie, Zhaohui Wang, Tianrui Qin, King Zhu, Yuqing Wang, Qianben Chen, Yuchen Eleanor Jiang, Wei Wang, Jiaheng Liu, Wangchunshu Zhou
Main category: cs.CL
TL;DR: FINDER benchmark with 100 human-curated research tasks and 419 checklist items addresses limitations in evaluating Deep Research Agents for comprehensive report generation, while DEFT taxonomy identifies 14 failure modes in current systems.
Details
Motivation: Existing Deep Research Agents are validated on question-answering benchmarks but overlook comprehensive report generation. Current report synthesis benchmarks suffer from task complexity and subjective metrics, failing to reflect user demands and limiting practical utility.Method: Created FINDER benchmark with 100 human-curated research tasks and 419 structured checklist items to standardize report structure, analytical depth, and factual grounding. Developed DEFT failure taxonomy through grounded theory with human-LLM co-annotation and inter-annotator reliability validation, analyzing ~1,000 reports from mainstream DRAs.
Result: DEFT taxonomy identifies 14 fine-grained failure modes across reasoning, retrieval, and generation. Experimental findings show current DRAs struggle with evidence integration, verification, and reasoning-resilient planning rather than task comprehension.
Conclusion: The FINDER benchmark and DEFT taxonomy provide standardized evaluation and diagnostic tools for Deep Research Agents, revealing critical gaps in evidence integration and verification that need to be addressed for practical report generation.
Abstract: Deep Research Agents (DRAs) aim to automatically produce analyst-level reports through iterative information retrieval and synthesis. However, most existing DRAs were validated on question-answering benchmarks, while research on generating comprehensive reports remains overlooked. Worse, current benchmarks for report synthesis suffer from task complexity and subjective metrics – this fails to reflect user demands and limits the practical utility of generated reports. To address these gaps, we present Fine-grained DEepResearch bench (FINDER), an enhanced benchmark consisting of 100 human-curated research tasks with 419 structured checklist items that standardize report structure, analytical depth, and factual grounding. Based on approximately 1,000 reports produced by mainstream DRAs, we further propose Deep rEsearch Failure Taxonomy (DEFT), the first failure taxonomy for deep research agents. DEFT contains 14 fine-grained failure modes across reasoning, retrieval, and generation, and is built upon grounded theory with human-LLM co-annotating and inter-annotator reliability validation. Our experimental findings reveal that current DRAs struggle not with task comprehension but with evidence integration, verification, and reasoning-resilient planning.
[79] The Art of Scaling Test-Time Compute for Large Language Models
Aradhye Agarwal, Ayan Sengupta, Tanmoy Chakraborty
Main category: cs.CL
TL;DR: First large-scale study of test-time scaling (TTS) strategies across 8 LLMs (7B-235B params) and 4 reasoning datasets, revealing no universal TTS winner and providing practical selection guidelines.
Details
Motivation: Lack of systematic comparison of TTS strategies under identical conditions, and unclear influence of model type and problem difficulty on performance.Method: Large-scale study spanning 30B+ tokens using 8 open-source LLMs across 4 reasoning datasets, analyzing TTS strategies under controlled conditions.
Result: Three key findings: 1) No single TTS strategy universally dominates; 2) Reasoning models show distinct trace-quality patterns forming short-horizon and long-horizon categories; 3) Optimal TTS performance scales monotonically with compute budget.
Conclusion: Provides practical recipe for selecting best TTS strategy based on problem difficulty, model type, and compute budget, offering effective inference-time scaling guidance.
Abstract: Test-time scaling (TTS) – the dynamic allocation of compute during inference – is a promising direction for improving reasoning in large language models (LLMs). However, a systematic comparison of well-known TTS strategies under identical conditions is missing, and the influence of model type and problem difficulty on performance remains unclear. To address these gaps, we conduct the first large-scale study of TTS, spanning over thirty billion tokens generated using eight open-source LLMs (7B to 235B parameters), across four reasoning datasets. We observe three consistent trends: (1) no single TTS strategy universally dominates; (2) reasoning models exhibit distinct trace-quality patterns across problem difficulty and trace length, forming short-horizon and long-horizon categories; and (3) for a given model type, the optimal TTS performance scales monotonically with compute budget. Based on these insights, we provide a practical recipe for selecting the best TTS strategy, considering problem difficulty, model type, and compute budget, providing a practical guide to effective inference-time scaling.
[80] Four Over Six: More Accurate NVFP4 Quantization with Adaptive Block Scaling
Jack Cook, Junxian Guo, Guangxuan Xiao, Yujun Lin, Song Han
Main category: cs.CL
TL;DR: 4/6 is a modified NVFP4 quantization algorithm that evaluates two scale factors per block to better represent near-maximal values, preventing training divergence and improving performance compared to standard NVFP4.
Details
Motivation: NVFP4 quantization accelerates LLM training but requires all operands to be quantized, often causing training divergence and performance degradation due to poor representation of near-maximal values in floating-point formats.Method: Proposes 4/6 algorithm that evaluates two potential scale factors per block instead of one, scaling to smaller FP4 values to make representable value distribution more uniform for near-maximal values. Efficiently implementable on NVIDIA Blackwell GPUs.
Result: Prevents training divergence in transformer/hybrid models, brings training loss significantly closer to BF16 baseline compared to state-of-the-art NVFP4 recipes. Also improves downstream accuracy when incorporated into post-training quantization methods.
Conclusion: 4/6 enables more stable NVFP4 training and better inference performance by addressing floating-point quantization’s near-maximal value representation problem, making NVFP4 more viable for LLM training and deployment.
Abstract: As large language models have grown larger, low-precision numerical formats such as NVFP4 have become increasingly popular due to the speed and memory benefits they provide. However, to accelerate computation with NVFP4, all matrix multiplication operands–weights and activations in the forward pass, and weights, activations, and gradients in the backward pass–must be quantized to NVFP4, often leading to divergence during training and performance degradation during inference. NVFP4 by evaluating multiple potential scale factors for each block of values. To address this issue, in this work we introduce Four Over Six (4/6), a modification to the NVFP4 quantization algorithm that evaluates two potential scale factors for each block of values. Unlike integer formats, floating-point formats such as FP4 have the most quantization error on near-maximal values in each block, which we find to be primarily responsible for downstream performance degradation. We find that for some blocks, scaling to smaller FP4 values makes the distribution of representable values more uniform, improving representation of near-maximal values. Importantly, 4/6 can be implemented efficiently on NVIDIA Blackwell GPUs, making it viable to use while training LLMs with NVFP4. In pre-training experiments with transformer and hybrid model architectures, we find that 4/6 prevents divergence in several cases, bringing training loss significantly closer to BF16 compared to models trained with current state-of-the-art NVFP4 training recipes. We also find that 4/6 can be easily incorporated into many different post-training quantization methods and generally improves downstream accuracy. We hope this inspires future work in training and deploying models with NVFP4.
[81] DeID-GPT: Zero-shot Medical Text De-Identification by GPT-4
Zhengliang Liu, Yue Huang, Xiaowei Yu, Lu Zhang, Zihao Wu, Chao Cao, Haixing Dai, Lin Zhao, Yiwei Li, Peng Shu, Fang Zeng, Lichao Sun, Wei Liu, Dinggang Shen, Quanzheng Li, Tianming Liu, Dajiang Zhu, Xiang Li
Main category: cs.CL
TL;DR: Researchers developed DeID-GPT, a GPT-4 enabled framework for de-identifying medical text data, achieving higher accuracy and reliability than existing methods while preserving text structure and meaning.
Details
Motivation: Healthcare digitization enables data sharing but raises privacy concerns. HIPAA requires de-identification before dissemination, but existing computer-assisted methods lack generalizability and need scenario-specific fine-tuning, limiting wider adoption.Method: Developed DeID-GPT, a novel framework leveraging GPT-4’s zero-shot in-context learning capabilities for medical text de-identification. Uses LLMs’ powerful named entity recognition to automatically identify and remove identifying information from unstructured medical text.
Result: DeID-GPT showed the highest accuracy and remarkable reliability compared to existing medical text de-identification methods. It effectively masks private information while preserving original text structure and meaning.
Conclusion: This early study demonstrates GPT-4’s potential for medical text de-identification, providing insights for further LLM research in healthcare. The framework offers an effective solution for HIPAA compliance in medical data sharing.
Abstract: The digitization of healthcare has facilitated the sharing and re-using of medical data but has also raised concerns about confidentiality and privacy. HIPAA (Health Insurance Portability and Accountability Act) mandates removing re-identifying information before the dissemination of medical records. Thus, effective and efficient solutions for de-identifying medical data, especially those in free-text forms, are highly needed. While various computer-assisted de-identification methods, including both rule-based and learning-based, have been developed and used in prior practice, such solutions still lack generalizability or need to be fine-tuned according to different scenarios, significantly imposing restrictions in wider use. The advancement of large language models (LLM), such as ChatGPT and GPT-4, have shown great potential in processing text data in the medical domain with zero-shot in-context learning, especially in the task of privacy protection, as these models can identify confidential information by their powerful named entity recognition (NER) capability. In this work, we developed a novel GPT4-enabled de-identification framework (``DeID-GPT") to automatically identify and remove the identifying information. Compared to existing commonly used medical text data de-identification methods, our developed DeID-GPT showed the highest accuracy and remarkable reliability in masking private information from the unstructured medical text while preserving the original structure and meaning of the text. This study is one of the earliest to utilize ChatGPT and GPT-4 for medical text data processing and de-identification, which provides insights for further research and solution development on the use of LLMs such as ChatGPT/GPT-4 in healthcare. Codes and benchmarking data information are available at https://github.com/yhydhx/ChatGPT-API.
[82] Template-assisted Contrastive Learning of Task-oriented Dialogue Sentence Embeddings
Minsik Oh, Jiwei Li, Guoyin Wang
Main category: cs.CL
TL;DR: TaDSE is a novel method that uses template information to learn better dialogue sentence embeddings through self-supervised contrastive learning, achieving state-of-the-art results on dialogue benchmarks.
Details
Motivation: Learning high-quality sentence embeddings from dialogues is essential for various dialogue tasks but difficult due to the challenge of annotating utterance relationships. Token-level annotations (entities, slots, templates) are easier to obtain than sentence-level annotations, but existing sentence embedding methods cannot effectively utilize this token-level knowledge.Method: Template-aware Dialogue Sentence Embedding (TaDSE) uses template information to learn utterance embeddings via self-supervised contrastive learning. The method includes synthetic data augmentation that diversifies utterance-template associations, with slot-filling as a preliminary step to enhance the learning process.
Result: TaDSE achieves significant improvements over previous state-of-the-art methods on five downstream benchmark dialogue datasets. The paper also introduces a novel semantic compression test as an analytic instrument and discovers correlations with uniformity and alignment metrics.
Conclusion: TaDSE effectively leverages template information to learn superior dialogue sentence embeddings, demonstrating the value of utilizing token-level annotations in self-supervised learning frameworks for dialogue understanding tasks.
Abstract: Learning high quality sentence embeddings from dialogues has drawn increasing attentions as it is essential to solve a variety of dialogue-oriented tasks with low annotation cost. Annotating and gathering utterance relationships in conversations are difficult, while token-level annotations, \eg, entities, slots and templates, are much easier to obtain. Other sentence embedding methods are usually sentence-level self-supervised frameworks and cannot utilize token-level extra knowledge. We introduce Template-aware Dialogue Sentence Embedding (TaDSE), a novel augmentation method that utilizes template information to learn utterance embeddings via self-supervised contrastive learning framework. We further enhance the effect with a synthetically augmented dataset that diversifies utterance-template association, in which slot-filling is a preliminary step. We evaluate TaDSE performance on five downstream benchmark dialogue datasets. The experiment results show that TaDSE achieves significant improvements over previous SOTA methods for dialogue. We further introduce a novel analytic instrument of semantic compression test, for which we discover a correlation with uniformity and alignment. Our code will be released upon acceptance.
[83] Extending Multilingual Machine Translation through Imitation Learning
Wen Lai, Viktor Hangya, Yingli Shen, Alexander Fraser
Main category: cs.CL
TL;DR: Imit-MNMT: A novel imitation learning approach for extending multilingual NMT models to new languages using only English-parallel data, preventing catastrophic forgetting.
Details
Motivation: Most world languages are unsupported by current multilingual NMT models, and extending models to new languages typically requires parallel data with all existing languages, which is impractical. Existing methods using only English-parallel data suffer from catastrophic forgetting of previously learned languages.Method: Proposes Imit-MNMT, treating language extension as imitation learning. Uses an expert model to generate pseudo-parallel corpora between new language and existing languages. Implements data distribution imitation with language-specific weighting and translation behavior imitation mechanisms.
Result: Extensive experiments show significant improvement in translation performance between new and existing languages while effectively mitigating catastrophic forgetting compared to previous methods.
Conclusion: Imit-MNMT successfully extends multilingual NMT models to new languages using only English-parallel data, addressing the critical problem of catastrophic forgetting through innovative imitation learning techniques.
Abstract: Despite the growing variety of languages supported by existing multilingual neural machine translation (MNMT) models, most of the world’s languages are still being left behind. We aim to extend large-scale MNMT models to incorporate a new language, enabling translations between this new language and all previously supported languages, even in the challenging scenario where only a parallel corpus between the new language and English is available. Previous methods, such as continued training on parallel data including the new language, often suffer from catastrophic forgetting, which degrades performance on other languages. We propose a novel approach Imit-MNMT which treats this task as an imitation learning problem, a technique widely used in computer vision but less explored in natural language processing. Specifically, we leverage an expert model to generate pseudo-parallel corpora between the new language and the existing languages. We then introduce a data distribution imitation strategy using language-specific weighting, alongside a translation behavior imitation mechanism. Extensive experiments show that our approach significantly improves translation performance between the new and existing languages while mitigating catastrophic forgetting.
[84] Reliable Reasoning Beyond Natural Language
Nasim Borazjanizadeh, Steven T. Piantadosi
Main category: cs.CL
TL;DR: LLMs struggle with non-linear reasoning, so researchers created NLR dataset and a neurosymbolic approach combining LLMs with Prolog for robust reasoning gains.
Details
Motivation: Large Language Models (LLMs) have linguistic competence but struggle with reliable and flexible reasoning, particularly with non-linear reasoning tasks that require iterative updates, backtracking, and parallel chains of thought.Method: Proposed a neurosymbolic reasoning approach that integrates Prolog (symbolic reasoning engine) into LLM inference pipeline. LLMs infer all information (explicit and implied through common sense) and encode it as logical code, while Prolog handles the iterative computations.
Result: Achieved large and robust performance gains across GSM8k and BIG-bench Navigate benchmarks. Near-perfect accuracy on NLR problems, maintaining robustness even as variable interdependence increases.
Conclusion: Neurosymbolic integration of LLMs with symbolic reasoning engines like Prolog effectively addresses LLMs’ reasoning limitations, particularly for non-linear reasoning tasks requiring iterative updates and parallel reasoning chains.
Abstract: Despite their linguistic competence, Large Language Models (LLMs) often struggle to reason reliably and flexibly. To identify these shortcomings, we introduce the Non-Linear Reasoning (NLR) dataset, a collection of 55 unique, hand-designed problems that target reasoning bottlenecks arising from the sequential prediction paradigm of LLMs and the inherently linear nature of natural language. NLR tasks require iterative updates, backtracking, and reasoning across multiple parallel chains of thought but only basic arithmetic to solve. To address these limitations, we propose a neurosymbolic reasoning approach that integrates Prolog, a symbolic reasoning engine, into the inference pipeline of LLMs. This division of labor shifts the LLM’s task from iterative computations to inferring all information, explicit or implied through common sense, and encoding it as logical code. Our method yields large and robust performance gains across the GSM8k and BIG-bench Navigate benchmarks and achieves near-perfect accuracy on NLR problems, maintaining robustness even as variable interdependence - the number of other variables on which the value of a single variable depends - increases.
[85] HEALTH-PARIKSHA: Assessing RAG Models for Health Chatbots in Real-World Multilingual Settings
Varun Gumma, Ananditha Raghunath, Mohit Jain, Sunayana Sitaram
Main category: cs.CL
TL;DR: This study evaluates 24 LLMs on real-world Indian patient queries in English and 4 Indic languages, finding significant performance variations, lower factual correctness for Indic queries, and challenges with code-mixed/cultural content.
Details
Motivation: There's a need to evaluate LLMs in real-world multilingual scenarios beyond translated benchmarks, as current evaluations often miss linguistic and cultural nuances, especially for languages like those used in India.Method: Used real-world data from Indian patients interacting with a medical chatbot in Indian English and 4 Indic languages. Employed a uniform Retrieval Augmented Generation framework to generate responses, evaluated using both automated techniques and human evaluators on four specific metrics.
Result: Models showed significant performance variations; instruction-tuned Indic models didn’t always perform well on Indic language queries; factual correctness was generally lower for Indic queries compared to English queries; code-mixed and culturally relevant queries posed challenges to models.
Conclusion: Real-world multilingual evaluation reveals important limitations in current LLMs, especially for Indic languages, highlighting the need for better handling of linguistic nuances, cultural context, and code-mixed content in medical applications.
Abstract: Assessing the capabilities and limitations of large language models (LLMs) has garnered significant interest, yet the evaluation of multiple models in real-world scenarios remains rare. Multilingual evaluation often relies on translated benchmarks, which typically do not capture linguistic and cultural nuances present in the source language. This study provides an extensive assessment of 24 LLMs on real world data collected from Indian patients interacting with a medical chatbot in Indian English and 4 other Indic languages. We employ a uniform Retrieval Augmented Generation framework to generate responses, which are evaluated using both automated techniques and human evaluators on four specific metrics relevant to our application. We find that models vary significantly in their performance and that instruction tuned Indic models do not always perform well on Indic language queries. Further, we empirically show that factual correctness is generally lower for responses to Indic queries compared to English queries. Finally, our qualitative work shows that code-mixed and culturally relevant queries in our dataset pose challenges to evaluated models.
[86] BoundingDocs: a Unified Dataset for Document Question Answering with Spatial Annotations
Simone Giovannini, Fabio Coppini, Andrea Gemelli, Simone Marinai
Main category: cs.CL
TL;DR: A unified document QA dataset combining multiple public datasets, reformulating Document AI tasks into QA format with OCR and bounding box annotations for LLM training and evaluation.
Details
Motivation: To create a comprehensive resource for training and evaluating Large Language Models on document understanding tasks by unifying existing datasets and reformulating Information Extraction and other Document AI tasks into a Question-Answering format.Method: Combined several public Document AI and VRDU datasets, reformulated tasks into QA format, released OCR of all documents, included bounding box annotations for answer locations, and explored different prompting techniques (with/without bounding box information) on open-weight models.
Result: Created a unified document QA dataset with OCR and spatial annotations, and identified the most effective prompting approaches for document comprehension through experimentation with different techniques.
Conclusion: The dataset provides a valuable resource for LLM training and evaluation on document understanding tasks, and the study reveals optimal prompting strategies that leverage spatial information for improved document comprehension performance.
Abstract: We present a unified dataset for document Question-Answering (QA), which is obtained combining several public datasets related to Document AI and visually rich document understanding (VRDU). Our main contribution is twofold: on the one hand we reformulate existing Document AI tasks, such as Information Extraction (IE), into a Question-Answering task, making it a suitable resource for training and evaluating Large Language Models; on the other hand, we release the OCR of all the documents and include the exact position of the answer to be found in the document image as a bounding box. Using this dataset, we explore the impact of different prompting techniques (that might include bounding box information) on the performance of open-weight models, identifying the most effective approaches for document comprehension.
[87] LLM-based Human Simulations Have Not Yet Been Reliable
Qian Wang, Jiaying Wu, Zichen Jiang, Zhenheng Tang, Bingqiao Luo, Nuo Chen, Wei Chen, Bingsheng He
Main category: cs.CL
TL;DR: LLM-based human simulations currently lack reliability due to significant discrepancies from real human behavior, requiring systematic improvements in data, LLM capabilities, and simulation design.
Details
Motivation: LLMs are increasingly used to simulate human behaviors across various domains, but current simulations show significant discrepancies from authentic human actions, indicating insufficient reliability that needs to be addressed.Method: Systematic review of LLM-based human simulations in social, economic, policy, and psychological contexts to identify common frameworks, advances, and limitations. Analysis reveals problems stem from LLM limitations and flawed simulation design. Proposes solution framework focusing on data enrichment, LLM advancement, and robust simulation design.
Result: Identified that discrepancies in LLM-based human simulations primarily come from inherent LLM limitations and flawed simulation design. Proposed a systematic solution framework and structured algorithm to enhance reliability and human alignment.
Conclusion: Current LLM-based human simulations are insufficiently reliable, requiring systematic improvements in data foundations, LLM capabilities, and simulation design. The proposed framework and algorithm aim to guide more credible and human-aligned simulations, with resources provided for further research.
Abstract: Large Language Models (LLMs) are increasingly employed for simulating human behaviors across diverse domains. However, our position is that current LLM-based human simulations remain insufficiently reliable, as evidenced by significant discrepancies between their outcomes and authentic human actions. Our investigation begins with a systematic review of LLM-based human simulations in social, economic, policy, and psychological contexts, identifying their common frameworks, recent advances, and persistent limitations. This review reveals that such discrepancies primarily stem from inherent limitations of LLMs and flaws in simulation design, both of which are examined in detail. Building on these insights, we propose a systematic solution framework that emphasizes enriching data foundations, advancing LLM capabilities, and ensuring robust simulation design to enhance reliability. Finally, we introduce a structured algorithm that operationalizes the proposed framework, aiming to guide credible and human-aligned LLM-based simulations. To facilitate further research, we provide a curated list of related literature and resources at https://github.com/Persdre/awesome-llm-human-simulation.
[88] Generating Text from Uniform Meaning Representation
Emma Markle, Reihaneh Iranmanesh, Shira Wein
Main category: cs.CL
TL;DR: This paper investigates approaches for generating text from multilingual UMR graphs, leveraging AMR technologies and achieving promising results with fine-tuning methods.
Details
Motivation: To develop a UMR technological ecosystem for downstream tasks by creating effective methods for text generation from UMR graphs, despite limited available UMR annotations.Method: Three approaches: (1) baseline using AMR-to-text models directly on UMR graphs, (2) pipeline conversion of UMR to AMR then using AMR-to-text models, and (3) fine-tuning foundation models and AMR-to-text models with UMR data.
Result: Best models achieved multilingual BERTscores of 0.825 for English and 0.882 for Chinese, showing effectiveness of fine-tuning approaches even with limited UMR data.
Conclusion: Fine-tuning approaches show promise for UMR-to-text generation, and leveraging AMR technologies provides a practical pathway for developing UMR applications despite data limitations.
Abstract: Uniform Meaning Representation (UMR) is a recently developed graph-based semantic representation, which expands on Abstract Meaning Representation (AMR) in a number of ways, in particular through the inclusion of document-level information and multilingual flexibility. In order to effectively adopt and leverage UMR for downstream tasks, efforts must be placed toward developing a UMR technological ecosystem. Though only a small amount of UMR annotations have been produced to date, in this work, we investigate the first approaches to producing text from multilingual UMR graphs. Exploiting the structural similarity between UMR and AMR graphs and the wide availability of AMR technologies, we introduce (1) a baseline approach which passes UMR graphs to AMR-to-text generation models, (2) a pipeline conversion of UMR to AMR, then using AMR-to-text generation models, and (3) a fine-tuning approach for both foundation models and AMR-to-text generation models with UMR data. Our best performing models achieve multilingual BERTscores of 0.825 for English and 0.882 for Chinese, a promising indication of the effectiveness of fine-tuning approaches for UMR-to-text generation even with limited UMR data.
[89] Med-gte-hybrid: A contextual embedding transformer model for extracting actionable information from clinical texts
Aditya Kumar, Simon Rauch, Mario Cypko, Oliver Amft
Main category: cs.CL
TL;DR: The paper introduces med-gte-hybrid, a contextual embedding model derived from gte-large sentence transformer for extracting information from clinical narratives, using a hybrid tuning strategy combining contrastive learning and denoising autoencoder, showing improved performance on clinical prediction tasks and text embedding benchmarks.
Details
Motivation: To develop an improved contextual embedding model specifically for clinical narratives that can better extract information from unstructured medical text, enabling more accurate clinical predictions and decision support.Method: Derived med-gte-hybrid from gte-large sentence transformer using a hybrid tuning strategy combining contrastive learning and denoising autoencoder. Evaluated on clinical prediction tasks using MIMIC-IV dataset including CKD prognosis, eGFR prediction, and mortality prediction.
Result: Med-gte-hybrid outperforms current state-of-the-art models on the Massive Text Embedding Benchmark (MTEB) and improves patient stratification, clustering, and text retrieval. Shows strong performance on clinical prediction tasks using MIMIC-IV data.
Conclusion: The hybrid tuning approach for sentence transformers can be transferred to other medical domains and has potential to improve clinical decision-making and personalized treatment pathways in various healthcare applications.
Abstract: We introduce a novel contextual embedding model med-gte-hybrid that was derived from the gte-large sentence transformer to extract information from unstructured clinical narratives. Our model tuning strategy for med-gte-hybrid combines contrastive learning and a denoising autoencoder. To evaluate the performance of med-gte-hybrid, we investigate several clinical prediction tasks in large patient cohorts extracted from the MIMIC-IV dataset, including Chronic Kidney Disease (CKD) patient prognosis, estimated glomerular filtration rate (eGFR) prediction, and patient mortality prediction. Furthermore, we demonstrate that the med-gte-hybrid model improves patient stratification, clustering, and text retrieval, thus outperforms current state-of-the-art models on the Massive Text Embedding Benchmark (MTEB). While some of our evaluations focus on CKD, our hybrid tuning of sentence transformers could be transferred to other medical domains and has the potential to improve clinical decision-making and personalised treatment pathways in various healthcare applications.
[90] MimeQA: Towards Socially-Intelligent Nonverbal Foundation Models
Hengzhi Li, Megan Tjandrasuwita, Yi R. Fung, Armando Solar-Lezama, Paul Pu Liang
Main category: cs.CL
TL;DR: MimeQA dataset introduces video-based nonverbal social reasoning benchmark using mime videos, revealing VideoLLMs’ poor performance (20-30%) vs humans (86%) due to failure to ground imagined objects and over-reliance on text prompts.
Details
Motivation: Current AI social reasoning relies on language-only approaches, leading to systems good at verbal communication but poor at nonverbal social understanding. The authors aim to address this gap by leveraging nonverbal social interactions in mime videos.Method: Created MimeQA dataset with ~8 hours of mime video clips from YouTube, annotated with 806 question-answer pairs to probe nonverbal social reasoning. Evaluated state-of-the-art VideoLLMs on this benchmark.
Result: VideoLLMs achieved only 20-30% accuracy on MimeQA, while humans scored 86%. Analysis showed VideoLLMs fail to ground imagined objects and over-rely on text prompts while ignoring subtle nonverbal interactions.
Conclusion: The paper highlights the limitations of current VideoLLMs in nonverbal social reasoning and introduces MimeQA as a benchmark to inspire development of AI models with true social intelligence capable of interpreting nonverbal human interactions.
Abstract: As AI becomes more closely integrated with peoples’ daily activities, socially intelligent AI that can understand and interact seamlessly with humans in daily lives is increasingly important. However, current works in AI social reasoning all rely on language-only or language-dominant approaches to benchmark and training models, resulting in systems that are improving in verbal communication but struggle with nonverbal social understanding. To address this limitation, we tap into a novel data source rich in nonverbal social interactions – mime videos. Mimes refer to the art of expression through gesture and movement without spoken words, which presents unique challenges and opportunities in interpreting nonverbal social communication. We contribute a new dataset called MimeQA, obtained by sourcing ~8 hours of videos clips from YouTube and developing a comprehensive video question-answering benchmark comprising 806 carefully annotated and verified question-answer pairs, designed to probe nonverbal social reasoning capabilities. Using MimeQA, we evaluate state-of-the-art video large language models (VideoLLMs) and find that they achieve low accuracy, generally ranging from 20-30%, while humans score 86%. Our analysis reveals that VideoLLMs often fail to ground imagined objects and over-rely on the text prompt while ignoring subtle nonverbal interactions. We hope to inspire future work in AI models that embody true social intelligence capable of interpreting non-verbal human interactions.
[91] LLM-based Automated Grading with Human-in-the-Loop
Yucheng Chu, Hang Li, Kaiqi Yang, Yasemin Copur-Gencturk, Jiliang Tang
Main category: cs.CL
TL;DR: GradeHITL: A human-in-the-loop LLM framework for automatic short answer grading that uses LLMs to ask questions to human experts, dynamically refining rubrics to achieve near-human grading accuracy.
Details
Motivation: While LLMs have improved automatic short answer grading (ASAG), existing methods still fall short of human-level performance in rubric-based assessments due to their fully automated nature. The paper aims to bridge this gap by leveraging LLMs' interactive capabilities.Method: Proposes GradeHITL framework that uses LLMs’ generative properties to pose questions to human experts, incorporating their insights to dynamically refine grading rubrics through a human-in-the-loop approach.
Result: The adaptive human-in-the-loop process significantly improves grading accuracy, outperforming existing methods and bringing ASAG closer to human-level evaluation performance.
Conclusion: Human-in-the-loop approaches using LLMs’ interactive capabilities can effectively enhance automatic short answer grading, achieving near-human performance by dynamically incorporating expert insights into rubric refinement.
Abstract: The rise of artificial intelligence (AI) technologies, particularly large language models (LLMs), has brought significant advancements to the field of education. Among various applications, automatic short answer grading (ASAG), which focuses on evaluating open-ended textual responses, has seen remarkable progress with the introduction of LLMs. These models not only enhance grading performance compared to traditional ASAG approaches but also move beyond simple comparisons with predefined “golden” answers, enabling more sophisticated grading scenarios, such as rubric-based evaluation. However, existing LLM-powered methods still face challenges in achieving human-level grading performance in rubric-based assessments due to their reliance on fully automated approaches. In this work, we explore the potential of LLMs in ASAG tasks by leveraging their interactive capabilities through a human-in-the-loop (HITL) approach. Our proposed framework, GradeHITL, utilizes the generative properties of LLMs to pose questions to human experts, incorporating their insights to refine grading rubrics dynamically. This adaptive process significantly improves grading accuracy, outperforming existing methods and bringing ASAG closer to human-level evaluation.
[92] A Method for Handling Negative Similarities in Explainable Graph Spectral Clustering of Text Documents – Extended Version
Mieczysław A. Kłopotek, Sławomir T. Wierzchoń, Bartłomiej Starosta, Dariusz Czerski, Piotr Borkowski
Main category: cs.CL
TL;DR: This paper examines Graph Spectral Clustering with negative similarities from modern document embeddings (doc2vec, GloVe), evaluates 6 solutions for combinatorial/normalized Laplacians, and shows that curing negative similarities improves accuracy and enables explanation methods for GloVe embeddings.
Details
Motivation: Traditional Graph Spectral Clustering (GSC) assumes non-negative similarities, but modern document embeddings like doc2vec and GloVe can produce negative similarities, causing failures in normalized Laplacian-based GSC. The paper aims to address this problem and find effective solutions.Method: The paper investigates 6 different solutions from literature and new research for handling negative similarities in GSC. It analyzes both combinatorial Laplacians and normalized Laplacians, and experimentally compares these approaches on document embeddings.
Result: GloVe embeddings frequently cause failures of normalized Laplacian-based GSC due to negative similarities. Applying methods that cure similarity negativity improves accuracy for both combinatorial and normalized Laplacian-based GSC, and enables application of explanation methods originally developed for Term Vector Space embeddings to GloVe embeddings.
Conclusion: Negative similarities from modern document embeddings pose significant challenges for GSC, particularly for normalized Laplacian approaches. The proposed solutions effectively address these issues, improving clustering accuracy and extending the applicability of explanation methods to newer embedding techniques like GloVe.
Abstract: This paper investigates the problem of Graph Spectral Clustering with negative similarities, resulting from document embeddings different from the traditional Term Vector Space (like doc2vec, GloVe, etc.). Solutions for combinatorial Laplacians and normalized Laplacians are discussed. An experimental investigation shows the advantages and disadvantages of 6 different solutions proposed in the literature and in this research. The research demonstrates that GloVe embeddings frequently cause failures of normalized Laplacian based GSC due to negative similarities. Furthermore, application of methods curing similarity negativity leads to accuracy improvement for both combinatorial and normalized Laplacian based GSC. It also leads to applicability for GloVe embeddings of explanation methods developed originally bythe authors for Term Vector Space embeddings.
[93] Nemotron-CLIMB: CLustering-based Iterative Data Mixture Bootstrapping for Language Model Pre-training
Shizhe Diao, Yu Yang, Yonggan Fu, Xin Dong, Dan Su, Markus Kliegl, Zijia Chen, Peter Belcak, Yoshi Suhara, Hongxu Yin, Mostofa Patwary, Yingyan, Lin, Jan Kautz, Pavlo Molchanov
Main category: cs.CL
TL;DR: Nemotron-CLIMB is an automated framework that discovers optimal pre-training data mixtures through semantic clustering and iterative search, achieving SOTA performance with 2% improvement over Llama-3.2-1B.
Details
Motivation: Pre-training datasets lack domain labels and manual curation is labor-intensive, making optimal data mixture identification challenging despite its significant impact on pre-training performance.Method: CLIMB embeds and clusters large-scale datasets in semantic space, then iteratively searches for optimal mixtures using a smaller proxy model and predictor to discover effective data combinations.
Result: 1B model trained on 400B tokens with CLIMB mixture exceeds Llama-3.2-1B by 2.0%, with domain-specific optimization achieving 5% improvement over random sampling. Also released Nemotron-ClimbLab (1.2T tokens, 20 clusters) and Nemotron-ClimbMix (400B tokens).
Conclusion: Automated data mixture discovery through semantic clustering and iterative search effectively optimizes pre-training performance, with released datasets enabling further research in optimal data composition.
Abstract: Pre-training datasets are typically collected from web content and lack inherent domain divisions. For instance, widely used datasets like Common Crawl do not include explicit domain labels, while manually curating labeled datasets such as The Pile is labor-intensive. Consequently, identifying an optimal pre-training data mixture remains a challenging problem, despite its significant benefits for pre-training performance. To address these challenges, we propose CLustering-based Iterative Data Mixture Bootstrapping (Nemotron-CLIMB), an automated framework that discovers, evaluates, and refines data mixtures in a pre-training setting. Specifically, Nemotron-CLIMB embeds and clusters large-scale datasets in a semantic space and then iteratively searches for optimal mixtures using a smaller proxy model and a predictor. When continuously trained on 400B tokens with this mixture, our 1B model exceeds the state-of-the-art Llama-3.2-1B by 2.0%. Moreover, we observe that optimizing for a specific domain (e.g., Social Sciences) yields a 5% improvement over random sampling. Finally, we introduce Nemotron-ClimbLab, a filtered 1.2-trillion-token corpus with 20 clusters as a research playground, and Nemotron-ClimbMix, a compact yet powerful 400-billion-token dataset designed for efficient pre-training that delivers superior performance under an equal token budget. We analyze the final data mixture, elucidating the characteristics of an optimal data mixture. Our data is available at: https://research.nvidia.com/labs/lpr/climb/
[94] Do different prompting methods yield a common task representation in language models?
Guy Davidson, Todd M. Gureckis, Brenden M. Lake, Adina Williams
Main category: cs.CL
TL;DR: Function vectors extracted from demonstrations vs. instructions show different task representations in LLMs, suggesting they leverage different model mechanisms rather than a common task representation.
Details
Motivation: To understand whether identical tasks presented through demonstrations vs. instructions result in similar task representations in language models, and to gain interpretability insights into how LLMs represent tasks differently based on prompting format.Method: Generalized function vectors (FVs) to extract task representations from alternative task presentations, focusing on short textual instruction prompts. Extracted instruction function vectors and compared them with demonstration-based FVs. Used controls to dissociate their contributions to task performance.
Result: Successfully extracted instruction function vectors that promote zero-shot task accuracy. Found evidence that demonstration- and instruction-based function vectors leverage different model components. Different task prompting forms do not induce a common task representation through FVs but elicit different, partly overlapping mechanisms.
Conclusion: Different prompting formats (demonstrations vs. instructions) create distinct task representations in LLMs, supporting the practice of combining both approaches. This presents challenges for universally monitoring task inference across presentation forms and encourages further examination of LLM task inference mechanisms.
Abstract: Demonstrations and instructions are two primary approaches for prompting language models to perform in-context learning (ICL) tasks. Do identical tasks elicited in different ways result in similar representations of the task? An improved understanding of task representation mechanisms would offer interpretability insights and may aid in steering models. We study this through \textit{function vectors} (FVs), recently proposed as a mechanism to extract few-shot ICL task representations. We generalize FVs to alternative task presentations, focusing on short textual instruction prompts, and successfully extract instruction function vectors that promote zero-shot task accuracy. We find evidence that demonstration- and instruction-based function vectors leverage different model components, and offer several controls to dissociate their contributions to task performance. Our results suggest that different task promptings forms do not induce a common task representation through FVs but elicit different, partly overlapping mechanisms. Our findings offer principled support to the practice of combining instructions and task demonstrations, imply challenges in universally monitoring task inference across presentation forms, and encourage further examinations of LLM task inference mechanisms.
[95] Extracting memorized pieces of (copyrighted) books from open-weight language models
A. Feder Cooper, Aaron Gokaslan, Ahmed Ahmed, Amy B. Cyphert, Christopher De Sa, Mark A. Lemley, Daniel E. Ho, Percy Liang
Main category: cs.CL
TL;DR: LLM memorization of copyrighted books varies significantly by model and book, with most models not memorizing most books, but some models like Llama 3.1 70B completely memorizing specific works like Harry Potter and 1984.
Details
Motivation: To address polarized claims in copyright lawsuits about whether LLMs memorize copyrighted training data, by providing empirical evidence on actual memorization patterns across different models and books.Method: Extended probabilistic extraction technique to measure memorization of 50 books in 17 open-weight LLMs, conducting thousands of experiments to assess memorization patterns.
Result: Most LLMs do not memorize most books, but Llama 3.1 70B entirely memorizes some books (e.g., Harry Potter, 1984) - can generate entire books near-verbatim from minimal prompts.
Conclusion: Memorization varies significantly by model and book, with implications for copyright cases that don’t unambiguously favor either plaintiffs or defendants.
Abstract: Plaintiffs and defendants in copyright lawsuits over generative AI often make sweeping, opposing claims about the extent to which large language models (LLMs) have memorized plaintiffs’ protected expression in their training data. Drawing on both machine learning and copyright law, we show that these polarized positions dramatically oversimplify the relationship between memorization and copyright. To do so, we extend a recent probabilistic extraction technique to measure memorization of 50 books in 17 open-weight LLMs. Through thousands of experiments, we show that the extent of memorization varies both by model and by book. With respect to our specific extraction methodology, we find that most LLMs do not memorize most books – either in whole or in part. However, we also find that Llama 3.1 70B entirely memorizes some books, like the first Harry Potter book and 1984. In fact, the first Harry Potter is so memorized that, using a seed prompt consisting of just the first few tokens of the first chapter, we can deterministically generate the entire book near-verbatim. We discuss why our results have significant implications for copyright cases, though not ones that unambiguously favor either side.
[96] A Token is Worth over 1,000 Tokens: Efficient Knowledge Distillation through Low-Rank Clone
Jitai Hao, Qiang Huang, Hao Liu, Xinyan Xiao, Zhaochun Ren, Jun Yu
Main category: cs.CL
TL;DR: LRC is an efficient pre-training method that creates small language models by using low-rank projection matrices to softly prune teacher weights and clone teacher activations, achieving high performance with dramatically reduced training tokens.
Details
Motivation: Existing methods for training small language models suffer from information loss from hard pruning, inefficient representation alignment, and underutilization of informative activations (especially from FFNs), making training costly even with knowledge distillation.Method: Low-Rank Clone (LRC) trains low-rank projection matrices that perform soft pruning by compressing teacher weights and activation cloning by aligning student activations (including FFN signals) with teacher activations, eliminating the need for explicit alignment modules.
Result: LRC matches or surpasses state-of-the-art models trained on trillions of tokens while using only 20B tokens, achieving over 1,000x training efficiency with open-source teachers like Llama-3.2-3B-Instruct and Qwen2.5-3B/7B-Instruct.
Conclusion: LRC provides an efficient unified approach for training high-performing small language models that maximizes knowledge transfer from teacher models while dramatically reducing training costs.
Abstract: Training high-performing Small Language Models (SLMs) remains costly, even with knowledge distillation and pruning from larger teacher models. Existing work often faces three key challenges: (1) information loss from hard pruning, (2) inefficient alignment of representations, and (3) underutilization of informative activations, particularly from Feed-Forward Networks (FFNs). To address these challenges, we introduce Low-Rank Clone (LRC), an efficient pre-training method that constructs SLMs aspiring to behavioral equivalence with strong teacher models. LRC trains a set of low-rank projection matrices that jointly enable soft pruning by compressing teacher weights, and activation clone by aligning student activations, including FFN signals, with those of the teacher. This unified design maximizes knowledge transfer while removing the need for explicit alignment modules. Extensive experiments with open-source teachers (e.g., Llama-3.2-3B-Instruct, Qwen2.5-3B/7B-Instruct) show that LRC matches or surpasses state-of-the-art models trained on trillions of tokens–while using only 20B tokens, achieving over 1,000x training efficiency. Our codes and model checkpoints are available at https://github.com/CURRENTF/LowRankClone and https://huggingface.co/collections/JitaiHao/low-rank-clone-lrc-6828389e96a93f1d4219dfaf.
[97] Attributional Safety Failures in Large Language Models under Code-Mixed Perturbations
Somnath Banerjee, Pratyush Chatterjee, Shanu Kumar, Sayan Layek, Parag Agrawal, Rima Hazra, Animesh Mukherjee
Main category: cs.CL
TL;DR: Code-mixing (blending languages) causes catastrophic safety failures in LLMs, with attack success rates jumping from 9% in English to 69% in code-mixed inputs, and over 90% in non-Western languages like Arabic and Hindi.
Details
Motivation: While LLMs appear robustly safety-aligned in English, there's an overlooked weakness: attributional collapse under code-mixed perturbations. Code-mixing (blending languages within a single conversation) can cause safety guardrails to fail dramatically, revealing serious risks for billions of users worldwide.Method: 1) Systematic evaluation of open models on code-mixed inputs; 2) Testing on both controlled synthetic datasets and real-world social media traces; 3) Introducing saliency drift attribution (SDA) interpretability framework to analyze attention drift; 4) Proposing a lightweight translation-based restoration strategy to recover safety.
Result: Attack success rates spike from 9% in monolingual English to 69% under code-mixed inputs, with rates exceeding 90% in non-Western contexts (Arabic, Hindi). SDA reveals that under code-mixing, model attention drifts away from safety-critical tokens, blinding the model to harmful intent. The proposed restoration strategy recovers roughly 80% of lost safety.
Conclusion: Code-mixing presents a catastrophic safety weakness in LLMs that disproportionately affects non-Western language users. The saliency drift attribution framework explains the mechanism, and translation-based restoration offers a practical mitigation path toward more equitable and robust LLM safety.
Abstract: While LLMs appear robustly safety-aligned in English, we uncover a catastrophic, overlooked weakness: attributional collapse under code-mixed perturbations. Our systematic evaluation of open models shows that the linguistic camouflage of code-mixing – blending languages within a single conversation'' -- can cause safety guardrails to fail dramatically. Attack success rates (ASR) spike from a benign 9\% in monolingual English to 69\% under code-mixed inputs, with rates exceeding 90\% in non-Western contexts such as Arabic and Hindi. These effects hold not only on controlled synthetic datasets but also on real-world social media traces, revealing a serious risk for billions of users. To explain why this happens, we introduce saliency drift attribution (SDA), an interpretability framework that shows how, under code-mixing, the model's internal attention drifts away from safety-critical tokens (e.g., violence’’ or ``corruption’’), effectively blinding it to harmful intent. Finally, we propose a lightweight translation-based restoration strategy that recovers roughly 80% of the safety lost to code-mixing, offering a practical path toward more equitable and robust LLM safety.
[98] Comprehensive Evaluation on Lexical Normalization: Boundary-Aware Approaches for Unsegmented Languages
Shohei Higashiyama, Masao Utiyama
Main category: cs.CL
TL;DR: This paper addresses lexical normalization for informal text in unsegmented languages, creating a Japanese dataset, developing methods based on pretrained models, and conducting multi-perspective evaluations showing both encoder-only and decoder-only approaches perform well.
Details
Motivation: The motivation is to address the lack of comprehensive evaluations in lexical normalization research for informal expressions in user-generated text, particularly for unsegmented languages like Japanese, where it's unclear which methods perform best across multiple perspectives.Method: The method involves three key contributions: (1) creating a large-scale, multi-domain Japanese normalization dataset, (2) developing normalization methods based on state-of-the-art pretrained models, and (3) conducting experiments across multiple evaluation perspectives to compare encoder-only and decoder-only approaches.
Result: Experiments show that both encoder-only and decoder-only approaches achieve promising results in both accuracy and efficiency for Japanese lexical normalization.
Conclusion: The paper demonstrates that modern pretrained models (both encoder-only and decoder-only architectures) can effectively handle lexical normalization tasks for unsegmented languages like Japanese, providing a comprehensive evaluation framework and dataset for future research.
Abstract: Lexical normalization research has sought to tackle the challenge of processing informal expressions in user-generated text, yet the absence of comprehensive evaluations leaves it unclear which methods excel across multiple perspectives. Focusing on unsegmented languages, we make three key contributions: (1) creating a large-scale, multi-domain Japanese normalization dataset, (2) developing normalization methods based on state-of-the-art pretrained models, and (3) conducting experiments across multiple evaluation perspectives. Our experiments show that both encoder-only and decoder-only approaches achieve promising results in both accuracy and efficiency.
[99] Eye of Judgement: Dissecting the Evaluation of Russian-speaking LLMs with POLLUX
Nikita Martynov, Anastasia Mordasheva, Dmitriy Gorbetskiy, Danil Astafurov, Ulyana Isaeva, Elina Basyrova, Sergey Skachkov, Victoria Berestova, Nikolay Ivanov, Valeriia Zanina, Alena Fenogenova
Main category: cs.CL
TL;DR: POLLUX is an open-source Russian LLM benchmark with 2,100 prompts across 35 task types, featuring a novel LLM-as-a-Judge evaluation methodology with transparent scoring criteria.
Details
Motivation: To address the lack of comprehensive evaluation benchmarks for Russian language LLMs and improve interpretability of LLM assessments beyond traditional human evaluations.Method: Created a benchmark with 35 task types categorized by difficulty, developed detailed evaluation criteria for each task, and trained LLM-as-a-Judge evaluators (7B and 32B) to score responses with justifications.
Result: A comprehensive Russian LLM benchmark with 2,100 professionally crafted prompts, scalable evaluation tools, and transparent assessment methodology that reduces reliance on costly human judgments.
Conclusion: POLLUX provides a scalable, interpretable evaluation framework for Russian LLMs that effectively replaces traditional human evaluations while offering greater transparency and efficiency.
Abstract: We introduce POLLUX, a comprehensive open-source benchmark designed to evaluate the generative capabilities of large language models (LLMs) in Russian. Our main contribution is a novel evaluation methodology that enhances the interpretability of LLM assessment. For each task type, we define a set of detailed criteria and develop a scoring protocol where models evaluate responses and provide justifications for their ratings. This enables transparent, criteria-driven evaluation beyond traditional resource-consuming, side-by-side human comparisons. POLLUX includes a detailed, fine-grained taxonomy of 35 task types covering diverse generative domains such as code generation, creative writing, and practical assistant use cases, totaling 2,100 manually crafted and professionally authored prompts. Each task is categorized by difficulty (easy/medium/hard), with experts constructing the dataset entirely from scratch. We also release a family of LLM-as-a-Judge (7B and 32B) evaluators trained for nuanced assessment of generative outputs. This approach provides scalable, interpretable evaluation and annotation tools for model development, effectively replacing costly and less precise human judgments.
[100] Measuring and Guiding Monosemanticity
Ruben Härle, Felix Friedrich, Manuel Brack, Stephan Wäldchen, Björn Deiseroth, Patrick Schramowski, Kristian Kersting
Main category: cs.CL
TL;DR: The paper introduces Guided Sparse Autoencoders (G-SAE) that condition latent representations on labeled concepts to improve feature monosemanticity and control in LLMs, addressing limitations of current sparse autoencoder methods.
Details
Motivation: Current mechanistic interpretability methods struggle with reliably localizing and manipulating feature representations in LLMs. Sparse Autoencoders (SAEs) show promise but suffer from incomplete feature isolation and unreliable monosemanticity, limiting their effectiveness for understanding and controlling LLM behavior.Method: The authors first introduce Feature Monosemanticity Score (FMS) to quantify feature monosemanticity in latent representations. Building on this, they propose Guided Sparse Autoencoders (G-SAE), which conditions latent representations on labeled concepts during training to improve localization and disentanglement of target concepts.
Result: Evaluations on toxicity detection, writing style identification, and privacy attribute recognition show that G-SAE enhances monosemanticity and enables more effective, fine-grained steering with less quality degradation compared to existing methods.
Conclusion: The paper provides actionable guidelines for measuring and advancing mechanistic interpretability and control of LLMs through improved feature localization and manipulation, with G-SAE demonstrating practical improvements in interpretability, behavior detection, and control.
Abstract: There is growing interest in leveraging mechanistic interpretability and controllability to better understand and influence the internal dynamics of large language models (LLMs). However, current methods face fundamental challenges in reliably localizing and manipulating feature representations. Sparse Autoencoders (SAEs) have recently emerged as a promising direction for feature extraction at scale, yet they, too, are limited by incomplete feature isolation and unreliable monosemanticity. To systematically quantify these limitations, we introduce Feature Monosemanticity Score (FMS), a novel metric to quantify feature monosemanticity in latent representation. Building on these insights, we propose Guided Sparse Autoencoders (G-SAE), a method that conditions latent representations on labeled concepts during training. We demonstrate that reliable localization and disentanglement of target concepts within the latent space improve interpretability, detection of behavior, and control. Specifically, our evaluations on toxicity detection, writing style identification, and privacy attribute recognition show that G-SAE not only enhances monosemanticity but also enables more effective and fine-grained steering with less quality degradation. Our findings provide actionable guidelines for measuring and advancing mechanistic interpretability and control of LLMs.
[101] Quantifying Cognitive Bias Induction in LLM-Generated Content
Abeer Alessa, Param Somane, Akshaya Lakshminarasimhan, Julian Skirzynski, Julian McAuley, Jessica Echterhoff
Main category: cs.CL
TL;DR: LLMs exhibit significant biases (framing, hallucination, primacy) that affect human decisions, with users 32% more likely to purchase products after reading LLM-generated summaries vs. original reviews. Targeted interventions show promise for mitigation.
Details
Motivation: LLMs are increasingly used in applications that influence human decisions (shopping reviews, summarization, medical diagnosis), raising concerns about whether they expose users to biased content and how this affects decision-making.Method: Evaluated five LLM families on summarization and news fact-checking tasks, assessing consistency with context and hallucination tendencies using a new self-updating dataset. Measured three biases: framing (sentiment changes), hallucination (post-cutoff knowledge), and primacy (early prompt emphasis). Also conducted human experiments on purchasing decisions and tested 18 mitigation methods across three LLM families.
Result: LLMs changed context sentiment in 26.42% of cases (framing bias), hallucinated on 60.33% of post-knowledge-cutoff questions, and showed primacy bias in 10.12% of cases. Humans were 32% more likely to purchase products after reading LLM summaries vs. original reviews. Targeted interventions showed effectiveness in mitigating these issues.
Conclusion: LLMs systematically expose users to biased content that significantly impacts human decision-making, with concerning rates of framing bias, hallucination, and primacy bias. While targeted mitigation methods show promise, these findings highlight the need for careful deployment and oversight of LLMs in decision-influencing applications.
Abstract: Large language models (LLMs) are integrated into applications like shopping reviews, summarization, or medical diagnosis support, where their use affects human decisions. We investigate the extent to which LLMs expose users to biased content and demonstrate its effect on human decision-making. We assess five LLM families in summarization and news fact-checking tasks, evaluating the consistency of LLMs with their context and their tendency to hallucinate on a new self-updating dataset. Our findings show that LLMs expose users to content that changes the context’s sentiment in 26.42% of cases (framing bias), hallucinate on 60.33% of post-knowledge-cutoff questions, and highlight context from earlier parts of the prompt (primacy bias) in 10.12% of cases, averaged across all tested models. We further find that humans are 32% more likely to purchase the same product after reading a summary of the review generated by an LLM rather than the original review. To address these issues, we evaluate 18 mitigation methods across three LLM families and find the effectiveness of targeted interventions.
[102] On the Robustness of Verbal Confidence of LLMs in Adversarial Attacks
Stephen Obadinma, Xiaodan Zhu
Main category: cs.CL
TL;DR: LLM verbal confidence is vulnerable to adversarial attacks via perturbation and jailbreak methods, compromising transparency and safety in human-AI interactions.
Details
Motivation: Robust verbal confidence from LLMs is essential for transparency, trust, and safety in applications involving human-AI interactions, but current confidence mechanisms lack robustness against adversarial manipulation.Method: Introduce attack frameworks targeting verbal confidence scores through perturbation-based and jailbreak-based methods. Examine various prompting strategies, model sizes, and application domains to assess vulnerabilities.
Result: Attacks significantly impair verbal confidence estimates and cause frequent answer changes. Current verbal confidence is vulnerable, and common defense techniques are largely ineffective or counterproductive.
Conclusion: There is an urgent need to design robust mechanisms for confidence expression in LLMs, as even subtle semantic-preserving modifications can lead to misleading confidence in responses.
Abstract: Robust verbal confidence generated by large language models (LLMs) is crucial for the deployment of LLMs to help ensure transparency, trust, and safety in many applications, including those involving human-AI interactions. In this paper, we present the first comprehensive study on the robustness of verbal confidence under adversarial attacks. We introduce attack frameworks targeting verbal confidence scores through both perturbation and jailbreak-based methods, and demonstrate that these attacks can significantly impair verbal confidence estimates and lead to frequent answer changes. We examine a variety of prompting strategies, model sizes, and application domains, revealing that current verbal confidence is vulnerable and that commonly used defence techniques are largely ineffective or counterproductive. Our findings underscore the need to design robust mechanisms for confidence expression in LLMs, as even subtle semantic-preserving modifications can lead to misleading confidence in responses.
[103] Confident RAG: Enhancing the Performance of LLMs for Mathematics Question Answering through Multi-Embedding and Confidence Scoring
Shiting Chen, Zijian Zhao, Jinsong Chen
Main category: cs.CL
TL;DR: Confident RAG improves mathematical reasoning by generating multiple answers and selecting the highest-confidence one, achieving 10% accuracy gain over vanilla LLMs and 5% over vanilla RAG.
Details
Motivation: LLMs struggle with complex mathematical reasoning, and while RAG helps by grounding them in external knowledge, its effectiveness is unstable due to dependency on single embedding models.Method: Proposed two approaches: Mixture-Embedding RAG (fusing retrieved documents from multiple embeddings) and Confident RAG (generating multiple answers and selecting the one with highest confidence score).
Result: Confident RAG achieved average accuracy improvements of ~10% over vanilla LLMs and ~5% over vanilla RAG. Mixture-Embedding RAG showed limited gains.
Conclusion: Confident RAG is an efficient plug-and-play solution for trustworthy mathematical AI assistants and lays groundwork for deploying Agentic RAG systems in educational settings.
Abstract: Large Language Models (LLMs) hold significant promise for mathematics education, yet they often struggle with complex mathematical reasoning. While Retrieval-Augmented Generation (RAG) mitigates these issues by grounding LLMs in external knowledge, its effectiveness remains unstable, heavily dependent on the choice of a single embedding model. Moving beyond static RAG workflows, we draw on agentic workflow patterns, a paradigm that introduces structured task decomposition and collaboration to enhance system performance. We propose and examine two novel approaches that combine the benefits of multiple embedding models. While our Mixture-Embedding RAG approach (fusing retrieved documents) shows limited gains, our Confident RAG method (generating multiple answers and selecting the one with the highest confidence score) demonstrates significant improvement. Experimental results show that Confident RAG achieved average accuracy improvements of approximately 10% over vanilla LLMs and 5% over vanilla RAG. The consistent results across different LLMs and embedding models indicate that Confident RAG is an efficient plug-and-play solution for trustworthy mathematical AI assistants. Finally, we discuss how this work lays the groundwork for deploying Agentic RAG systems in educational settings, where autonomous planning and iterative refinement can be built upon our robust retrieval foundation.
[104] Checklists Are Better Than Reward Models For Aligning Language Models
Vijay Viswanathan, Yanchao Sun, Shuang Ma, Xiang Kong, Meng Cao, Graham Neubig, Tongshuang Wu
Main category: cs.CL
TL;DR: RLCF (Reinforcement Learning from Checklist Feedback) improves instruction following by using flexible, instruction-specific criteria instead of fixed criteria like “helpfulness” and “harmfulness”, achieving consistent performance gains across multiple benchmarks.
Details
Motivation: Current reinforcement learning for language model alignment typically uses fixed criteria (e.g., "helpfulness", "harmfulness"), which limits their effectiveness in handling diverse user instructions with specific requirements. The authors aim to broaden RL's impact by using flexible, instruction-specific criteria that can better capture the multitude of needs expressed in user queries.Method: Proposes Reinforcement Learning from Checklist Feedback (RLCF): 1) Extract checklists from instructions, 2) Evaluate how well responses satisfy each checklist item using both AI judges and specialized verifier programs, 3) Combine these scores to compute rewards for reinforcement learning training.
Result: RLCF applied to Qwen2.5-7B-Instruct improved performance on all five benchmarks tested: 4-point boost in hard satisfaction rate on FollowBench, 6-point increase on InFoBench, and 3-point rise in win rate on Arena-Hard. It was the only method to improve performance on every benchmark.
Conclusion: Checklist feedback is established as a key tool for improving language models’ support of diverse user queries, demonstrating that flexible, instruction-specific criteria can significantly enhance RL’s effectiveness in eliciting instruction following compared to fixed criteria approaches.
Abstract: Language models must be adapted to understand and follow user instructions. Reinforcement learning is widely used to facilitate this – typically using fixed criteria such as “helpfulness” and “harmfulness”. In our work, we instead propose using flexible, instruction-specific criteria as a means of broadening the impact that reinforcement learning can have in eliciting instruction following. We propose “Reinforcement Learning from Checklist Feedback” (RLCF). From instructions, we extract checklists and evaluate how well responses satisfy each item - using both AI judges and specialized verifier programs - then combine these scores to compute rewards for RL. We compare RLCF with other alignment methods applied to a strong instruction following model (Qwen2.5-7B-Instruct) on five widely-studied benchmarks – RLCF is the only method to improve performance on every benchmark, including a 4-point boost in hard satisfaction rate on FollowBench, a 6-point increase on InFoBench, and a 3-point rise in win rate on Arena-Hard. These results establish checklist feedback as a key tool for improving language models’ support of queries that express a multitude of needs.
[105] Debating Truth: Debate-driven Claim Verification with Multiple Large Language Model Agents
Haorui He, Yupeng Li, Dacheng Wen, Yang Chen, Reynold Cheng, Donglong Chen, Francis C. M. Lau
Main category: cs.CL
TL;DR: DebateCV introduces a debate-driven claim verification framework using multiple LLM agents, with Debate-SFT training to improve adjudication of complex debates for better misinformation detection.
Details
Motivation: Single-agent claim verification methods struggle with complex claims requiring nuanced analysis of multifaceted evidence, while real-world human fact-checking often involves debate and discussion to surface subtle errors.Method: Proposes DebateCV framework with two Debaters arguing opposing stances over multiple rounds, and a Moderator weighing evidential strength. Introduces Debate-SFT post-training framework using synthetic data to enhance agents’ ability to adjudicate debates effectively.
Result: Methods surpass state-of-the-art non-debate approaches in both accuracy across various evidence conditions and justification quality, improving misinformation detection capabilities.
Conclusion: The debate-driven approach strengthens societal resilience against misinformation and contributes to a more trustworthy online information ecosystem by mimicking real-world fact-checking practices through multi-agent LLM debates.
Abstract: Claim verification is essential for digital literacy, yet state-of-the-art single-agent methods often struggle with complex claims that require nuanced analysis of multifaceted online evidence. Inspired by real-world human fact-checking practices, we propose \textbf{DebateCV}, the first debate-driven claim verification framework powered by multiple LLM agents. In DebateCV, two \textit{Debaters} argue opposing stances over multiple rounds to surface subtle errors in single-agent assessments. A decisive \textit{Moderator} is then required to weigh the evidential strength of conflicting arguments to deliver an accurate verdict. Yet zero-shot agents struggle to adjudicate multi-round debates for verifying complex claims, often defaulting to neutral judgements, and no datasets exist for training agents for this role. To bridge this gap, we propose \textbf{Debate-SFT}, a post-training framework that leverages synthetic data to enhance agents’ ability to effectively adjudicate debates for claim verification. Results show that our methods surpass state-of-the-art non-debate approaches in both accuracy (across various evidence conditions) and justification quality, which strengthens societal resilience against misinformation and contributes to a more trustworthy online information ecosystem.
[106] SpeechRole: A Large-Scale Dataset and Benchmark for Evaluating Speech Role-Playing Agents
Changhao Jiang, Jiajun Sun, Yifei Cao, Jiabao Zhuang, Hui Li, Baoyu Fan, Tao Ji, Tao Gui, Qi Zhang
Main category: cs.CL
TL;DR: The paper introduces SpeechRole-Data (large speech role-playing dataset) and SpeechRole-Eval (evaluation benchmark) to address the lack of systematic evaluation for Speech Role-Playing Agents (SRPAs).
Details
Motivation: Existing role-playing research focuses mainly on text modality, neglecting speech in realistic interactive scenarios, with no systematic evaluation framework for speech role-playing agents.Method: Constructed SpeechRole-Data (98 diverse roles, 112k speech conversations) with distinct vocal characteristics, and proposed SpeechRole-Eval benchmark for multidimensional evaluation of SRPAs.
Result: Experimental results reveal advantages and challenges of both cascaded and end-to-end speech role-playing agents in maintaining vocal style consistency and role coherence.
Conclusion: The work provides foundational resources (data, code, models) for speech-driven multimodal role-playing research and highlights future development directions in this field.
Abstract: Recently, role-playing agents have emerged as a promising paradigm for achieving personalized interaction and emotional resonance. Existing research primarily focuses on the textual modality, neglecting the critical dimension of speech in realistic interactive scenarios. In particular, there is a lack of systematic evaluation for Speech Role-Playing Agents (SRPAs). To address this gap, we construct SpeechRole-Data, a large-scale, high-quality dataset that comprises 98 diverse roles and 112k speech-based single-turn and multi-turn conversations. Each role demonstrates distinct vocal characteristics, including timbre and prosody, thereby enabling more sophisticated speech role-playing. Furthermore, we propose SpeechRole-Eval, a multidimensional evaluation benchmark that systematically assesses SRPAs performance in key aspects such as fundamental interaction ability, speech expressiveness, and role-playing fidelity. Experimental results reveal the advantages and challenges of both cascaded and end-to-end speech role-playing agents in maintaining vocal style consistency and role coherence. We release all data, code, and baseline models to provide a solid foundation for speech-driven multimodal role-playing research and to foster further developments in this field.
[107] ShoppingBench: A Real-World Intent-Grounded Shopping Benchmark for LLM-based Agents
Jiangyuan Wang, Kejun Xiao, Qi Sun, Huaipeng Zhao, Tao Luo, Jian Dong Zhang, Xiaoyi Zeng
Main category: cs.CL
TL;DR: ShoppingBench: A novel e-commerce benchmark for complex user intents like voucher application, budget management, and multi-product seller finding, featuring a scalable simulation framework and large-scale shopping sandbox with over 2.5M real products.
Details
Motivation: Existing e-commerce benchmarks focus only on basic user intents (finding/purchasing products), but real-world users have more complex goals like applying vouchers, managing budgets, and finding multi-product sellers. There's a gap in evaluating AI agents on these realistic, challenging shopping scenarios.Method: 1) Proposed ShoppingBench benchmark with increasingly challenging grounded intents; 2) Scalable framework to simulate user instructions based on various intents derived from real-world products; 3) Large-scale shopping sandbox with over 2.5 million real products as interactive simulated environment; 4) Trajectory distillation strategy with supervised fine-tuning and reinforcement learning on synthetic trajectories to distill large agent capabilities into smaller ones.
Result: State-of-the-art language agents (including GPT-4.1) achieve absolute success rates under 50% on ShoppingBench tasks, highlighting significant challenges. The proposed trajectory distillation approach enables smaller trained agents to achieve competitive performance compared to GPT-4.1.
Conclusion: ShoppingBench addresses the gap in evaluating AI agents on complex real-world shopping intents, revealing current limitations of even advanced models. The proposed distillation methods show promise for creating more efficient, capable shopping agents that can handle realistic e-commerce scenarios.
Abstract: Existing benchmarks in e-commerce primarily focus on basic user intents, such as finding or purchasing products. However, real-world users often pursue more complex goals, such as applying vouchers, managing budgets, and finding multi-products seller. To bridge this gap, we propose ShoppingBench, a novel end-to-end shopping benchmark designed to encompass increasingly challenging levels of grounded intent. Specifically, we propose a scalable framework to simulate user instructions based on various intents derived from sampled real-world products. To facilitate consistent and reliable evaluations, we provide a large-scale shopping sandbox that serves as an interactive simulated environment, incorporating over 2.5 million real-world products. Experimental results demonstrate that even state-of-the-art language agents (such as GPT-4.1) achieve absolute success rates under 50% on our benchmark tasks, highlighting the significant challenges posed by our ShoppingBench. In addition, we propose a trajectory distillation strategy and leverage supervised fine-tuning, along with reinforcement learning on synthetic trajectories, to distill the capabilities of a large language agent into a smaller one. As a result, our trained agent achieves competitive performance compared to GPT-4.1.
[108] Persistent Instability in LLM’s Personality Measurements: Effects of Scale, Reasoning, and Conversation History
Tommaso Tosato, Saskia Helbling, Yorguin-Jose Mantilla-Ramos, Mahmood Hegazy, Alberto Tosato, David John Lemay, Irina Rish, Guillaume Dumas
Main category: cs.CL
TL;DR: Current LLMs show persistent personality instability across scales and interventions, with question reordering causing large shifts, limited stability gains from scaling, and paradoxical effects from reasoning/conversation history.
Details
Motivation: Large language models require consistent behavioral patterns for safe deployment, but there are indications of large variability in personality trait expression that may lead to unstable behavior.Method: PERSIST framework evaluates 25 open-source models (1B-685B parameters) across 2M+ responses using traditional (BFI, SD3) and novel LLM-adapted personality questionnaires, systematically varying model size, personas, reasoning modes, question order/paraphrasing, and conversation history.
Result: (1) Question reordering alone introduces large personality measurement shifts; (2) Scaling provides limited stability gains (400B+ models show SD >0.3 on 5-point scales); (3) Reasoning and conversation history paradoxically increase variability; (4) Detailed personas show mixed effects with misaligned personas having higher variability; (5) LLM-adapted questionnaires show comparable instability to human-centric versions.
Conclusion: Persistent instability across scales and mitigation strategies suggests current LLMs lack architectural foundations for genuine behavioral consistency, indicating current alignment strategies may be inadequate for safety-critical applications requiring predictable behavior.
Abstract: Large language models require consistent behavioral patterns for safe deployment, yet there are indications of large variability that may lead to an instable expression of personality traits in these models. We present PERSIST (PERsonality Stability in Synthetic Text), a comprehensive evaluation framework testing 25 open-source models (1B-685B parameters) across 2 million+ responses. Using traditional (BFI, SD3) and novel LLM-adapted personality questionnaires, we systematically vary model size, personas, reasoning modes, question order or paraphrasing, and conversation history. Our findings challenge fundamental assumptions: (1) Question reordering alone can introduce large shifts in personality measurements; (2) Scaling provides limited stability gains: even 400B+ models exhibit standard deviations >0.3 on 5-point scales; (3) Interventions expected to stabilize behavior, such as reasoning and inclusion of conversation history, can paradoxically increase variability; (4) Detailed persona instructions produce mixed effects, with misaligned personas showing significantly higher variability than the helpful assistant baseline; (5) The LLM-adapted questionnaires, despite their improved ecological validity, exhibit instability comparable to human-centric versions. This persistent instability across scales and mitigation strategies suggests that current LLMs lack the architectural foundations for genuine behavioral consistency. For safety-critical applications requiring predictable behavior, these findings indicate that current alignment strategies may be inadequate.
[109] CCFQA: A Benchmark for Cross-Lingual and Cross-Modal Speech and Text Factuality Evaluation
Yexing Du, Kaiyuan Liu, Youcheng Pan, Zheng Chu, Bo Yang, Xiaocheng Feng, Ming Liu, Yang Xiang
Main category: cs.CL
TL;DR: CCFQA is a new benchmark for evaluating cross-lingual and cross-modal factuality in MLLMs, focusing on multilingual speech-text factual questions across 8 languages, revealing current MLLMs’ limitations and proposing effective few-shot transfer learning.
Details
Motivation: Existing MLLM benchmarks focus on English and textual/visual modalities, creating a gap for evaluating multilingual speech understanding and factuality, which is crucial as LLMs become more widely used in multilingual contexts.Method: Created CCFQA benchmark with parallel speech-text factual questions across 8 languages; proposed few-shot transfer learning strategy to transfer English QA capabilities to multilingual Spoken QA tasks using just 5-shot training.
Result: Current MLLMs face substantial challenges on CCFQA; the proposed few-shot transfer learning achieves competitive performance with GPT-4o-mini-Audio using minimal training data.
Conclusion: CCFQA addresses the multilingual speech factuality evaluation gap, reveals MLLMs’ current limitations, and demonstrates effective transfer learning for multilingual Spoken QA, providing a valuable resource for developing more robust MLLMs.
Abstract: As Large Language Models (LLMs) are increasingly popularized in the multilingual world, ensuring hallucination-free factuality becomes markedly crucial. However, existing benchmarks for evaluating the reliability of Multimodal Large Language Models (MLLMs) predominantly focus on textual or visual modalities with a primary emphasis on English, which creates a gap in evaluation when processing multilingual input, especially in speech. To bridge this gap, we propose a novel Cross-lingual and Cross-modal Factuality benchmark (CCFQA). Specifically, the CCFQA benchmark contains parallel speech-text factual questions across 8 languages, designed to systematically evaluate MLLMs’ cross-lingual and cross-modal factuality capabilities. Our experimental results demonstrate that current MLLMs still face substantial challenges on the CCFQA benchmark. Furthermore, we propose a few-shot transfer learning strategy that effectively transfers the Question Answering (QA) capabilities of LLMs in English to multilingual Spoken Question Answering (SQA) tasks, achieving competitive performance with GPT-4o-mini-Audio using just 5-shot training. We release CCFQA as a foundational research resource to promote the development of MLLMs with more robust and reliable speech understanding capabilities. Our code and dataset are available at https://github.com/yxduir/ccfqa.
[110] DESIGNER: Design-Logic-Guided Multidisciplinary Data Synthesis for LLM Reasoning
Weize Liu, Yongchi Zhao, Yijia Luo, Mingyu Xu, Jiaheng Liu, Yanan Li, Xiguo Hu, Zhiqi Bai, Yuchi Xu, Wenbo Su, Bo Zheng
Main category: cs.CL
TL;DR: DESIGNER: A pipeline that uses “design logic” to generate large-scale, multidisciplinary reasoning questions from book and web corpora, creating challenging datasets that improve LLM reasoning capabilities beyond existing benchmarks.
Details
Motivation: LLMs struggle with complex multi-step reasoning across diverse disciplines. Existing reasoning datasets lack disciplinary breadth, reasoning depth, diversity, and guiding principles for question synthesis.Method: Introduce “design logic” concept to mimic human educators’ question-creation process. Reverse-engineer 120,000+ design logics from existing questions, match them with source documents (book and web corpora) to generate questions with controllable types and difficulty levels.
Result: Created two large-scale datasets: DLR-Book (3.04M questions from books) and DLR-Web (1.66M questions from web). Questions show greater difficulty and diversity than baselines. SFT on Qwen3 and Llama3 models with this data substantially enhances multidisciplinary reasoning, even surpassing official final models.
Conclusion: The DESIGNER pipeline successfully generates high-quality, challenging reasoning questions at scale, significantly improving LLMs’ multidisciplinary reasoning capabilities beyond existing approaches.
Abstract: Large language models (LLMs) have achieved remarkable success in many natural language tasks but still struggle with complex, multi-step reasoning, particularly across diverse disciplines. Existing reasoning datasets often lack disciplinary breadth, reasoning depth, and diversity, as well as guiding principles for question synthesis. We propose DESIGNER: a DESIGN-logic-guidEd Reasoning data synthesis pipeline that leverages naturally available, extensive raw documents (e.g., book corpus and web corpus) to generate multidisciplinary challenging questions. We introduce the concept of “design logic” and instruct LLMs to mimic human educators’ question-creation process, enabling the automated synthesis of large-scale, high-difficulty questions. We use LLMs to reverse-engineer and abstract over 120,000 design logics from existing questions across various disciplines. By matching these design logics with source documents, we are able to generate reasoning questions with controllable question types and difficulty levels. Using this pipeline, we synthesized two large-scale reasoning datasets that span 75 disciplines: DLR-Book (3.04 million questions from the book corpus) and DLR-Web (1.66 million questions from the web corpus). Data analysis indicates that the questions synthesized by our method exhibit greater difficulty and diversity compared to those in the baseline datasets. We validate our synthesized data through supervised fine-tuning (SFT) on the Qwen3 and Llama3 model families. Our data substantially enhances their multidisciplinary reasoning capabilities, outperforming existing datasets. Notably, by applying SFT on the base versions of these models using only our data, we even surpass their official final models that have undergone the full post-training process.
[111] Chain-of-Query: Unleashing the Power of LLMs in SQL-Aided Table Understanding via Multi-Agent Collaboration
Songyuan Sui, Hongyi Liu, Serena Liu, Li Li, Soo-Hyun Choi, Rui Chen, Xia Hu
Main category: cs.CL
TL;DR: Chain-of-Query (CoQ) is a multi-agent framework that improves table understanding by using natural-language schema representations, clause-by-clause SQL generation, and hybrid reasoning to reduce structural complexity and execution dependency.
Details
Motivation: LLMs struggle with table understanding due to structural complexity of tabular data. Existing multi-agent SQL generation approaches have limitations: poor table structure comprehension, error propagation causing invalid queries, and over-reliance on execution correctness.Method: CoQ uses natural-language-style schema representations to abstract structural noise, employs clause-by-clause SQL generation for better query quality, and implements hybrid reasoning that separates SQL-based mechanical reasoning from LLM-based logical inference.
Result: Extensive experiments across four models and five benchmarks show CoQ achieves substantial accuracy improvements and significantly lowers invalid SQL rates compared to prior LLM-based, SQL-aided, and hybrid baselines.
Conclusion: CoQ demonstrates superior effectiveness in table understanding by addressing key limitations of existing approaches through its novel multi-agent framework design.
Abstract: Table understanding requires structured, multi-step reasoning. Large Language Models (LLMs) struggle with it due to the structural complexity of tabular data. Recently, multi-agent frameworks for SQL generation have shown promise in tackling the challenges of understanding tabular data, but existing approaches often suffer from limitations such as the inability to comprehend table structure for reliable SQL generation, error propagation that results in invalid queries, and over-reliance on execution correctness. To address these issues, we propose Chain-of-Query (CoQ), a novel multi-agent framework for SQL-aided table understanding. CoQ adopts natural-language-style representations of table schemas to abstract away structural noise and enhance understanding. It employs a clause-by-clause SQL generation strategy to improve query quality and introduces a hybrid reasoning division that separates SQL-based mechanical reasoning from LLM-based logical inference, thereby reducing reliance on execution outcomes. Extensive experiments across four models and five widely used benchmarks demonstrate that CoQ achieves substantial accuracy improvements and significantly lowers invalid SQL rates compared to prior generic LLM-based, SQL-aided, and hybrid baselines, confirming its superior effectiveness in table understanding. The code is available at https://github.com/SongyuanSui/ChainofQuery.
[112] Reasoning-Intensive Regression
Diane Tchuindjo, Omar Khattab
Main category: cs.CL
TL;DR: MENTAT combines batch-reflective prompt optimization with neural ensemble learning to improve performance on reasoning-intensive regression tasks where LLMs and standard finetuning struggle.
Details
Motivation: AI researchers increasingly use LLMs for reasoning-intensive regression (RiR) tasks that require deep contextual analysis but have limited training data. Current approaches like prompting frozen LLMs or finetuning Transformer encoders often struggle with these subtle numerical scoring problems.Method: MENTAT: a simple lightweight method combining batch-reflective prompt optimization with neural ensemble learning. The approach optimizes prompts based on batch performance and uses ensemble techniques to improve reasoning capabilities.
Result: MENTAT achieves up to 65% improvement over baseline methods (prompting frozen LLMs and finetuning Transformer encoders) on four realistic RiR benchmark tasks.
Conclusion: MENTAT effectively addresses reasoning-intensive regression challenges, though substantial room remains for future advances in this emerging area of AI research.
Abstract: AI researchers and practitioners increasingly apply large language models (LLMs) to what we call reasoning-intensive regression (RiR), i.e., deducing subtle numerical scores from text. Unlike standard language regression tasks, e.g., for sentiment or similarity, RiR often appears instead in ad-hoc problems such as rubric-based scoring, modeling dense rewards in complex environments, or domain-specific retrieval, where much deeper analysis of context is required while only limited task-specific training data and computation are available. We cast four realistic problems as RiR tasks to establish an initial benchmark, and use that to test our hypothesis that prompting frozen LLMs and finetuning Transformer encoders via gradient descent will both often struggle in RiR. We then propose MENTAT, a simple and lightweight method that combines batch-reflective prompt optimization with neural ensemble learning. MENTAT achieves up to 65% improvement over both baselines, though substantial room remains for future advances in RiR.
[113] Explainable Semantic Text Relations: A Question-Answering Framework for Comparing Document Content
Yehudit Aperstein, Alon Gottlib, Gal Benita, Alexander Apartsin
Main category: cs.CL
TL;DR: This paper proposes a framework for understanding semantic relations between texts using Answerable Question Sets (AQS), creates a synthetic benchmark for evaluation, and tests various models on STR classification.
Details
Motivation: Current approaches lack formal definitions for semantic relations between texts and don't provide explainable outputs about what information is present, missing, or added between text pairs.Method: Define semantic relations through set-theoretic relations between Answerable Question Sets (AQS) - the sets of questions each text can answer. Create synthetic benchmark using controlled paraphrasing and information removal via AQS manipulations.
Result: Developed a formal framework for Semantic Text Relations (STR) and created a synthetic dataset. Evaluated discriminative and generative models for STR classification to assess how well different architectures capture semantic relations beyond surface similarity.
Conclusion: The AQS-based framework provides a formal, explainable approach to semantic text relations. The released dataset and generation code support further research in understanding fine-grained informational relations between texts.
Abstract: Understanding semantic relations between two texts is crucial for many information and document management tasks, in which one must determine whether the content fully overlaps, is completely superseded by another document, or overlaps only partially, with unique information in each. Beyond establishing this relation, it is equally important to provide explainable outputs that specify which pieces of information are present, missing, or newly added between the text pair. In this study, we formally define semantic relations between two texts through the set-theoretic relation between their respective Answerable Question Sets (AQS), the sets of questions each text can answer. Under this formulation, Semantic Text Relation (STR), such as equivalence, inclusion, and mutual overlap, becomes a well-defined set relation between the corresponding texts’ AQSs. The set differences between the AQSs also serve as an explanation or diagnostic tool for identifying how the information in the texts diverges. Using this definition, we construct a synthetic benchmark that captures fine-grained informational relations through controlled paraphrasing and deliberate information removal supported by AQS manipulations. We then use this dataset to evaluate several discriminative and generative models for classifying text pairs into STR categories, assessing how well different model architectures capture semantic relations beyond surface-level similarity. We publicly release both the dataset and the data generation code to support further research.
[114] Advancing Natural Language Formalization to First Order Logic with Fine-tuned LLMs
Felix Vossel, Till Mossakowski, Björn Gehrke
Main category: cs.CL
TL;DR: Fine-tuned Flan-T5-XXL achieves 70% accuracy in translating natural language to first-order logic, outperforming GPT-4o, DeepSeek-R1-0528, and symbolic systems like ccg2lambda.
Details
Motivation: Automating natural language to first-order logic translation is crucial for knowledge representation and formal methods, but remains challenging despite advances in LLMs.Method: Systematic evaluation of fine-tuned LLMs comparing encoder-decoder vs. decoder-only architectures, using MALLS and Willow datasets. Techniques include vocabulary extension, predicate conditioning, multilingual training, with metrics for exact match, logical equivalence, and predicate alignment.
Result: Fine-tuned Flan-T5-XXL achieves 70% accuracy with predicate lists, outperforming GPT-4o and DeepSeek-R1-0528. Key findings: predicate availability boosts performance 15-20%, T5 models surpass larger decoder-only LLMs, models generalize to unseen logical arguments (FOLIO dataset) without specific training.
Conclusion: While structural logic translation proves robust, predicate extraction emerges as the main bottleneck. Fine-tuned encoder-decoder models show strong performance in FOL translation tasks.
Abstract: Automating the translation of natural language to first-order logic (FOL) is crucial for knowledge representation and formal methods, yet remains challenging. We present a systematic evaluation of fine-tuned LLMs for this task, comparing architectures (encoder-decoder vs. decoder-only) and training strategies. Using the MALLS and Willow datasets, we explore techniques like vocabulary extension, predicate conditioning, and multilingual training, introducing metrics for exact match, logical equivalence, and predicate alignment. Our fine-tuned Flan-T5-XXL achieves 70% accuracy with predicate lists, outperforming GPT-4o and even the DeepSeek-R1-0528 model with CoT reasoning ability as well as symbolic systems like ccg2lambda. Key findings show: (1) predicate availability boosts performance by 15-20%, (2) T5 models surpass larger decoder-only LLMs, and (3) models generalize to unseen logical arguments (FOLIO dataset) without specific training. While structural logic translation proves robust, predicate extraction emerges as the main bottleneck.
[115] Beyond Token Length: Step Pruner for Efficient and Accurate Reasoning in Large Language Models
Canhui Wu, Qiong Cao, Chang Li, Zhenfang Wang, Chao Xue, Yuwei Fan, Wei Xi, Xiaodong He
Main category: cs.CL
TL;DR: Step Pruner (SP) is an RL framework that reduces overthinking in Large Reasoning Models by penalizing redundant reasoning steps rather than just tokens, achieving SOTA accuracy with 69.7% token reduction on AIME24.
Details
Motivation: Large Reasoning Models suffer from excessive verbosity (overthinking). Existing RL methods that penalize tokens have two problems: fewer tokens don't always mean fewer reasoning steps, and models can hack training by discarding reasoning steps to minimize token usage.Method: Step Pruner (SP) uses step-aware RL reward function that prioritizes correctness while penalizing redundant steps, withholds rewards for incorrect responses, and includes dynamic stopping mechanism to prevent step merging hacking behavior.
Result: SP achieves state-of-the-art accuracy across four reasoning benchmarks while significantly reducing response length. On AIME24, it reduces token usage by 69.7%.
Conclusion: Step Pruner effectively addresses overthinking in LRMs by focusing on reasoning step efficiency rather than just token count, preventing hacking behaviors while maintaining high accuracy.
Abstract: Large Reasoning Models (LRMs) demonstrate strong performance on complex tasks but often suffer from excessive verbosity, known as “overthinking.” Existing solutions via reinforcement learning (RL) typically penalize generated tokens to promote conciseness. However, these methods encounter two challenges: responses with fewer tokens do not always correspond to fewer reasoning steps, and models may develop hacking behavior in later stages of training by discarding reasoning steps to minimize token usage. In this work, we introduce \textbf{Step Pruner (SP)}, an RL framework that steers LRMs toward more efficient reasoning by favoring compact reasoning steps. Our step-aware reward function prioritizes correctness while imposing penalties for redundant steps, and withholds rewards for incorrect responses to prevent the reinforcement of erroneous reasoning. Moreover, we propose a dynamic stopping mechanism: when the model’s output no longer shortens, training is halted to prevent hacking behavior caused by the merging of steps. Extensive experiments across four reasoning benchmarks demonstrate that SP achieves state-of-the-art accuracy while significantly reducing response length. For instance, on AIME24, SP reduces token usage by \textbf{69.7%}.
[116] SECA: Semantically Equivalent and Coherent Attacks for Eliciting LLM Hallucinations
Buyun Liang, Liangzu Peng, Jinqi Luo, Darshan Thaker, Kwan Ho Ryan Chan, René Vidal
Main category: cs.CL
TL;DR: SECA proposes realistic adversarial attacks to elicit hallucinations in LLMs through semantically equivalent and coherent prompt modifications, achieving higher attack success rates than existing methods.
Details
Motivation: Existing adversarial attacks for hallucination elicitation in LLMs often produce unrealistic prompts (gibberish tokens or altered meaning), limiting practical insights into how hallucinations occur in real-world scenarios.Method: Formulates realistic attacks as constrained optimization over input prompt space under semantic equivalence and coherence constraints, and introduces a constraint-preserving zeroth-order method to search for adversarial yet feasible prompts.
Result: SECA achieves higher attack success rates while maintaining almost no semantic equivalence or coherence errors compared to existing methods on open-ended multiple-choice question answering tasks.
Conclusion: SECA demonstrates that both open-source and commercial LLMs are sensitive to realistic and plausible prompt variations, highlighting the need for more robust models against such attacks.
Abstract: Large Language Models (LLMs) are increasingly deployed in high-risk domains. However, state-of-the-art LLMs often produce hallucinations, raising serious concerns about their reliability. Prior work has explored adversarial attacks for hallucination elicitation in LLMs, but it often produces unrealistic prompts, either by inserting gibberish tokens or by altering the original meaning. As a result, these approaches offer limited insight into how hallucinations may occur in practice. While adversarial attacks in computer vision often involve realistic modifications to input images, the problem of finding realistic adversarial prompts for eliciting LLM hallucinations has remained largely underexplored. To address this gap, we propose Semantically Equivalent and Coherent Attacks (SECA) to elicit hallucinations via realistic modifications to the prompt that preserve its meaning while maintaining semantic coherence. Our contributions are threefold: (i) we formulate finding realistic attacks for hallucination elicitation as a constrained optimization problem over the input prompt space under semantic equivalence and coherence constraints; (ii) we introduce a constraint-preserving zeroth-order method to effectively search for adversarial yet feasible prompts; and (iii) we demonstrate through experiments on open-ended multiple-choice question answering tasks that SECA achieves higher attack success rates while incurring almost no semantic equivalence or semantic coherence errors compared to existing methods. SECA highlights the sensitivity of both open-source and commercial gradient-inaccessible LLMs to realistic and plausible prompt variations. Code is available at https://github.com/Buyun-Liang/SECA.
[117] TRepLiNa: Layer-wise CKA+REPINA Alignment Improves Low-Resource Machine Translation in Aya-23 8B
Toshiki Nakai, Ravi Kiran Chikkala, Lena Sophie Oberkircher, Nicholas Jennings, Natalia Skachkova, Tatiana Anikina, Jesujoba Oluwadara Alabi
Main category: cs.CL
TL;DR: TRepLiNa method combines CKA and REPINA to align mid-level layers in multilingual LLMs, improving low-resource language translation with minimal data.
Details
Motivation: Addresses India's linguistic gap by improving translation for low-resource languages (Mundari, Santali, Bhili) to high-resource languages (Hindi/English) using limited data.Method: Combines Centered Kernel Alignment (CKA) for cross-lingual representation alignment with REPINA regularization to constrain parameter updates, applied to Aya-23 8B with QLoRA in zero-shot, few-shot, and fine-tuning settings.
Result: TRepLiNa (CKA+REPINA) alignment of mid-level layers improves low-resource language translation quality, especially in data-scarce settings, as a low-cost practical approach.
Conclusion: Enforcing cross-lingual similarity in specific internal layers of multilingual LLMs through TRepLiNa effectively enhances translation for low-resource languages with minimal computational resources.
Abstract: The 2025 Multimodal Models for Low-Resource Contexts and Social Impact (MMLoSo) Language Challenge addresses one of India’s most pressing linguistic gaps: the lack of resources for its diverse low-resource languages (LRLs). In this study, we investigate whether enforcing cross-lingual similarity in specific internal layers of a decoder-only multilingual large language model (LLM) can improve translation quality from LRL to high-resource language (HRL). Specifically, we combine Centered Kernel Alignment (CKA), a similarity metric that encourages representations of different languages to align, with REPINA, a regularization method that constrains parameter updates to remain close to the pretrained model, into a joint method we call TRepLiNa. In this research project, we experiment with zero-shot, few-shot, and fine-tuning settings using Aya-23 8B with QLoRA across MMLoSo shared task language pairs (Mundari, Santali, Bhili) with Hindi/English pivots. Our results show that aligning mid-level layers using TRepLiNa (CKA+REPINA) is a low-cost, practical approach to improving LRL translation, especially in data-scarce settings.
[118] NarraBench: A Comprehensive Framework for Narrative Benchmarking
Sil Hamilton, Matthew Wilkens, Andrew Piper
Main category: cs.CL
TL;DR: NarraBench introduces a taxonomy of narrative-understanding tasks, surveys 78 existing benchmarks, and finds only 27% of narrative tasks are well-covered, with significant gaps in events, style, perspective, and revelation.
Details
Motivation: There's a need to systematically understand and evaluate narrative understanding in NLP, as current benchmarks may not adequately cover important aspects of narrative comprehension that are crucial for testing LLMs' narrative capabilities.Method: Developed a theory-informed taxonomy of narrative-understanding tasks, conducted a comprehensive survey of 78 existing benchmarks, and analyzed their coverage and alignment with narrative aspects.
Result: Only 27% of narrative tasks are well captured by existing benchmarks, with significant gaps in narrative events, style, perspective, and revelation. There’s also a need for benchmarks that can assess subjective and perspectival aspects of narrative.
Conclusion: The taxonomy, survey, and methodology provide valuable tools for NLP researchers to better evaluate LLM narrative understanding, highlighting the need for new benchmarks that cover overlooked narrative aspects and can handle subjective elements.
Abstract: We present NarraBench, a theory-informed taxonomy of narrative-understanding tasks, as well as an associated survey of 78 existing benchmarks in the area. We find significant need for new evaluations covering aspects of narrative understanding that are either overlooked in current work or are poorly aligned with existing metrics. Specifically, we estimate that only 27% of narrative tasks are well captured by existing benchmarks, and we note that some areas – including narrative events, style, perspective, and revelation – are nearly absent from current evaluations. We also note the need for increased development of benchmarks capable of assessing constitutively subjective and perspectival aspects of narrative, that is, aspects for which there is generally no single correct answer. Our taxonomy, survey, and methodology are of value to NLP researchers seeking to test LLM narrative understanding.
[119] Closing the Data-Efficiency Gap Between Autoregressive and Masked Diffusion LLMs
Xu Pan, Ely Hahami, Jingxuan Fan, Ziqian Xie, Haim Sompolinsky
Main category: cs.CL
TL;DR: Masked diffusion LLMs (dLLMs) outperform autoregressive LLMs (arLLMs) in knowledge acquisition through fine-tuning, showing better generalization without data augmentation and being free from the reversal curse. A novel masked fine-tuning paradigm for arLLMs is introduced, significantly improving their data efficiency and closing the performance gap with dLLMs.
Details
Motivation: Autoregressive LLMs struggle with knowledge injection via fine-tuning due to issues like the reversal curse, while masked diffusion LLMs show promise in pre-training but their post-training knowledge acquisition capabilities are unknown. The research aims to compare knowledge acquisition in both paradigms and develop better fine-tuning methods.Method: Fine-tuned arLLMs and dLLMs on three diverse datasets, evaluating with forward and backward style QA to probe knowledge generalization and reversal curse. Introduced a novel masked fine-tuning paradigm for arLLMs that improves data efficiency. Extended the approach to supervised fine-tuning for mathematical capability.
Result: dLLMs achieved high accuracies on both forward and backward QAs without paraphrases, while arLLMs required extensive data augmentation and only worked when paraphrase order matched QA style. The proposed masked fine-tuning paradigm drastically improved arLLM data efficiency, closing the performance gap with dLLMs. Masked SFT also outperformed regular SFT for mathematical capability.
Conclusion: Masked diffusion LLMs have superior knowledge acquisition capabilities compared to autoregressive LLMs, being free from reversal curse and requiring less data augmentation. The novel masked fine-tuning paradigm successfully improves arLLM performance, demonstrating that architectural advantages of dLLMs can be transferred to arLLMs through specialized training methods.
Abstract: Despite autoregressive large language models (arLLMs) being the current dominant paradigm in language modeling, effectively updating these models to incorporate new factual knowledge still remains difficult. They resist knowledge injection via fine-tuning due to inherent shortcomings such as the “reversal curse” – the challenge of answering questions that reverse the original information order in the training sample. Masked diffusion large language models (dLLMs) are rapidly emerging as a powerful alternative to the arLLM paradigm, with evidence of better data efficiency and free of the “reversal curse” in pre-training. However, it is unknown whether these advantages extend to the post-training phase, i.e. whether pre-trained dLLMs can easily acquire new knowledge through fine-tuning. On three diverse datasets, we fine-tune arLLMs and dLLMs, evaluating them with forward and backward style Question Answering (QA) to probe knowledge generalization and the reversal curse. Our results confirm that arLLMs critically rely on extensive data augmentation via paraphrases for QA generalization, and paraphrases are only effective when their information order matches the QA style. Conversely, dLLMs achieve high accuracies on both forward and backward QAs without paraphrases; adding paraphrases yields only marginal gains. Inspired by the dLLM’s performance, we introduce a novel masked fine-tuning paradigm for knowledge injection into pre-trained arLLMs. This proposed method successfully and drastically improves the data efficiency of arLLM fine-tuning, effectively closing its performance gap with dLLMs. We further show that the masked fine-tuning paradigm of arLLMs can be extended to the supervised fine-tuning (SFT) of mathematical capability. Across two models and two datasets, our masked SFT outperforms regular SFT.
[120] DND: Boosting Large Language Models with Dynamic Nested Depth
Tieyuan Chen, Xiaodong Chen, Haoxing Chen, Zhenzhong Lan, Weiyao Lin, Jianguo Li
Main category: cs.CL
TL;DR: DND improves LLM performance by dynamically selecting critical tokens for reprocessing in a nested depth manner, boosting accuracy with minimal computational overhead.
Details
Motivation: To enhance off-the-shelf LLM performance without extensive retraining by efficiently reprocessing only the most challenging tokens rather than applying uniform computation to all tokens.Method: Dynamic Nested Depth (DND) identifies critical tokens at transformer layer ends using a router with loss control for better distinguishability, then feeds them back for extra processing via a threshold control scheme for stability.
Result: Improves dense Qwen3-1.7B by 1.88% and MoE Qwen3-30B-A3B by 0.87% on diverse benchmarks with minimal parameter and computing increase, demonstrating effectiveness across both dense and mixture-of-experts models.
Conclusion: DND provides an efficient post-training enhancement method that selectively reprocesses difficult tokens, offering significant performance gains with minimal computational overhead for both dense and MoE LLMs.
Abstract: We introduce Dynamic Nested Depth (DND), a novel method that improves performance for off-the-shelf LLMs by selecting critical tokens to reprocess in a nested depth manner. Specifically, at the end of the given transformer layer, DND identifies more critical tokens with a router and feeds them back for an extra round of processing, effectively ``reviewing" difficult tokens while avoiding redundant computation for easier ones. The dynamic selection mechanism is tailored for precise control via two novel strategies: a router controlling loss to enhance token selection distinguishability, and a threshold control scheme to ensure selection stability. We demonstrate the effectiveness of DND by directly integrating it into pre-trained dense and MoE models during a post-training phase. On diverse benchmarks, this approach boosts the performances of the dense Qwen3-1.7B by 1.88% and the MoE Qwen3-30B-A3B by 0.87%, all with a minimal parameter and computing increase.
[121] Recursive numeral systems are highly regular and easy to process
Ponrawee Prasertsom, Andrea Silvi, Jennifer Culbertson, Moa Johansson, Devdatt Dubhashi, Kenny Smith
Main category: cs.CL
TL;DR: The paper argues that recursive numeral systems optimize regularity and processing complexity, not just lexicon size vs. morphosyntactic complexity, and that MDL-based measures better distinguish natural from unnatural systems.
Details
Motivation: Previous work failed to show that only natural-language-like recursive numeral systems optimize the trade-off between lexicon size and average morphosyntactic complexity, requiring ad-hoc constraints to rule out unnatural systems. The authors argue this issue arises because regularity - a crucial aspect of complexity in human grammars - has been neglected.Method: The authors use the Minimum Description Length (MDL) approach to propose that recursive numeral systems should be evaluated based on their regularity and processing complexity. They develop MDL-based measures of regularity and processing complexity to analyze numeral systems.
Result: The MDL-based measures of regularity and processing complexity better capture the differences between attested natural numeral systems and unattested but possible ones, including “optimal” recursive numeral systems from previous work. The ad-hoc constraints from previous literature naturally emerge from considerations of regularity.
Conclusion: Recursive numeral systems are efficient with regard to regularity and processing complexity, not just lexicon size vs. morphosyntactic complexity. The study highlights the importance of incorporating regularity across sets of forms when measuring and explaining optimality in language.
Abstract: Previous work has argued that recursive numeral systems optimise the trade-off between lexicon size and average morphosyntatic complexity (Denić and Szymanik, 2024). However, showing that only natural-language-like systems optimise this tradeoff has proven elusive, and the existing solution has relied on ad-hoc constraints to rule out unnatural systems (Yang and Regier, 2025). Here, we argue that this issue arises because the proposed trade-off has neglected regularity, a crucial aspect of complexity central to human grammars in general. Drawing on the Minimum Description Length (MDL) approach, we propose that recursive numeral systems are better viewed as efficient with regard to their regularity and processing complexity. We show that our MDL-based measures of regularity and processing complexity better capture the key differences between attested, natural systems and unattested but possible ones, including “optimal” recursive numeral systems from previous work, and that the ad-hoc constraints from previous literature naturally follow from regularity. Our approach highlights the need to incorporate regularity across sets of forms in studies that attempt to measure and explain optimality in language.
[122] Prompt-R1: Collaborative Automatic Prompting Framework via End-to-end Reinforcement Learning
Wenjin Liu, Haoran Luo, Xueyuan Lin, Haoming Liu, Tiesunlong Shen, Jiapu Wang, Rui Mao, Erik Cambria
Main category: cs.CL
TL;DR: Prompt-R1 is a reinforcement learning framework that uses a small LLM to generate prompts for a large LLM, improving performance on complex tasks without requiring users to craft effective prompts.
Details
Motivation: Users often struggle to provide accurate and effective prompts for large language models when facing complex problems, which limits the performance of LLMs despite their advanced capabilities.Method: An end-to-end reinforcement learning framework where a small-scale LLM collaborates with a large-scale LLM through multi-turn prompt interactions. The small LLM thinks and generates prompts, while the large LLM performs complex reasoning. A dual-constrained reward optimizes for correctness, generation quality, and reasoning accuracy.
Result: Experiments on multiple public datasets show that Prompt-R1 significantly outperforms baseline models across various tasks.
Conclusion: Prompt-R1 provides a plug-and-play framework that supports both inference and training with various large-scale LLMs, effectively addressing the challenge of prompt engineering for complex problems.
Abstract: Recently, advanced large language models (LLMs) have emerged at an increasingly rapid pace. However, when faced with complex problems, most users are often unable to provide accurate and effective prompts to interact with LLMs, thus limiting the performance of LLMs. To address this challenge, we propose Prompt-R1, an end-to-end reinforcement learning framework that uses a small-scale LLM to collaborate with large-scale LLMs, replacing user interaction to solve problems better. This collaboration is cast as a multi-turn prompt interaction, where the small-scale LLM thinks and generates prompts, and the large-scale LLM performs complex reasoning. A dual-constrained reward is designed to optimize for correctness, generation quality, and reasoning accuracy. Prompt-R1 provides a plug-and-play framework that supports both inference and training with various large-scale LLMs. Experiments on multiple public datasets show that Prompt-R1 significantly outperforms baseline models across tasks. Our code is publicly available at https://github.com/QwenQKing/Prompt-R1.
[123] EQ-Negotiator: Dynamic Emotional Personas Empower Small Language Models for Edge-Deployable Credit Negotiation
Yunbo Long, Yuhan Liu, Alexandra Brintrup
Main category: cs.CL
TL;DR: EQ-Negotiator enables small language models to outperform much larger LLMs in credit negotiations by using emotional personas and game theory, achieving better debt recovery while preserving privacy.
Details
Motivation: LLMs are too computationally expensive and privacy-invasive for on-device applications like mobile assistants and private client interactions. Small language models (SLMs) are more practical but have a performance gap in handling emotionally charged negotiations like credit collection.Method: EQ-Negotiator framework integrates game theory with a Hidden Markov Model (HMM) to learn and track debtor emotional states online without pre-training. This reasoning system equips SLMs with strategic intelligence to counter manipulation while de-escalating conflict and upholding ethics.
Result: A 7B parameter language model with EQ-Negotiator achieves better debt recovery and negotiation efficiency than baseline LLMs more than 10 times its size, across diverse credit negotiation scenarios including adversarial strategies like cheating, threatening, and playing the victim.
Conclusion: Strategic emotional intelligence, not raw model scale, is critical for automated negotiation success. EQ-Negotiator advances persona modeling from descriptive profiles to dynamic emotional architectures that enable effective, ethical, privacy-preserving AI negotiators for edge computing.
Abstract: The deployment of large language models (LLMs) in automated negotiation has set a high performance benchmark, but their computational cost and data privacy requirements render them unsuitable for many privacy-sensitive, on-device applications such as mobile assistants, embodied AI agents or private client interactions. While small language models (SLMs) offer a practical alternative, they suffer from a significant performance gap compared to LLMs in playing emotionally charged complex personas, especially for credit negotiation. This paper introduces EQ-Negotiator, a novel framework that bridges this capability gap using emotional personas. Its core is a reasoning system that integrates game theory with a Hidden Markov Model(HMM) to learn and track debtor emotional states online, without pre-training. This allows EQ-Negotiator to equip SLMs with the strategic intelligence to counter manipulation while de-escalating conflict and upholding ethical standards. Through extensive agent-to-agent simulations across diverse credit negotiation scenarios, including adversarial debtor strategies like cheating, threatening, and playing the victim, we show that a 7B parameter language model with EQ-Negotiator achieves better debt recovery and negotiation efficiency than baseline LLMs more than 10 times its size. This work advances persona modeling from descriptive character profiles to dynamic emotional architectures that operate within privacy constraints. Besides, this paper establishes that strategic emotional intelligence, not raw model scale, is the critical factor for success in automated negotiation, paving the way for effective, ethical, and privacy-preserving AI negotiators that can operate on the edge.
[124] DPRM: A Dual Implicit Process Reward Model in Multi-Hop Question Answering
Xinyi Wang, Yiping Song, Zhiliang Tian, Bo Liu, Tingjin Luo, Minlie Huang
Main category: cs.CL
TL;DR: DPRM introduces a dual implicit process reward model that trains separate PRMs for Chain of Thought and Knowledge Graph reasoning in multi-hop QA, using outcome signals without explicit annotations, with structural constraints and consistency verification between reasoning paths.
Details
Motivation: Existing implicit PRMs only work for plain text and cannot handle graph structure constraints in KGs or capture inconsistencies between CoT and KG paths in multi-hop QA tasks.Method: Proposes DPRM with two implicit PRMs: KG-PRM (learns structural constraints from KGs using preference pairs) and CoT-PRM (for Chain of Thought reasoning). Both derive step-level rewards from outcome signals via reward parameterization without explicit annotations. Introduces consistency constraint between CoT and KG reasoning steps for mutual verification and collaborative optimization.
Result: Outperforms 13 baselines on multiple datasets with up to 16.6% improvement on Hit@1 metric.
Conclusion: DPRM effectively addresses limitations of existing implicit PRMs by handling graph structure constraints and ensuring consistency between different reasoning paths in multi-hop QA, achieving significant performance improvements.
Abstract: In multi-hop question answering (MHQA) tasks, Chain of Thought (CoT) improves the quality of generation by guiding large language models (LLMs) through multi-step reasoning, and Knowledge Graphs (KGs) reduce hallucinations via semantic matching. Outcome Reward Models (ORMs) provide feedback after generating the final answers but fail to evaluate the process for multi-step reasoning. Traditional Process Reward Models (PRMs) evaluate the reasoning process but require costly human annotations or rollout generation. While implicit PRM is trained only with outcome signals and derives step rewards through reward parameterization without explicit annotations, it is more suitable for multi-step reasoning in MHQA tasks. However, existing implicit PRM has only been explored for plain text scenarios. When adapting to MHQA tasks, it cannot handle the graph structure constraints in KGs and capture the potential inconsistency between CoT and KG paths. To address these limitations, we propose the DPRM (Dual Implicit Process Reward Model). It trains two implicit PRMs for CoT and KG reasoning in MHQA tasks. Both PRMs, namely KG-PRM and CoT-PRM, derive step-level rewards from outcome signals via reward parameterization without additional explicit annotations. Among them, KG-PRM uses preference pairs to learn structural constraints from KGs. DPRM further introduces a consistency constraint between CoT and KG reasoning steps, making the two PRMs mutually verify and collaboratively optimize the reasoning paths. We also provide a theoretical demonstration of the derivation of process rewards. Experimental results show that our method outperforms 13 baselines on multiple datasets with up to 16.6% improvement on Hit@1.
[125] Black-Box On-Policy Distillation of Large Language Models
Tianzhu Ye, Li Dong, Zewen Chi, Xun Wu, Shaohan Huang, Furu Wei
Main category: cs.CL
TL;DR: GAD (Generative Adversarial Distillation) enables on-policy black-box distillation of LLMs by framing the student as a generator and training a discriminator to distinguish its outputs from the teacher’s, creating a minimax game.
Details
Motivation: Black-box distillation of LLMs traditionally relies on learning from teacher model outputs alone without access to internal logits or parameters, which can be limiting. The authors aim to create a more effective distillation method that works in this constrained setting.Method: GAD frames the student LLM as a generator and trains a discriminator to distinguish student responses from teacher responses. This creates a minimax game where the discriminator acts as an on-policy reward model that co-evolves with the student, providing stable, adaptive feedback for distillation.
Result: GAD consistently surpasses commonly used sequence-level knowledge distillation. Qwen2.5-14B-Instruct trained with GAD becomes comparable to its teacher, GPT-5-Chat, on the LMSYS-Chat automatic evaluation.
Conclusion: GAD establishes a promising and effective paradigm for black-box LLM distillation, enabling student models to achieve performance comparable to proprietary teacher models using only text outputs.
Abstract: Black-box distillation creates student large language models (LLMs) by learning from a proprietary teacher model’s text outputs alone, without access to its internal logits or parameters. In this work, we introduce Generative Adversarial Distillation (GAD), which enables on-policy and black-box distillation. GAD frames the student LLM as a generator and trains a discriminator to distinguish its responses from the teacher LLM’s, creating a minimax game. The discriminator acts as an on-policy reward model that co-evolves with the student, providing stable, adaptive feedback. Experimental results show that GAD consistently surpasses the commonly used sequence-level knowledge distillation. In particular, Qwen2.5-14B-Instruct (student) trained with GAD becomes comparable to its teacher, GPT-5-Chat, on the LMSYS-Chat automatic evaluation. The results establish GAD as a promising and effective paradigm for black-box LLM distillation.
[126] Spark-Prover-X1: Formal Theorem Proving Through Diverse Data Training
Xinyuan Zhou, Yi Lei, Xiaoyu Zhou, Jingyi Sun, Yu Zhu, Zhongyi Ye, Weitai Zhang, Quan Liu, Si Wei, Cong Liu
Main category: cs.CL
TL;DR: Spark-Prover-X1 is a 7B parameter LLM for theorem proving, trained via a three-stage framework with novel data tasks and benchmarked on new ExamFormal-Bench, achieving SOTA performance among similarly-sized models.
Details
Motivation: Progress in LLM-based automated theorem proving is constrained by scarcity of diverse, high-quality formal language data, limiting the reasoning potential of moderately-sized models.Method: Three-stage framework: 1) Continuous pre-training on broad mathematical corpus with novel tasks including “CoT-augmented state prediction” for fine-grained reasoning; 2) Supervised Fine-tuning with expert iteration loop for Spark-Prover-X1-7B and Spark-Formalizer-X1-7B; 3) Targeted Group Relative Policy Optimization (GRPO) on challenging problems.
Result: Spark-Prover achieves SOTA performance among similarly-sized open-source models in “Whole-Proof Generation” paradigm: solves 27 problems on PutnamBench (pass@32) and achieves 24.0% on CombiBench (pass@32).
Conclusion: Diverse training data and progressively refined training pipeline effectively enhance formal reasoning capabilities of lightweight LLMs; models and ExamFormal-Bench dataset will be released.
Abstract: Large Language Models (LLMs) have shown significant promise in automated theorem proving, yet progress is often constrained by the scarcity of diverse and high-quality formal language data. To address this issue, we introduce Spark-Prover-X1, a 7B parameter model trained via an three-stage framework designed to unlock the reasoning potential of more accessible and moderately-sized LLMs. The first stage infuses deep knowledge through continuous pre-training on a broad mathematical corpus, enhanced by a suite of novel data tasks. Key innovation is a “CoT-augmented state prediction” task to achieve fine-grained reasoning. The second stage employs Supervised Fine-tuning (SFT) within an expert iteration loop to specialize both the Spark-Prover-X1-7B and Spark-Formalizer-X1-7B models. Finally, a targeted round of Group Relative Policy Optimization (GRPO) is applied to sharpen the prover’s capabilities on the most challenging problems. To facilitate robust evaluation, particularly on problems from real-world examinations, we also introduce ExamFormal-Bench, a new benchmark dataset of 402 formal problems. Experimental results demonstrate that Spark-Prover achieves state-of-the-art performance among similarly-sized open-source models within the “Whole-Proof Generation” paradigm. It shows exceptional performance on difficult competition benchmarks, notably solving 27 problems on PutnamBench (pass@32) and achieving 24.0% on CombiBench (pass@32). Our work validates that this diverse training data and progressively refined training pipeline provides an effective path for enhancing the formal reasoning capabilities of lightweight LLMs. We will release both Spark-Prover-X1-7B and Spark-Formalizer-X1-7B, along with the ExamFormal-Bench dataset, in the near future.
[127] COMPASS: Context-Modulated PID Attention Steering System for Hallucination Mitigation
Kenji Sahay, Snigdha Pandya, Rohan Nagale, Anna Lin, Shikhar Shiromani, Kevin Zhu, Dev Sunishchal
Main category: cs.CL
TL;DR: COMPASS introduces a lightweight feedback control system that dynamically modulates attention heads during LLM decoding to reduce factual hallucinations by steering context reliance.
Details
Motivation: LLMs often generate fluent but factually incorrect statements despite having access to relevant evidence, due to improper attention allocation between contextual and parametric knowledge. Understanding and steering this internal behavior is crucial for trustworthy deployment and scientific interpretability.Method: COMPASS (Context-Modulated PID Attention Steering System) embeds a model-based feedback loop directly within decoding. It quantifies context reliance via a transparent Context Reliance Score (CRS) that probes how attention heads ground generation in evidence. A PID controller then dynamically modulates attention heads to maintain factual consistency without retraining or multi-pass decoding.
Result: Across benchmarks (HotpotQA, XSum, HaluEval, RAGTruth), COMPASS consistently reduces contextual hallucination rates by 2.8 to 5.8 percent absolute. The system reveals how distinct attention heads contribute to evidence alignment.
Conclusion: Feedback-driven interpretability provides a pathway toward scientific understanding of LLM behavior, demonstrating that lightweight control frameworks can effectively steer attention allocation to reduce factual inconsistencies without requiring retraining.
Abstract: Large language models (LLMs) often generate fluent but factually incorrect statements despite having access to relevant evidence, a failure mode rooted in how they allocate attention between contextual and parametric knowledge. Understanding and steering this internal behavior is key both for trustworthy deployment and for scientific interpretability of model mechanisms. We introduce COMPASS (Context-Modulated PID Attention Steering System), a lightweight, interpretable control framework that embeds a model-based feedback loop directly within decoding. COMPASS quantifies context reliance via a transparent metric, the Context Reliance Score (CRS), which serves as an online probe of how attention heads ground generation in evidence. Using this interpretable signal, a PID controller dynamically modulates attention heads to maintain factual consistency without retraining or multi-pass decoding. Across benchmarks (HotpotQA, XSum, HaluEval, RAGTruth), COMPASS consistently reduces contextual hallucination rates (2.8 to 5.8 percent absolute) while revealing how distinct attention heads contribute to evidence alignment. These results highlight feedback-driven interpretability as a pathway toward scientific understanding of LLM behavior.
[128] PEPPER: Perception-Guided Perturbation for Robust Backdoor Defense in Text-to-Image Diffusion Models
Oscar Chew, Po-Yi Lu, Jayden Lin, Kuan-Hao Huang, Hsuan-Tien Lin
Main category: cs.CL
TL;DR: PEPPER is a defense against backdoor attacks in text-to-image diffusion models that rewrites input captions to disrupt triggers while preserving visual similarity.
Details
Motivation: Text-to-image diffusion models are vulnerable to backdoor attacks where triggers in input prompts can steer generation toward harmful content, creating security risks that need to be addressed.Method: PEPPER rewrites input captions into semantically distant but visually similar versions while adding unobtrusive elements, disrupting embedded triggers and diluting trigger token influence.
Result: PEPPER is particularly effective against text encoder based attacks, substantially reducing attack success while preserving generation quality, and can be combined with existing defenses for stronger robustness.
Conclusion: PEPPER provides an effective defense against backdoor attacks in T2I models through caption rewriting, offering enhanced robustness that can be further strengthened when combined with other defense methods.
Abstract: Recent studies show that text to image (T2I) diffusion models are vulnerable to backdoor attacks, where a trigger in the input prompt can steer generation toward harmful or unintended content. To address this, we introduce PEPPER (PErcePtion Guided PERturbation), a backdoor defense that rewrites the caption into a semantically distant yet visually similar caption while adding unobstructive elements. With this rewriting strategy, PEPPER disrupt the trigger embedded in the input prompt, dilute the influence of trigger tokens and thereby achieve enhanced robustness. Experiments show that PEPPER is particularly effective against text encoder based attacks, substantially reducing attack success while preserving generation quality. Beyond this, PEPPER can be paired with any existing defenses yielding consistently stronger and generalizable robustness than any standalone method. Our code will be released on Github.
[129] PARROT: Persuasion and Agreement Robustness Rating of Output Truth – A Sycophancy Robustness Benchmark for LLMs
Yusuf Çelebi, Özay Ezerceli, Mahmoud El Hussieni
Main category: cs.CL
TL;DR: PARROT framework measures LLM accuracy degradation under social pressure/sycophancy, showing advanced models resist authority while older/smaller models collapse epistemically.
Details
Motivation: To systematically measure how LLMs degrade in accuracy when subjected to social pressure/authority (sycophancy), which is crucial for safe real-world deployment beyond just accuracy metrics.Method: PARROT framework: (i) double-blind comparison of neutral vs. authoritatively false versions of same questions, (ii) log-likelihood calibration tracking of confidence shifts, (iii) eight-state behavioral taxonomy classification of failure modes. Evaluated 22 models on 1,302 MMLU-style questions across 13 domains with authority templates.
Result: Advanced models (GPT-5, GPT-4.1, Claude Sonnet 4.5) show low follow rates (≤11%) and minimal accuracy loss, while older/smaller models exhibit severe epistemic collapse (GPT-4: 80%, Qwen 2.5-1.5B: 94%). Weak models reduce confidence in correct answers while increasing confidence in imposed incorrect ones. International law/global knowledge domains are fragile; elementary math is resilient.
Conclusion: Resistance to overfitting pressure should be a primary objective alongside accuracy, harm avoidance, and privacy for safe LLM deployment in real-world settings where social pressure exists.
Abstract: This study presents PARROT (Persuasion and Agreement Robustness Rating of Output Truth), a robustness focused framework designed to measure the degradation in accuracy that occurs under social pressure exerted on users through authority and persuasion in large language models (LLMs) the phenomenon of sycophancy (excessive conformity). PARROT (i) isolates causal effects by comparing the neutral version of the same question with an authoritatively false version using a double-blind evaluation, (ii) quantifies confidence shifts toward the correct and imposed false responses using log-likelihood-based calibration tracking, and (iii) systematically classifies failure modes (e.g., robust correct, sycophantic agreement, reinforced error, stubborn error, self-correction, etc.) using an eight-state behavioral taxonomy. We evaluated 22 models using 1,302 MMLU-style multiple-choice questions across 13 domains and domain-specific authority templates. Findings show marked heterogeneity: advanced models (e.g., GPT-5, GPT-4.1, Claude Sonnet 4.5) exhibit low “follow rates” ($\leq 11%$, GPT-5: 4%) and minimal accuracy loss, while older/smaller models show severe epistemic collapse (GPT-4: 80%, Qwen 2.5-1.5B: 94%). The danger is not limited to response changes; weak models reduce confidence in the correct response while increasing confidence in the imposed incorrect response. While international law and global knowledge at the domain level exhibit high fragility, elementary mathematics is relatively resilient. Consequently, we argue that the goal of “resistance to overfitting pressure” should be addressed as a primary objective alongside accuracy, harm avoidance, and privacy for safe deployment in the real world.
[130] Evaluating Large Language Models on the 2026 Korean CSAT Mathematics Exam: Measuring Mathematical Ability in a Zero-Data-Leakage Setting
Goun Pyeon, Inbum Heo, Jeesu Jung, Taewook Hwang, Hyuk Namgoong, Hyein Seo, Yerim Han, Eunbin Kim, Hyeonseok Kang, Sangkeun Jung
Main category: cs.CL
TL;DR: LLMs evaluated on 2026 Korean CSAT math with contamination-free setup; GPT-5 achieved perfect scores, smaller models like gpt-oss-20B showed high cost-effectiveness; text input outperformed images, and excessive reasoning improved scores but reduced efficiency.
Details
Motivation: To address data leakage issues in existing benchmarks and provide a completely contamination-free evaluation of LLMs' mathematical reasoning capabilities using real-world exam questions that couldn't have been in training data.Method: Digitized all 46 questions from the 2026 Korean CSAT Mathematics section within 2 hours of public release; evaluated 24 state-of-the-art LLMs across different input modalities (Text-only, Image-only, Text+Figure) and prompt languages (Korean, English); conducted reasoning enhancement experiments with GPT-5 series.
Result: GPT-5 family achieved perfect scores (100 points) in certain configurations; Grok 4, Qwen 3 235B, and Gemini 2.5 pro scored above 97 points; gpt-oss-20B achieved 95.7 points despite small size; Calculus was the weakest domain; text input consistently outperformed image input; increased reasoning intensity improved performance but drastically reduced efficiency.
Conclusion: The study establishes a contamination-free evaluation framework, demonstrates that smaller models can be highly cost-effective, reveals domain-specific weaknesses in LLMs, and suggests that practical deployment should balance performance with computational efficiency rather than maximizing reasoning intensity.
Abstract: This study systematically evaluated the mathematical reasoning capabilities of Large Language Models (LLMs) using the 2026 Korean College Scholastic Ability Test (CSAT) Mathematics section, ensuring a completely contamination-free evaluation environment. To address data leakage issues in existing benchmarks, we digitized all 46 questions (22 common and 24 elective) within two hours of the exam’s public release, eliminating any possibility of inclusion in model training data. We conducted comprehensive evaluations of 24 state-of-the-art LLMs across varying input modalities (Text-only, Image-only, Text+Figure) and prompt languages (Korean, English). The GPT-5 family models achieved perfect scores (100 points) under a limited set of language-modality configurations, while Grok 4, Qwen 3 235B, and Gemini 2.5 pro also scored above 97 points. Notably, gpt-oss-20B achieved 95.7 points despite its relatively small size, demonstrating high cost-effectiveness. Problem-specific analysis revealed Calculus as the weakest domain with significant performance degradation on 4-point high-difficulty problems. Text input consistently outperformed image input, while prompt language effects varied by model scale. In reasoning enhancement experiments with GPT-5 series, increased reasoning intensity improved performance (82.6->100 points) but quadrupled token usage and drastically reduced efficiency, suggesting that models with minimal reasoning may be more practical. This research contributes: (1) implementation of a completely unexposed evaluation environment, (2) a standardized digitization pipeline that converts human-targeted exam materials into LLM-ready evaluation data, and (3) a practical evaluation perspective integrating performance, cost, and time considerations. Detailed results and model comparisons are available at the 2026 Korean CSAT LLM Evaluation Leaderboard; https://isoft.cnu.ac.kr/csat2026/
[131] Robust Multimodal Sentiment Analysis with Distribution-Based Feature Recovery and Fusion
Daiqing Wu, Dongbao Yang, Yu Zhou, Can Ma
Main category: cs.CL
TL;DR: DRF method for robust multimodal sentiment analysis using distribution-based feature recovery and fusion to handle low-quality and missing modalities in image-text pairs.
Details
Motivation: Existing multimodal sentiment analysis methods lack consideration for low-quality and missing modalities, which frequently occur in real-world social media applications, creating need for robust models.Method: Distribution-based feature Recovery and Fusion (DRF) maintains feature queues for each modality to approximate distributions, reduces contributions of low-quality modalities based on quality estimation, and recovers missing modalities through inter-modal mapping supervised by samples and distributions.
Result: DRF demonstrates universal improvements compared to SOTA methods on three publicly available image-text datasets under disruption strategies mimicking real-world low-quality and missing modalities.
Conclusion: DRF provides an effective unified framework for robust multimodal sentiment analysis that handles both low-quality and missing modalities in real-world scenarios.
Abstract: As posts on social media increase rapidly, analyzing the sentiments embedded in image-text pairs has become a popular research topic in recent years. Although existing works achieve impressive accomplishments in simultaneously harnessing image and text information, they lack the considerations of possible low-quality and missing modalities. In real-world applications, these issues might frequently occur, leading to urgent needs for models capable of predicting sentiment robustly. Therefore, we propose a Distribution-based feature Recovery and Fusion (DRF) method for robust multimodal sentiment analysis of image-text pairs. Specifically, we maintain a feature queue for each modality to approximate their feature distributions, through which we can simultaneously handle low-quality and missing modalities in a unified framework. For low-quality modalities, we reduce their contributions to the fusion by quantitatively estimating modality qualities based on the distributions. For missing modalities, we build inter-modal mapping relationships supervised by samples and distributions, thereby recovering the missing modalities from available ones. In experiments, two disruption strategies that corrupt and discard some modalities in samples are adopted to mimic the low-quality and missing modalities in various real-world scenarios. Through comprehensive experiments on three publicly available image-text datasets, we demonstrate the universal improvements of DRF compared to SOTA methods under both two strategies, validating its effectiveness in robust multimodal sentiment analysis.
[132] A Machine Learning Approach for Detection of Mental Health Conditions and Cyberbullying from Social Media
Edward Ajayi, Martha Kachweka, Mawuli Deku, Emily Aiken
Main category: cs.CL
TL;DR: A unified multiclass classification framework detects 10 mental health and cyberbullying categories from social media using domain-adapted MentalBERT, achieving 0.92 accuracy with human-in-the-loop explainable AI.
Details
Motivation: Increasing prevalence of mental health challenges and cyberbullying in digital spaces requires scalable, interpretable detection systems for online safety and computational mental health applications.Method: Unified multiclass classification framework using curated Twitter/Reddit datasets with “split-then-balance” pipeline; comprehensive evaluation of lexical models, hybrid approaches, and fine-tuned transformers; introduces hybrid SHAP-LLM explainability framework and prototype dashboard for human-in-the-loop screening.
Result: Domain-adapted MentalBERT emerged as top performer with 0.92 accuracy and 0.76 Macro F1 score, surpassing generic transformers and zero-shot LLM baselines; framework successfully detects 10 distinct mental health and cyberbullying categories.
Conclusion: End-to-end fine-tuning is critical for performance; system is framed as human-in-the-loop screening aid (not diagnostic tool) with explainable AI; provides robust baseline highlighting need for multi-label, clinically-validated datasets at intersection of online safety and computational mental health.
Abstract: Mental health challenges and cyberbullying are increasingly prevalent in digital spaces, necessitating scalable and interpretable detection systems. This paper introduces a unified multiclass classification framework for detecting ten distinct mental health and cyberbullying categories from social media data. We curate datasets from Twitter and Reddit, implementing a rigorous “split-then-balance” pipeline to train on balanced data while evaluating on a realistic, held-out imbalanced test set. We conducted a comprehensive evaluation comparing traditional lexical models, hybrid approaches, and several end-to-end fine-tuned transformers. Our results demonstrate that end-to-end fine-tuning is critical for performance, with the domain-adapted MentalBERT emerging as the top model, achieving an accuracy of 0.92 and a Macro F1 score of 0.76, surpassing both its generic counterpart and a zero-shot LLM baseline. Grounded in a comprehensive ethical analysis, we frame the system as a human-in-the-loop screening aid, not a diagnostic tool. To support this, we introduce a hybrid SHAPLLM explainability framework and present a prototype dashboard (“Social Media Screener”) designed to integrate model predictions and their explanations into a practical workflow for moderators. Our work provides a robust baseline, highlighting future needs for multi-label, clinically-validated datasets at the critical intersection of online safety and computational mental health.
[133] Adversarial Confusion Attack: Disrupting Multimodal Large Language Models
Jakub Hoscilowicz, Artur Janicki
Main category: cs.CL
TL;DR: Adversarial Confusion Attack disrupts MLLMs by maximizing next-token entropy, causing incoherent outputs that transfer to both open-source and proprietary models.
Details
Motivation: Current adversarial attacks focus on jailbreaks or targeted misclassification, but there's a need for systematic disruption attacks that make MLLMs generate incoherent or confidently incorrect outputs, preventing reliable operation of MLLM-powered AI agents.Method: Uses Projected Gradient Descent (PGD) to maximize next-token entropy using a small ensemble of open-source MLLMs. Works in both full-image and Adversarial CAPTCHA settings, with single adversarial images disrupting all models in the ensemble.
Result: The attack successfully disrupts all models in the ensemble and transfers to unseen open-source (Qwen3-VL) and proprietary (GPT-5.1) models, demonstrating strong transferability despite using basic adversarial techniques.
Conclusion: Adversarial Confusion Attack represents a new class of threats against MLLMs that can systematically disrupt model performance, highlighting vulnerabilities in multimodal systems and the need for robust defenses against such attacks.
Abstract: We introduce the Adversarial Confusion Attack, a new class of threats against multimodal large language models (MLLMs). Unlike jailbreaks or targeted misclassification, the goal is to induce systematic disruption that makes the model generate incoherent or confidently incorrect outputs. Practical applications include embedding such adversarial images into websites to prevent MLLM-powered AI Agents from operating reliably. The proposed attack maximizes next-token entropy using a small ensemble of open-source MLLMs. In the white-box setting, we show that a single adversarial image can disrupt all models in the ensemble, both in the full-image and Adversarial CAPTCHA settings. Despite relying on a basic adversarial technique (PGD), the attack generates perturbations that transfer to both unseen open-source (e.g., Qwen3-VL) and proprietary (e.g., GPT-5.1) models.
cs.CV
[134] MOTION: ML-Assisted On-Device Low-Latency Motion Recognition
Veeramani Pugazhenthi, Wei-Hsiang Chu, Junwei Lu, Jadyn N. Miyahira, Soheil Salehi
Main category: cs.CV
TL;DR: The paper presents an efficient motion-based gesture recognition system using triaxial accelerometer sensors and AutoML pipelines, achieving reliable real-time performance on wearable devices for medical monitoring applications.
Details
Motivation: There is growing demand for tiny devices with low-latency gesture recognition capabilities in human-computer interaction and medical monitoring fields like fall detection, rehabilitation tracking, and patient supervision, requiring fast, efficient movement tracking while minimizing false alarms.Method: The study uses AutoML pipelines to extract important features from data segments collected by triaxial accelerometer sensors, then trains multiple lightweight machine learning algorithms using these features. The system is implemented on WeBe Band, a multi-sensor wearable device with a powerful MCU capable of on-device gesture recognition.
Result: Among the explored models, neural networks provided the best balance between accuracy, latency, and memory use. The system demonstrates reliable real-time gesture recognition on WeBe Band, showing great potential for secure, fast-response medical monitoring solutions.
Conclusion: The research successfully demonstrates that efficient motion-based models using only triaxial accelerometer sensors can achieve reliable real-time gesture recognition on wearable devices, offering promising solutions for medical monitoring applications requiring secure and fast response times.
Abstract: The use of tiny devices capable of low-latency gesture recognition is gaining momentum in everyday human-computer interaction and especially in medical monitoring fields. Embedded solutions such as fall detection, rehabilitation tracking, and patient supervision require fast and efficient tracking of movements while avoiding unwanted false alarms. This study presents an efficient solution on how to build very efficient motion-based models only using triaxial accelerometer sensors. We explore the capability of the AutoML pipelines to extract the most important features from the data segments. This approach also involves training multiple lightweight machine learning algorithms using the extracted features. We use WeBe Band, a multi-sensor wearable device that is equipped with a powerful enough MCU to effectively perform gesture recognition entirely on the device. Of the models explored, we found that the neural network provided the best balance between accuracy, latency, and memory use. Our results also demonstrate that reliable real-time gesture recognition can be achieved in WeBe Band, with great potential for real-time medical monitoring solutions that require a secure and fast response time.
[135] Closing the Gap: Data-Centric Fine-Tuning of Vision Language Models for the Standardized Exam Questions
Egemen Sert, Şeyda Ertekin
Main category: cs.CV
TL;DR: This paper shows that supervised fine-tuning with high-quality multimodal data can achieve near state-of-the-art performance on exam question reasoning, rivaling proprietary models, through careful data curation and optimized reasoning syntax.
Details
Motivation: Standardized exam questions provide rigorous testbeds for multimodal reasoning, but current research focuses too much on algorithmic advances while neglecting data-centric foundations. The authors aim to demonstrate that high-quality data curation can enable open-weight models to compete with proprietary approaches.Method: Compiled a 161.4 million token multimodal dataset combining textbook question-solution pairs, curriculum-aligned diagrams, and contextual materials. Fine-tuned Qwen-2.5VL-32B using an optimized reasoning syntax called QMSA. Evaluated on YKSUniform benchmark with 1,854 multimodal exam questions across 309 curriculum topics.
Result: The resulting model achieved 78.6% accuracy on YKSUniform benchmark, only 1.0% below Gemini 2.0 Flash. This demonstrates that supervised fine-tuning with carefully curated data can approach state-of-the-art performance.
Conclusion: Data composition and representational syntax play decisive roles in multimodal reasoning. The work establishes a data-centric framework showing that curriculum-grounded multimodal data can elevate supervised fine-tuning to near state-of-the-art performance for open-weight vision-language models.
Abstract: Multimodal reasoning has become a cornerstone of modern AI research. Standardized exam questions offer a uniquely rigorous testbed for such reasoning, providing structured visual contexts and verifiable answers. While recent progress has largely focused on algorithmic advances such as reinforcement learning (e.g., GRPO, DPO), the data centric foundations of vision language reasoning remain less explored. We show that supervised fine-tuning (SFT) with high-quality data can rival proprietary approaches. To this end, we compile a 161.4 million token multimodal dataset combining textbook question-solution pairs, curriculum aligned diagrams, and contextual materials, and fine-tune Qwen-2.5VL-32B using an optimized reasoning syntax (QMSA). The resulting model achieves 78.6% accuracy, only 1.0% below Gemini 2.0 Flash, on our newly released benchmark YKSUniform, which standardizes 1,854 multimodal exam questions across 309 curriculum topics. Our results reveal that data composition and representational syntax play a decisive role in multimodal reasoning. This work establishes a data centric framework for advancing open weight vision language models, demonstrating that carefully curated and curriculum-grounded multimodal data can elevate supervised fine-tuning to near state-of-the-art performance.
[136] PEFT-DML: Parameter-Efficient Fine-Tuning Deep Metric Learning for Robust Multi-Modal 3D Object Detection in Autonomous Driving
Abdolazim Rezaei, Mehdi Sookhak
Main category: cs.CV
TL;DR: PEFT-DML is a parameter-efficient deep metric learning framework for robust multi-modal 3D object detection in autonomous driving that handles diverse sensor modalities and maintains performance under sensor dropout or unseen modality combinations.
Details
Motivation: Conventional multi-modal 3D object detection models assume fixed sensor availability, but real-world autonomous driving scenarios often face sensor dropout, diverse modality combinations, and environmental challenges like weather variability and fast motion.Method: PEFT-DML maps diverse modalities (LiDAR, radar, camera, IMU, GNSS) into a shared latent space using parameter-efficient techniques like Low-Rank Adaptation (LoRA) and adapter layers, enabling robust detection under various sensor conditions.
Result: Experiments on nuScenes benchmarks demonstrate superior accuracy and robustness to sensor dropout, fast motion, weather variability, and domain shifts while achieving significant training efficiency.
Conclusion: PEFT-DML provides an efficient and robust solution for multi-modal 3D object detection in autonomous driving that can handle real-world sensor variability and environmental challenges while maintaining high accuracy.
Abstract: This study introduces PEFT-DML, a parameter-efficient deep metric learning framework for robust multi-modal 3D object detection in autonomous driving. Unlike conventional models that assume fixed sensor availability, PEFT-DML maps diverse modalities (LiDAR, radar, camera, IMU, GNSS) into a shared latent space, enabling reliable detection even under sensor dropout or unseen modality class combinations. By integrating Low-Rank Adaptation (LoRA) and adapter layers, PEFT-DML achieves significant training efficiency while enhancing robustness to fast motion, weather variability, and domain shifts. Experiments on benchmarks nuScenes demonstrate superior accuracy.
[137] DL-CapsNet: A Deep and Light Capsule Network
Pouya Shiri, Amirali Baniasadi
Main category: cs.CV
TL;DR: Deep Capsule Network (DL-CapsNet) with multiple capsule layers and a Capsule Summarization layer for reduced complexity, faster training, and improved accuracy on complex datasets.
Details
Motivation: Capsule Networks show promise as successors to CNNs with better accuracy on overlapping categories and affine-transformed images, but need improvements in complexity and scalability for complex datasets.Method: Proposes DL-CapsNet with multiple capsule layers and a Capsule Summarization layer to reduce parameters and complexity while maintaining accuracy.
Result: DL-CapsNet achieves high accuracy with fewer parameters, faster training/inference, and can handle complex datasets with many categories.
Conclusion: DL-CapsNet provides an efficient deep variant of CapsNet that balances accuracy and computational efficiency for complex image classification tasks.
Abstract: Capsule Network (CapsNet) is among the promising classifiers and a possible successor of the classifiers built based on Convolutional Neural Network (CNN). CapsNet is more accurate than CNNs in detecting images with overlapping categories and those with applied affine transformations. In this work, we propose a deep variant of CapsNet consisting of several capsule layers. In addition, we design the Capsule Summarization layer to reduce the complexity by reducing the number of parameters. DL-CapsNet, while being highly accurate, employs a small number of parameters and delivers faster training and inference. DL-CapsNet can process complex datasets with a high number of categories.
[138] Satellite to Street : Disaster Impact Estimator
Sreesritha Sai, Sai Venkata Suma Sreeja, Deepthi, Nikhil
Main category: cs.CV
TL;DR: A deep learning framework called “Satellite-to-Street: Disaster Impact Estimator” that uses modified dual-input U-Net with enhanced feature fusion to generate pixel-level damage maps from pre- and post-disaster satellite imagery, addressing class imbalance and subtle structural variations.
Details
Motivation: Manual satellite imagery interpretation for post-disaster damage assessment is slow, subjective, and hard to scale. Existing deep learning models struggle with subtle structural variations and severe class imbalance, leading to poor detection of highly damaged regions.Method: Modified dual-input U-Net architecture with enhanced feature fusion to jointly process pre- and post-disaster satellite images. Uses class-aware weighted loss functions to handle dominance of undamaged pixels and improve sensitivity to major and destroyed damage categories.
Result: Experiments on public disaster datasets show improved localization and classification of structural damage compared to traditional segmentation and baseline change-detection models.
Conclusion: The framework provides rapid, consistent damage assessment to support (not replace) expert decision-making, enabling more efficient, data-driven disaster management through fine-grained pixel-level damage maps.
Abstract: Accurate post-disaster damage assessment is of high importance for prioritizing emergency response; however, manual interpretation of satellite imagery is slow, subjective, and hard to scale. While deep-learning models for image segmentation, such as U-Net-based baselines and change-detection models, are useful baselines, they often struggle with subtle structural variations and severe class imbalance, yielding poor detection of highly damaged regions. The present work proposes a deep-learning framework that jointly processes pre- and post-disaster satellite images to obtain fine-grained pixel-level damage maps: Satellite-to-Street: Disaster Impact Estimator. The model uses a modified dual-input U-Net architecture with enhanced feature fusion to capture both the local structural changes as well as the broader contextual cues. Class-aware weighted loss functions are integrated in order to handle the dominance of undamaged pixels in real disaster datasets, thus enhancing sensitivity toward major and destroyed categories. Experimentation on publicly available disaster datasets shows improved localization and classification of structural damage when compared to traditional segmentation and baseline change-detection models. The resulting damage maps provide a rapid and consistent assessment mechanism to support and not replace expert decision-making, thus allowing more efficient, data-driven disaster management.
[139] ProvRain: Rain-Adaptive Denoising and Vehicle Detection via MobileNet-UNet and Faster R-CNN
Aswinkumar Varathakumaran, Nirmala Paramanandham
Main category: cs.CV
TL;DR: ProvRain pipeline improves night-time vehicle detection in rainy conditions using lightweight MobileNet-U-Net with curricula training, achieving 8.94% accuracy and 10.25% recall improvements over baseline.
Details
Motivation: Night-time vehicle detection faces challenges with noise from weather conditions (rain/snow) and camera issues, which degrade detection performance before vehicles become directly visible.Method: Developed ProvRain pipeline using lightweight MobileNet-U-Net architecture with curricula training to generalize to robust weather conditions, trained on synthetic and real PVDN dataset data, compared against Faster RCNN baseline.
Result: 8.94% increase in accuracy and 10.25% increase in recall for vehicle detection in rainy night-time frames; MobileNet-U-Net shows 10-15% PSNR improvement, 5-6% SSIM increase, and up to 67% reduction in perceptual error compared to transformer approaches.
Conclusion: The ProvRain pipeline effectively handles noise in night-time rainy conditions while maintaining detection accuracy, demonstrating significant improvements over baseline methods and superior performance compared to transformer-based approaches.
Abstract: Provident vehicle detection has a lot of scope in the detection of vehicle during night time. The extraction of features other than the headlamps of vehicles allows us to detect oncoming vehicles before they appear directly on the camera. However, it faces multiple issues especially in the field of night vision, where a lot of noise caused due to weather conditions such as rain or snow as well as camera conditions. This paper focuses on creating a pipeline aimed at dealing with such noise while at the same time maintaining the accuracy of provident vehicular detection. The pipeline in this paper, ProvRain, uses a lightweight MobileNet-U-Net architecture tuned to generalize to robust weather conditions by using the concept of curricula training. A mix of synthetic as well as available data from the PVDN dataset is used for this. This pipeline is compared to the base Faster RCNN architecture trained on the PVDN dataset to see how much the addition of a denoising architecture helps increase the detection model’s performance in rainy conditions. The system boasts an 8.94% increase in accuracy and a 10.25% increase in recall in the detection of vehicles in rainy night time frames. Similarly, the custom MobileNet-U-Net architecture that was trained also shows a 10-15% improvement in PSNR, a 5-6% increase in SSIM, and upto a 67% reduction in perceptual error (LPIPS) compared to other transformer approaches.
[140] Adapter Shield: A Unified Framework with Built-in Authentication for Preventing Unauthorized Zero-Shot Image-to-Image Generation
Jun Jia, Hongyi Miao, Yingjie Zhou, Wangqiu Zhou, Jianbo Zhang, Linhan Cao, Dandan Zhu, Hua Yang, Xiongkuo Min, Wei Sun, Guangtao Zhai
Main category: cs.CV
TL;DR: Adapter Shield is a universal defense system that protects personal images from unauthorized zero-shot image generation by using reversible encryption and adversarial perturbations to block misuse while allowing authorized access.
Details
Motivation: Zero-shot image-to-image generation enables high-fidelity replication of facial identities and artistic styles without model modification, creating significant intellectual property risks including unauthorized identity cloning and stylistic imitation.Method: The method investigates how zero-shot methods use image encoders to extract embeddings fed into diffusion models via cross-attention. It constructs a reversible encryption system mapping original embeddings to encrypted representations using secret keys, and designs multi-target adversarial perturbations to shift embeddings toward encrypted patterns.
Result: Extensive evaluations show the method surpasses existing state-of-the-art defenses in blocking unauthorized zero-shot image synthesis while supporting flexible and secure access control for verified users.
Conclusion: Adapter Shield provides the first universal and authentication-integrated solution for protecting personal images from misuse in zero-shot generation scenarios, effectively preventing unauthorized identity cloning and stylistic imitation while maintaining authorized functionality.
Abstract: With the rapid progress in diffusion models, image synthesis has advanced to the stage of zero-shot image-to-image generation, where high-fidelity replication of facial identities or artistic styles can be achieved using just one portrait or artwork, without modifying any model weights. Although these techniques significantly enhance creative possibilities, they also pose substantial risks related to intellectual property violations, including unauthorized identity cloning and stylistic imitation. To counter such threats, this work presents Adapter Shield, the first universal and authentication-integrated solution aimed at defending personal images from misuse in zero-shot generation scenarios. We first investigate how current zero-shot methods employ image encoders to extract embeddings from input images, which are subsequently fed into the UNet of diffusion models through cross-attention layers. Inspired by this mechanism, we construct a reversible encryption system that maps original embeddings into distinct encrypted representations according to different secret keys. The authorized users can restore the authentic embeddings via a decryption module and the correct key, enabling normal usage for authorized generation tasks. For protection purposes, we design a multi-target adversarial perturbation method that actively shifts the original embeddings toward designated encrypted patterns. Consequently, protected images are embedded with a defensive layer that ensures unauthorized users can only produce distorted or encrypted outputs. Extensive evaluations demonstrate that our method surpasses existing state-of-the-art defenses in blocking unauthorized zero-shot image synthesis, while supporting flexible and secure access control for verified users.
[141] Diffusion-Based Synthetic Brightfield Microscopy Images for Enhanced Single Cell Detection
Mario de Jesus da Graca, Jörg Dahlkemper, Peer Stelldinger
Main category: cs.CV
TL;DR: Using diffusion models to generate synthetic brightfield microscopy images improves object detection performance while reducing annotation costs, with synthetic images being indistinguishable from real ones by human experts.
Details
Motivation: Data scarcity and annotation bottlenecks in brightfield microscopy limit deep learning progress for single cell detection, creating need for synthetic data generation methods.Method: Trained U-Net based diffusion model to generate synthetic brightfield microscopy images, created datasets with varying synthetic-real ratios, tested with YOLOv8, YOLOv9 and RT-DETR object detectors, and conducted human expert survey for realism evaluation.
Result: Synthetic data improved detection accuracies at minimal costs; human experts could not distinguish synthetic from real images (50% accuracy, random chance level).
Conclusion: Diffusion-based synthetic data generation is promising for augmenting microscopy datasets, reducing manual annotation reliance, and potentially improving cell detection model robustness.
Abstract: Accurate single cell detection in brightfield microscopy is crucial for biological research, yet data scarcity and annotation bottlenecks limit the progress of deep learning methods. We investigate the use of unconditional models to generate synthetic brightfield microscopy images and evaluate their impact on object detection performance. A U-Net based diffusion model was trained and used to create datasets with varying ratios of synthetic and real images. Experiments with YOLOv8, YOLOv9 and RT-DETR reveal that training with synthetic data can achieve improved detection accuracies (at minimal costs). A human expert survey demonstrates the high realism of generated images, with experts not capable to distinguish them from real microscopy images (accuracy 50%). Our findings suggest that diffusion-based synthetic data generation is a promising avenue for augmenting real datasets in microscopy image analysis, reducing the reliance on extensive manual annotation and potentially improving the robustness of cell detection models.
[142] TinyViT: Field Deployable Transformer Pipeline for Solar Panel Surface Fault and Severity Screening
Ishwaryah Pandiarajan, Mohamed Mansoor Roomi Sindha, Uma Maheswari Pandyan, Sharafia N
Main category: cs.CV
TL;DR: TinyViT: A compact deep learning pipeline using only visible-light images to detect and grade severity of 7 types of PV panel surface faults, eliminating need for expensive multi-modal sensors.
Details
Motivation: Current solar PV fault detection relies on multi-modal imaging (electroluminescence/IR sensors) which is expensive and logistically challenging for routine farm-level deployment, creating barriers for resource-limited installations.Method: TinyViT pipeline combines Transformer-based segmentation with spectral-spatial feature engineering and ensemble regression. It uses only consumer-grade color camera images to classify 7 nuanced surface faults and generate actionable severity grades.
Result: The system achieves competitive accuracy and interpretability on real-world datasets for both classification and regression tasks, validating its effectiveness without specialized sensors.
Conclusion: By eliminating reliance on expensive multi-modal sensors, this approach enables affordable, scalable solar health monitoring with universal field accessibility for resource-limited installations.
Abstract: Sustained operation of solar photovoltaic assets hinges on accurate detection and prioritization of surface faults across vast, geographically distributed modules. While multi modal imaging strategies are popular, they introduce logistical and economic barriers for routine farm level deployment. This work demonstrates that deep learning and classical machine learning may be judiciously combined to achieve robust surface anomaly categorization and severity estimation from planar visible band imagery alone. We introduce TinyViT which is a compact pipeline integrating Transformer based segmentation, spectral-spatial feature engineering, and ensemble regression. The system ingests consumer grade color camera mosaics of PV panels, classifies seven nuanced surface faults, and generates actionable severity grades for maintenance triage. By eliminating reliance on electroluminescence or IR sensors, our method enables affordable, scalable upkeep for resource limited installations, and advances the state of solar health monitoring toward universal field accessibility. Experiments on real public world datasets validate both classification and regression sub modules, achieving accuracy and interpretability competitive with specialized approaches.
[143] Conceptual Evaluation of Deep Visual Stereo Odometry for the MARWIN Radiation Monitoring Robot in Accelerator Tunnels
André Dehne, Juri Zach, Peer Stelldinger
Main category: cs.CV
TL;DR: The paper proposes using deep visual stereo odometry (DVSO) as an alternative navigation method for the MARWIN robot at European XFEL, aiming to improve flexibility in unknown tunnel environments compared to current lidar/QR-code based systems.
Details
Motivation: Current MARWIN robot navigation at European XFEL uses lidar-based edge detection with QR-code referencing, which works well in predefined sections but lacks flexibility for unknown geometries and obstacles in long accelerator tunnels.Method: Proposes deep visual stereo odometry (DVSO) using stereo disparity, optical flow, and self-supervised learning to jointly estimate depth and ego-motion without labeled data, with potential fusion with absolute references or other sensors for global consistency.
Result: Conceptual evaluation shows expected benefits include reduced scale drift via stereo, low-cost sensing, and scalable data collection, but challenges remain in low-texture surfaces, lighting variability, computational load, and radiation robustness.
Conclusion: DVSO offers a promising alternative for autonomous navigation in constrained accelerator tunnels, defining a research agenda to enable MARWIN to navigate more autonomously in safety-critical infrastructure environments.
Abstract: The MARWIN robot operates at the European XFEL to perform autonomous radiation monitoring in long, monotonous accelerator tunnels where conventional localization approaches struggle. Its current navigation concept combines lidar-based edge detection, wheel/lidar odometry with periodic QR-code referencing, and fuzzy control of wall distance, rotation, and longitudinal position. While robust in predefined sections, this design lacks flexibility for unknown geometries and obstacles. This paper explores deep visual stereo odometry (DVSO) with 3D-geometric constraints as a focused alternative. DVSO is purely vision-based, leveraging stereo disparity, optical flow, and self-supervised learning to jointly estimate depth and ego-motion without labeled data. For global consistency, DVSO can subsequently be fused with absolute references (e.g., landmarks) or other sensors. We provide a conceptual evaluation for accelerator tunnel environments, using the European XFEL as a case study. Expected benefits include reduced scale drift via stereo, low-cost sensing, and scalable data collection, while challenges remain in low-texture surfaces, lighting variability, computational load, and robustness under radiation. The paper defines a research agenda toward enabling MARWIN to navigate more autonomously in constrained, safety-critical infrastructures.
[144] SizeGS: Size-aware Compression of 3D Gaussian Splatting via Mixed Integer Programming
Shuzhao Xie, Jiahang Liu, Weixiang Zhang, Shijia Ge, Sicheng Pan, Chen Tang, Yunpeng Bai, Cong Zhang, Xiaoyi Fan, Zhi Wang
Main category: cs.CV
TL;DR: A size-aware compression method for 3D Gaussian Splatting that optimizes hyperparameters to meet specific size constraints while maximizing visual quality.
Details
Motivation: 3D Gaussian Splatting produces large data sizes that are challenging for transmission and storage. Existing compression methods don't efficiently adapt to fluctuating network bandwidth, leading to resource wastage.Method: Formulates compression as a mixed-integer nonlinear programming (MINLP) problem, decouples it into discrete reserve ratio sampling and integer linear programming for bit-width settings, and implements quality/size estimators with CUDA kernels.
Result: Achieves state-of-the-art performance in post-training compression on multiple 3DGS variants, and achieves comparable quality to leading training-required methods after fine-tuning.
Conclusion: Proposes an effective size-aware compression framework for 3DGS that can adapt to bandwidth constraints while maintaining visual quality, bridging the gap between post-training and training-required compression methods.
Abstract: Recent advances in 3D Gaussian Splatting (3DGS) have greatly improved 3D reconstruction. However, its substantial data size poses a significant challenge for transmission and storage. While many compression techniques have been proposed, they fail to efficiently adapt to fluctuating network bandwidth, leading to resource wastage. We address this issue from the perspective of size-aware compression, where we aim to compress 3DGS to a desired size by quickly searching for suitable hyperparameters. Through a measurement study, we identify key hyperparameters that affect the size – namely, the reserve ratio of Gaussians and bit-width settings for Gaussian attributes. Then, we formulate this hyperparameter optimization problem as a mixed-integer nonlinear programming (MINLP) problem, with the goal of maximizing visual quality while respecting the size budget constraint. To solve the MINLP, we decouple this problem into two parts: discretely sampling the reserve ratio and determining the bit-width settings using integer linear programming (ILP). To solve the ILP more quickly and accurately, we design a quality loss estimator and a calibrated size estimator, as well as implement a CUDA kernel. Extensive experiments on multiple 3DGS variants demonstrate that our method achieves state-of-the-art performance in post-training compression. Furthermore, our method can achieve comparable quality to leading training-required methods after fine-tuning.
[145] Exploring Diagnostic Prompting Approach for Multimodal LLM-based Visual Complexity Assessment: A Case Study of Amazon Search Result Pages
Divendar Murtadak, Yoon Kim, Trilokya Akula
Main category: cs.CV
TL;DR: Diagnostic prompting improves MLLM reliability for visual complexity assessment of Amazon search pages, showing 858% F1-score improvement but still modest absolute performance with Cohen’s κ = 0.071.
Details
Motivation: To investigate whether diagnostic prompting can improve Multimodal Large Language Model (MLLM) reliability for visual complexity assessment of Amazon Search Results Pages, addressing the gap between MLLM performance and human expert judgments.Method: Compared diagnostic prompting with standard gestalt principles-based prompting using 200 Amazon SRP pages and human expert annotations, analyzed decision trees to understand reasoning patterns, and conducted failure case analysis.
Result: Diagnostic prompting showed notable improvements with F1-score increasing from 0.031 to 0.297 (+858% relative improvement), though absolute performance remains modest (Cohen’s κ = 0.071). Models prioritize visual design elements while humans emphasize content similarity, revealing partial alignment in reasoning patterns.
Conclusion: Diagnostic prompting represents a promising initial step toward human-aligned MLLM-based evaluation, but persistent challenges in visual perception (product similarity, color intensity) require continued research, refinement in prompting approaches, and larger ground truth datasets for reliable practical deployment.
Abstract: This study investigates whether diagnostic prompting can improve Multimodal Large Language Model (MLLM) reliability for visual complexity assessment of Amazon Search Results Pages (SRP). We compare diagnostic prompting with standard gestalt principles-based prompting using 200 Amazon SRP pages and human expert annotations. Diagnostic prompting showed notable improvements in predicting human complexity judgments, with F1-score increasing from 0.031 to 0.297 (+858% relative improvement), though absolute performance remains modest (Cohen’s $κ$ = 0.071). The decision tree revealed that models prioritize visual design elements (badge clutter: 38.6% importance) while humans emphasize content similarity, suggesting partial alignment in reasoning patterns. Failure case analysis reveals persistent challenges in MLLM visual perception, particularly for product similarity and color intensity assessment. Our findings indicate that diagnostic prompting represents a promising initial step toward human-aligned MLLM-based evaluation, though failure cases with consistent human-MLLM disagreement require continued research and refinement in prompting approaches with larger ground truth datasets for reliable practical deployment.
[146] Prompt-OT: An Optimal Transport Regularization Paradigm for Knowledge Preservation in Vision-Language Model Adaptation
Xiwen Chen, Wenhui Zhu, Peijie Qiu, Hao Wang, Huayu Li, Haiyu Wu, Aristeidis Sotiras, Yalin Wang, Abolfazl Razi
Main category: cs.CV
TL;DR: OT-guided prompt learning framework for VLMs that uses optimal transport to preserve structural consistency between pre-trained and fine-tuned feature distributions, improving adaptation-generalization trade-off.
Details
Motivation: Existing prompt learning methods for vision-language models lead to overfitting and degrade zero-shot generalization, creating a need for better adaptation strategies that preserve pre-trained knowledge while adapting to downstream tasks.Method: Proposes an optimal transport-guided prompt learning framework that enforces joint constraints on both vision and text representations to preserve structural consistency of feature distributions between pre-trained and fine-tuned models, capturing cross-instance relationships and expanding feasible parameter space.
Result: Outperforms existing prompt learning strategies in base-to-novel generalization, cross-dataset evaluation, and domain generalization without additional augmentation or ensemble techniques, as demonstrated in extensive experiments on benchmark datasets.
Conclusion: The OT-guided framework provides a simple yet effective solution to mitigate forgetting in prompt learning for VLMs, achieving better trade-off between adaptation and generalization through holistic feature alignment.
Abstract: Vision-language models (VLMs) such as CLIP demonstrate strong performance but struggle when adapted to downstream tasks. Prompt learning has emerged as an efficient and effective strategy to adapt VLMs while preserving their pre-trained knowledge. However, existing methods still lead to overfitting and degrade zero-shot generalization. To address this challenge, we propose an optimal transport (OT)-guided prompt learning framework that mitigates forgetting by preserving the structural consistency of feature distributions between pre-trained and fine-tuned models. Unlike conventional point-wise constraints, OT naturally captures cross-instance relationships and expands the feasible parameter space for prompt tuning, allowing a better trade-off between adaptation and generalization. Our approach enforces joint constraints on both vision and text representations, ensuring a holistic feature alignment. Extensive experiments on benchmark datasets demonstrate that our simple yet effective method can outperform existing prompt learning strategies in base-to-novel generalization, cross-dataset evaluation, and domain generalization without additional augmentation or ensemble techniques. The code is available at https://github.com/ChongQingNoSubway/Prompt-OT
[147] Efficient Edge-Compatible CNN for Speckle-Based Material Recognition in Laser Cutting Systems
Mohamed Abdallah Salem, Nourhan Zein Diab
Main category: cs.CV
TL;DR: A lightweight CNN for laser speckle material recognition achieves 95% accuracy on 59 material classes with only 341k parameters, enabling edge deployment on devices like Raspberry Pi for safer laser cutting.
Details
Motivation: Accurate material recognition is critical for safe laser cutting to prevent poor cut quality, machine damage, and hazardous fumes. Existing speckle-based methods use computationally expensive networks or handle limited material subsets.Method: Proposes a lightweight convolutional neural network specifically designed for speckle patterns, minimizing parameters while maintaining discriminative power. Uses the complete SensiCut dataset of 59 material classes across various categories.
Result: Achieves 95.05% test accuracy with 0.951 F1-scores, using only 341k parameters (70x fewer than ResNet-50). Inference speed of 295 images/sec enables Raspberry Pi/Jetson deployment. Material family regrouping achieves 98-100% recall for practical laser cutter settings.
Conclusion: Compact, domain-specific CNNs outperform large backbone networks for speckle-based material classification, advancing material-aware, edge-deployable laser cutting systems with practical applications in power/speed preset selection.
Abstract: Accurate material recognition is critical for safe and effective laser cutting, as misidentification can lead to poor cut quality, machine damage, or the release of hazardous fumes. Laser speckle sensing has recently emerged as a low-cost and non-destructive modality for material classification; however, prior work has either relied on computationally expensive backbone networks or addressed only limited subsets of materials. In this study, A lightweight convolutional neural network (CNN) tailored for speckle patterns is proposed, designed to minimize parameters while maintaining high discriminative power. Using the complete SensiCut dataset of 59 material classes spanning woods, acrylics, composites, textiles, metals, and paper-based products, the proposed model achieves 95.05% test accuracy, with macro and weighted F1-scores of 0.951. The network contains only 341k trainable parameters (~1.3 MB) – over 70X fewer than ResNet-50 – and achieves an inference speed of 295 images per second, enabling deployment on Raspberry Pi and Jetson-class devices. Furthermore, when materials are regrouped into nine and five practical families, recall exceeds 98% and approaches 100%, directly supporting power and speed preset selection in laser cutters. These results demonstrate that compact, domain-specific CNNs can outperform large backbones for speckle-based material classification, advancing the feasibility of material-aware, edge-deployable laser cutting systems.
[148] A Fast and Efficient Modern BERT based Text-Conditioned Diffusion Model for Medical Image Segmentation
Venkata Siddharth Dhara, Pawan Kumar
Main category: cs.CV
TL;DR: FastTextDiff is a label-efficient diffusion-based medical image segmentation model that integrates medical text annotations to enhance semantic representations, using ModernBERT for processing clinical notes instead of Clinical BioBERT.
Details
Motivation: Traditional diffusion-based segmentation models require dense pixel-wise labels that are expensive, time-consuming, and require expert knowledge. There's a need for more label-efficient approaches that can leverage available medical text annotations.Method: Proposes FastTextDiff, which integrates medical text annotations with diffusion models using ModernBERT (a transformer for long clinical notes) to link textual annotations with semantic image content. Uses cross-modal attention between visual and textual features, trained on MIMIC-III and MIMIC-IV datasets.
Result: ModernBERT serves as a fast, scalable alternative to Clinical BioBERT in diffusion-based segmentation pipelines. FastTextDiff benefits from FlashAttention 2, alternating attention mechanism, and a 2-trillion-token corpus, improving both segmentation accuracy and training efficiency over traditional diffusion-based models.
Conclusion: The study validates ModernBERT as an effective alternative to Clinical BioBERT and highlights the promise of multi-modal techniques for medical image analysis, demonstrating improved segmentation performance with reduced labeling requirements.
Abstract: In recent times, denoising diffusion probabilistic models (DPMs) have proven effective for medical image generation and denoising, and as representation learners for downstream segmentation. However, segmentation performance is limited by the need for dense pixel-wise labels, which are expensive, time-consuming, and require expert knowledge. We propose FastTextDiff, a label-efficient diffusion-based segmentation model that integrates medical text annotations to enhance semantic representations. Our approach uses ModernBERT, a transformer capable of processing long clinical notes, to tightly link textual annotations with semantic content in medical images. Trained on MIMIC-III and MIMIC-IV, ModernBERT encodes clinical knowledge that guides cross-modal attention between visual and textual features. This study validates ModernBERT as a fast, scalable alternative to Clinical BioBERT in diffusion-based segmentation pipelines and highlights the promise of multi-modal techniques for medical image analysis. By replacing Clinical BioBERT with ModernBERT, FastTextDiff benefits from FlashAttention 2, an alternating attention mechanism, and a 2-trillion-token corpus, improving both segmentation accuracy and training efficiency over traditional diffusion-based models.
[149] Proxy-Tuning: Tailoring Multimodal Autoregressive Models for Subject-Driven Image Generation
Yi Wu, Shengju Qian, Lingting Zhu, Lei Liu, Wandi Qiao, Ziqiang Li, Lequan Yu, Bin Li
Main category: cs.CV
TL;DR: Proxy-Tuning enhances multimodal autoregressive models for subject-driven image generation by leveraging diffusion models, revealing a weak-to-strong phenomenon where fine-tuned AR models outperform their diffusion model supervisors.
Details
Motivation: Multimodal autoregressive models based on next-token prediction show strong performance in general text-to-image tasks but initially struggle with subject-driven image generation compared to diffusion models. The authors aim to address this limitation and explore the potential of AR models for subject-specific image generation.Method: Introduces Proxy-Tuning, which leverages diffusion models to enhance AR models’ capabilities in subject-specific image generation. The method involves using diffusion models as supervisors to fine-tune AR models, enabling them to learn subject-driven generation capabilities.
Result: Reveals a striking weak-to-strong phenomenon: fine-tuned AR models consistently outperform their diffusion model supervisors in both subject fidelity and prompt adherence. AR models excel particularly in multi-subject compositions and contextual understanding.
Conclusion: The work demonstrates impressive results in subject-driven AR image generation and unveils the potential of weak-to-strong generalization in the image generation domain. It contributes to a deeper understanding of different architectures’ strengths and limitations, showing that AR models can surpass diffusion models in specific subject-driven tasks after proper tuning.
Abstract: Multimodal autoregressive (AR) models, based on next-token prediction and transformer architecture, have demonstrated remarkable capabilities in various multimodal tasks including text-to-image (T2I) generation. Despite their strong performance in general T2I tasks, our research reveals that these models initially struggle with subject-driven image generation compared to dominant diffusion models. To address this limitation, we introduce Proxy-Tuning, leveraging diffusion models to enhance AR models’ capabilities in subject-specific image generation. Our method reveals a striking weak-to-strong phenomenon: fine-tuned AR models consistently outperform their diffusion model supervisors in both subject fidelity and prompt adherence. We analyze this performance shift and identify scenarios where AR models excel, particularly in multi-subject compositions and contextual understanding. This work not only demonstrates impressive results in subject-driven AR image generation, but also unveils the potential of weak-to-strong generalization in the image generation domain, contributing to a deeper understanding of different architectures’ strengths and limitations.
[150] AutocleanEEG ICVision: Automated ICA Artifact Classification Using Vision-Language AI
Zag ElSayed, Grace Westerkamp, Gavin Gammoh, Yanchen Liu, Peyton Siekierski, Craig Erickson, Ernest Pedapati
Main category: cs.CV
TL;DR: ICVision is an AI system that uses vision-language models to classify EEG ICA components by visually interpreting dashboard plots like experts do, achieving better performance than traditional classifiers while providing human-like explanations.
Details
Motivation: Current EEG ICA classification methods rely on handcrafted features and lack the visual reasoning capabilities that human experts use when examining EEG component visualizations. There's a need for AI systems that can emulate expert-level visual cognition in neurophysiology.Method: ICVision uses GPT-4 Vision to directly interpret ICA dashboard visualizations (topography, time series, power spectra, and ERP plots) as multimodal inputs. It classifies components into six canonical categories and provides both confidence scores and natural language explanations.
Result: Evaluated on 3,168 ICA components from 124 EEG datasets, ICVision achieved k = 0.677 agreement with expert consensus, surpassing MNE ICLabel. Over 97% of outputs were rated as interpretable and actionable by experts, and it preserved clinically relevant brain signals in ambiguous cases.
Conclusion: ICVision represents a paradigm shift toward AI agents capable of expert-level visual decision-making in brain science. It enables scalable, explainable, and reproducible EEG workflows and opens doors for AI visual cognition applications beyond neurophysiology.
Abstract: We introduce EEG Autoclean Vision Language AI (ICVision) a first-of-its-kind system that emulates expert-level EEG ICA component classification through AI-agent vision and natural language reasoning. Unlike conventional classifiers such as ICLabel, which rely on handcrafted features, ICVision directly interprets ICA dashboard visualizations topography, time series, power spectra, and ERP plots, using a multimodal large language model (GPT-4 Vision). This allows the AI to see and explain EEG components the way trained neurologists do, making it the first scientific implementation of AI-agent visual cognition in neurophysiology. ICVision classifies each component into one of six canonical categories (brain, eye, heart, muscle, channel noise, and other noise), returning both a confidence score and a human-like explanation. Evaluated on 3,168 ICA components from 124 EEG datasets, ICVision achieved k = 0.677 agreement with expert consensus, surpassing MNE ICLabel, while also preserving clinically relevant brain signals in ambiguous cases. Over 97% of its outputs were rated as interpretable and actionable by expert reviewers. As a core module of the open-source EEG Autoclean platform, ICVision signals a paradigm shift in scientific AI, where models do not just classify, but see, reason, and communicate. It opens the door to globally scalable, explainable, and reproducible EEG workflows, marking the emergence of AI agents capable of expert-level visual decision-making in brain science and beyond.
[151] Multi-modal On-Device Learning for Monocular Depth Estimation on Ultra-low-power MCUs
Davide Nadalini, Manuele Rusci, Elia Cereda, Luca Benini, Francesco Conti, Daniele Palossi
Main category: cs.CV
TL;DR: On-device learning for monocular depth estimation on ultra-low-power IoT devices using sparse updates and multi-modal pseudo-labeling.
Details
Motivation: Monocular depth estimation on ULP IoT platforms suffers from domain shift when sensor data differs from training data, requiring adaptation to new environments without sacrificing energy efficiency.Method: Multi-modal on-device learning with tiny μPyD-Net model (107k params), using monocular camera for inference and activating low-power depth sensor (8x8 pixels) only for collecting pseudo-labels in new environments. Introduces memory-driven sparse update scheme to minimize fine-tuning memory to 1.2MB.
Result: Reduces RMSE from 4.9m to 0.6m with only 3k self-labeled samples in real deployment. Fine-tuning takes 17.8 minutes on IoT node with only 2% and 1.5% accuracy drops on KITTI and NYUv2 datasets compared to full update.
Conclusion: Demonstrates practical on-device learning for monocular depth estimation on ULP IoT platforms, enabling adaptation to new environments with minimal energy consumption and memory overhead.
Abstract: Monocular depth estimation (MDE) plays a crucial role in enabling spatially-aware applications in Ultra-low-power (ULP) Internet-of-Things (IoT) platforms. However, the limited number of parameters of Deep Neural Networks for the MDE task, designed for IoT nodes, results in severe accuracy drops when the sensor data observed in the field shifts significantly from the training dataset. To address this domain shift problem, we present a multi-modal On-Device Learning (ODL) technique, deployed on an IoT device integrating a Greenwaves GAP9 MicroController Unit (MCU), a 80 mW monocular camera and a 8 x 8 pixel depth sensor, consuming $\approx$300mW. In its normal operation, this setup feeds a tiny 107 k-parameter $μ$PyD-Net model with monocular images for inference. The depth sensor, usually deactivated to minimize energy consumption, is only activated alongside the camera to collect pseudo-labels when the system is placed in a new environment. Then, the fine-tuning task is performed entirely on the MCU, using the new data. To optimize our backpropagation-based on-device training, we introduce a novel memory-driven sparse update scheme, which minimizes the fine-tuning memory to 1.2 MB, 2.2x less than a full update, while preserving accuracy (i.e., only 2% and 1.5% drops on the KITTI and NYUv2 datasets). Our in-field tests demonstrate, for the first time, that ODL for MDE can be performed in 17.8 minutes on the IoT node, reducing the root mean squared error from 4.9 to 0.6m with only 3 k self-labeled samples, collected in a real-life deployment scenario.
[152] VIVAT: Virtuous Improving VAE Training through Artifact Mitigation
Lev Novitskiy, Viacheslav Vasilev, Maria Kovaleva, Vladimir Arkhipkin, Denis Dimitrov
Main category: cs.CV
TL;DR: VIVAT addresses common artifacts in KL-VAE training through systematic modifications, improving reconstruction and generation quality without major architectural changes.
Details
Motivation: VAEs suffer from persistent artifacts that degrade reconstruction and generation quality, but current solutions often require complex architectural changes. The paper aims to provide a systematic approach to mitigate these artifacts while preserving the simplicity of KL-VAE framework.Method: Introduces VIVAT with a taxonomy of five common artifacts (color shift, grid patterns, blur, corner and droplet artifacts), analyzes their root causes, and implements straightforward modifications including loss weight adjustments, padding strategies, and integration of Spatially Conditional Normalization.
Result: Achieves state-of-the-art results in image reconstruction metrics (PSNR and SSIM) across multiple benchmarks and improves text-to-image generation quality with superior CLIP scores.
Conclusion: VIVAT provides an effective, practical solution for mitigating VAE artifacts without radical architectural changes, offering actionable insights for researchers and practitioners to optimize VAE training while maintaining framework simplicity.
Abstract: Variational Autoencoders (VAEs) remain a cornerstone of generative computer vision, yet their training is often plagued by artifacts that degrade reconstruction and generation quality. This paper introduces VIVAT, a systematic approach to mitigating common artifacts in KL-VAE training without requiring radical architectural changes. We present a detailed taxonomy of five prevalent artifacts - color shift, grid patterns, blur, corner and droplet artifacts - and analyze their root causes. Through straightforward modifications, including adjustments to loss weights, padding strategies, and the integration of Spatially Conditional Normalization, we demonstrate significant improvements in VAE performance. Our method achieves state-of-the-art results in image reconstruction metrics (PSNR and SSIM) across multiple benchmarks and enhances text-to-image generation quality, as evidenced by superior CLIP scores. By preserving the simplicity of the KL-VAE framework while addressing its practical challenges, VIVAT offers actionable insights for researchers and practitioners aiming to optimize VAE training.
[153] Exploring Automated Recognition of Instructional Activity and Discourse from Multimodal Classroom Data
Ivo Bueno, Ruikun Hou, Babette Bühler, Tim Fütterer, James Drimalla, Jonathan Kyle Foster, Peter Youngs, Peter Gerjets, Ulrich Trautwein, Enkelejda Kasneci
Main category: cs.CV
TL;DR: AI-driven multimodal analysis of classroom recordings using video and transcript data to automate instructional activity and discourse recognition for scalable teacher feedback.
Details
Motivation: Manual classroom observation is resource-intensive and hard to scale, creating a need for automated AI-driven analysis to provide concrete feedback to teachers.Method: Parallel modality-specific pipelines: for video, evaluate zero-shot multimodal LLMs, fine-tuned vision-language models, and self-supervised video transformers on 24 activity labels; for transcripts, fine-tune transformer-based classifier with contextualized inputs and compare against prompting-based LLMs on 19 discourse labels. Use per-label thresholding, context windows, and imbalance-aware loss functions to handle class imbalance and multi-label complexity.
Result: Fine-tuned models consistently outperform prompting-based approaches, achieving macro-F1 scores of 0.577 for video analysis and 0.460 for transcript analysis.
Conclusion: The results demonstrate feasibility of automated classroom analysis and establish a foundation for scalable teacher feedback systems, showing that fine-tuned models are more effective than prompting-based approaches for this task.
Abstract: Observation of classroom interactions can provide concrete feedback to teachers, but current methods rely on manual annotation, which is resource-intensive and hard to scale. This work explores AI-driven analysis of classroom recordings, focusing on multimodal instructional activity and discourse recognition as a foundation for actionable feedback. Using a densely annotated dataset of 164 hours of video and 68 lesson transcripts, we design parallel, modality-specific pipelines. For video, we evaluate zero-shot multimodal LLMs, fine-tuned vision-language models, and self-supervised video transformers on 24 activity labels. For transcripts, we fine-tune a transformer-based classifier with contextualized inputs and compare it against prompting-based LLMs on 19 discourse labels. To handle class imbalance and multi-label complexity, we apply per-label thresholding, context windows, and imbalance-aware loss functions. The results show that fine-tuned models consistently outperform prompting-based approaches, achieving macro-F1 scores of 0.577 for video and 0.460 for transcripts. These results demonstrate the feasibility of automated classroom analysis and establish a foundation for scalable teacher feedback systems.
[154] Fine-grained Image Retrieval via Dual-Vision Adaptation
Xin Jiang, Meiqi Cao, Hao Tang, Fei Shen, Zechao Li
Main category: cs.CV
TL;DR: DVA: A dual-adaptation approach for fine-grained image retrieval that modifies samples and features to leverage pre-trained knowledge without overfitting, achieving strong performance with few parameters.
Details
Motivation: Current FGIR methods either use pairwise similarity constraints or localization networks, which tend to overfit training data and forget pre-trained knowledge, reducing generalization ability.Method: Proposes Dual-Vision Adaptation (DVA) with: 1) Object-Perceptual Adaptation that modifies input samples to help pre-trained models perceive critical objects; 2) In-Context Adaptation that introduces small parameter sets for feature adaptation; 3) Discrimination Perception Transfer to balance efficiency and performance via knowledge distillation.
Result: DVA has fewer learnable parameters and performs well on three in-distribution and three out-of-distribution fine-grained datasets.
Conclusion: DVA effectively guides frozen pre-trained models for FGIR through collaborative sample and feature adaptation, maintaining pre-trained knowledge while achieving strong generalization.
Abstract: Fine-Grained Image Retrieval~(FGIR) faces challenges in learning discriminative visual representations to retrieve images with similar fine-grained features. Current leading FGIR solutions typically follow two regimes: enforce pairwise similarity constraints in the semantic embedding space, or incorporate a localization sub-network to fine-tune the entire model. However, such two regimes tend to overfit the training data while forgetting the knowledge gained from large-scale pre-training, thus reducing their generalization ability. In this paper, we propose a Dual-Vision Adaptation (DVA) approach for FGIR, which guides the frozen pre-trained model to perform FGIR through collaborative sample and feature adaptation. Specifically, we design Object-Perceptual Adaptation, which modifies input samples to help the pre-trained model perceive critical objects and elements within objects that are helpful for category prediction. Meanwhile, we propose In-Context Adaptation, which introduces a small set of parameters for feature adaptation without modifying the pre-trained parameters. This makes the FGIR task using these adjusted features closer to the task solved during the pre-training. Additionally, to balance retrieval efficiency and performance, we propose Discrimination Perception Transfer to transfer the discriminative knowledge in the object-perceptual adaptation to the image encoder using the knowledge distillation mechanism. Extensive experiments show that DVA has fewer learnable parameters and performs well on three in-distribution and three out-of-distribution fine-grained datasets.
[155] SemImage: Semantic Image Representation for Text, a Novel Framework for Embedding Disentangled Linguistic Features
Mohammad Zare
Main category: cs.CV
TL;DR: SemImage converts text documents into 2D semantic images where words become pixels in HSV color space, with Hue encoding topic, Saturation encoding sentiment, and Value encoding intensity, processed by CNNs for document classification.
Details
Motivation: To create a visual representation of text documents that makes semantic features (topic, sentiment, intensity) visible and interpretable while leveraging the power of convolutional neural networks for document classification.Method: Represent each word as a pixel in a 2D image with rows as sentences and boundary rows between sentences. Map word embeddings to HSV color space where Hue encodes topic (with circular components), Saturation encodes sentiment, and Value encodes intensity. Use multi-task learning with auxiliary supervision on Hue and Saturation channels. Insert dynamic boundary rows to highlight semantic transitions.
Result: Achieves competitive or better accuracy than strong baselines (including BERT and hierarchical attention networks) on multi-label and single-label document classification benchmarks. Ablation studies confirm importance of multi-channel HSV representation and dynamic boundary rows.
Conclusion: SemImage provides an effective visual representation for text documents that enables competitive classification performance while offering enhanced interpretability through visible patterns of topic shifts and sentiment changes in the generated images.
Abstract: We propose SemImage, a novel method for representing a text document as a two-dimensional semantic image to be processed by convolutional neural networks (CNNs). In a SemImage, each word is represented as a pixel in a 2D image: rows correspond to sentences and an additional boundary row is inserted between sentences to mark semantic transitions. Each pixel is not a typical RGB value but a vector in a disentangled HSV color space, encoding different linguistic features: the Hue with two components H_cos and H_sin to account for circularity encodes the topic, Saturation encodes the sentiment, and Value encodes intensity or certainty. We enforce this disentanglement via a multi-task learning framework: a ColorMapper network maps each word embedding to the HSV space, and auxiliary supervision is applied to the Hue and Saturation channels to predict topic and sentiment labels, alongside the main task objective. The insertion of dynamically computed boundary rows between sentences yields sharp visual boundaries in the image when consecutive sentences are semantically dissimilar, effectively making paragraph breaks salient. We integrate SemImage with standard 2D CNNs (e.g., ResNet) for document classification. Experiments on multi-label datasets (with both topic and sentiment annotations) and single-label benchmarks demonstrate that SemImage can achieve competitive or better accuracy than strong text classification baselines (including BERT and hierarchical attention networks) while offering enhanced interpretability. An ablation study confirms the importance of the multi-channel HSV representation and the dynamic boundary rows. Finally, we present visualizations of SemImage that qualitatively reveal clear patterns corresponding to topic shifts and sentiment changes in the generated image, suggesting that our representation makes these linguistic features visible to both humans and machines.
[156] DMC$^3$: Dual-Modal Counterfactual Contrastive Construction for Egocentric Video Question Answering
Jiayi Zou, Chaofan Chen, Bing-Kun Bao, Changsheng Xu
Main category: cs.CV
TL;DR: DMC³ framework improves Egocentric VideoQA by generating counterfactual samples for contrastive learning to address first-person perspective challenges like multiple events and hand-object interactions.
Details
Motivation: Existing Egocentric VideoQA methods ignore unique first-person perspective challenges such as understanding multiple events and recognizing hand-object interactions, limiting their effectiveness.Method: Dual-Modal Counterfactual Contrastive Construction (DMC³) framework with: 1) counterfactual sample construction module generating positive/negative samples via event description paraphrasing (text) and core interaction mining (visual), 2) feeding these with original samples into baseline, 3) contrastive optimization minimizing distance to positive samples and maximizing distance to negative samples.
Result: Achieves 52.51% on normal split and 46.04% on indirect split of EgoTaskQA, and 13.2% on QAEGO4D, reaching state-of-the-art performance.
Conclusion: The DMC³ framework effectively addresses first-person perspective challenges in Egocentric VideoQA through counterfactual contrastive learning, demonstrating superior performance on benchmark datasets.
Abstract: Egocentric Video Question Answering (Egocentric VideoQA) plays an important role in egocentric video understanding, which refers to answering questions based on first-person videos. Although existing methods have made progress through the paradigm of pre-training and fine-tuning, they ignore the unique challenges posed by the first-person perspective, such as understanding multiple events and recognizing hand-object interactions. To deal with these challenges, we propose a Dual-Modal Counterfactual Contrastive Construction (DMC$^3$) framework, which contains an egocentric videoqa baseline, a counterfactual sample construction module and a counterfactual sample-involved contrastive optimization. Specifically, We first develop a counterfactual sample construction module to generate positive and negative samples for textual and visual modalities through event description paraphrasing and core interaction mining, respectively. Then, We feed these samples together with the original samples into the baseline. Finally, in the counterfactual sample-involved contrastive optimization module, we apply contrastive loss to minimize the distance between the original sample features and the positive sample features, while maximizing the distance from the negative samples. Experiments show that our method achieve 52.51% and 46.04% on the \textit{normal} and \textit{indirect} splits of EgoTaskQA, and 13.2% on QAEGO4D, both reaching the state-of-the-art performance.
[157] CourtMotion: Learning Event-Driven Motion Representations from Skeletal Data for Basketball
Omer Sela, Michael Chertok, Lior Wolf
Main category: cs.CV
TL;DR: CourtMotion is a spatiotemporal framework using skeletal tracking data with Graph Neural Networks and Transformers to predict basketball events by connecting physical motion patterns to tactical purposes.
Details
Motivation: Traditional position-only approaches fail to capture crucial motion indicators like body orientation, defensive stance, and shooting preparation motions needed to anticipate basketball events.Method: Two-stage approach: 1) Graph Neural Networks process skeletal tracking data to capture nuanced motion patterns, 2) Transformer with specialized attention models player interactions. Event projection heads explicitly connect movements to basketball events.
Result: 35% reduction in trajectory prediction error vs position-only baselines; consistent performance gains across key basketball analytics tasks including pick detection, shot taker identification, assist prediction, shot location classification, and shot type recognition.
Conclusion: CourtMotion demonstrates that incorporating skeletal motion data with semantic understanding significantly improves basketball event prediction, providing a powerful pretrained foundation for multiple downstream analytics tasks.
Abstract: This paper presents CourtMotion, a spatiotemporal modeling framework for analyzing and predicting game events and plays as they develop in professional basketball. Anticipating basketball events requires understanding both physical motion patterns and their semantic significance in the context of the game. Traditional approaches that use only player positions fail to capture crucial indicators such as body orientation, defensive stance, or shooting preparation motions. Our two-stage approach first processes skeletal tracking data through Graph Neural Networks to capture nuanced motion patterns, then employs a Transformer architecture with specialized attention mechanisms to model player interactions. We introduce event projection heads that explicitly connect player movements to basketball events like passes, shots, and steals, training the model to associate physical motion patterns with their tactical purposes. Experiments on NBA tracking data demonstrate significant improvements over position-only baselines: 35% reduction in trajectory prediction error compared to state-of-the-art position-based models and consistent performance gains across key basketball analytics tasks. The resulting pretrained model serves as a powerful foundation for multiple downstream tasks, with pick detection, shot taker identification, assist prediction, shot location classification, and shot type recognition demonstrating substantial improvements over existing methods.
[158] TeleViT1.0: Teleconnection-aware Vision Transformers for Subseasonal to Seasonal Wildfire Pattern Forecasts
Ioannis Prapas, Nikolaos Papadopoulos, Nikolaos-Ioannis Bountos, Dimitrios Michail, Gustau Camps-Valls, Ioannis Papoutsis
Main category: cs.CV
TL;DR: TeleViT: A Vision Transformer model that integrates local fire drivers, global climate fields, and teleconnection indices for wildfire forecasting weeks to months in advance, outperforming existing methods across all lead times.
Details
Motivation: Long-term wildfire forecasting is challenging but essential for resource planning. While short-term predictions rely on local weather, long-term forecasting requires accounting for global interconnected climate patterns and teleconnections that influence fire conditions weeks to months in advance.Method: TeleViT uses an asymmetric tokenization strategy to create heterogeneous tokens from (i) fine-scale local fire drivers, (ii) coarsened global fields, and (iii) teleconnection indices. These tokens are processed jointly by a transformer encoder, with a decoder that preserves spatial structure by mapping local tokens to prediction patches.
Result: TeleViT outperforms U-Net++, ViT, and climatology across all lead times (up to 4 months) on the global SeasFire dataset (2001-2021, 8-day resolution). At zero lead: AUPRC 0.630 vs ViT 0.617/U-Net 0.620. At 4-month lead: TeleViT maintains 0.601-0.603 vs ViT 0.582/U-Net 0.578. Best performance in African savannas, lower in boreal/arid regions.
Conclusion: Architectures that explicitly encode large-scale Earth-system context (global fields and teleconnections) can extend wildfire predictability on subseasonal-to-seasonal timescales. Predictions mainly rely on local tokens, with global fields and indices providing coarse contextual information.
Abstract: Forecasting wildfires weeks to months in advance is difficult, yet crucial for planning fuel treatments and allocating resources. While short-term predictions typically rely on local weather conditions, long-term forecasting requires accounting for the Earth’s interconnectedness, including global patterns and teleconnections. We introduce TeleViT, a Teleconnection-aware Vision Transformer that integrates (i) fine-scale local fire drivers, (ii) coarsened global fields, and (iii) teleconnection indices. This multi-scale fusion is achieved through an asymmetric tokenization strategy that produces heterogeneous tokens processed jointly by a transformer encoder, followed by a decoder that preserves spatial structure by mapping local tokens to their corresponding prediction patches. Using the global SeasFire dataset (2001-2021, 8-day resolution), TeleViT improves AUPRC performance over U-Net++, ViT, and climatology across all lead times, including horizons up to four months. At zero lead, TeleViT with indices and global inputs reaches AUPRC 0.630 (ViT 0.617, U-Net 0.620), at 16x8day lead (around 4 months), TeleViT variants using global input maintain 0.601-0.603 (ViT 0.582, U-Net 0.578), while surpassing the climatology (0.572) at all lead times. Regional results show the highest skill in seasonally consistent fire regimes, such as African savannas, and lower skill in boreal and arid regions. Attention and attribution analyses indicate that predictions rely mainly on local tokens, with global fields and indices contributing coarse contextual information. These findings suggest that architectures explicitly encoding large-scale Earth-system context can extend wildfire predictability on subseasonal-to-seasonal timescales.
[159] Cross Modal Fine-Grained Alignment via Granularity-Aware and Region-Uncertain Modeling
Jiale Liu, Haoming Zhou, Yishu Liu, Bingzhi Chen, Yuncheng Jiang
Main category: cs.CV
TL;DR: Proposes a unified approach for fine-grained image-text alignment using significance-aware modeling and region-level uncertainty modeling to address limitations in existing methods.
Details
Motivation: Existing fine-grained image-text alignment methods suffer from noisy attention mechanisms and oversimplified cross-modal relationships, lacking robust intra-modal mechanisms to assess token significance and fine-grained uncertainty modeling for complex region-word correspondences.Method: Incorporates significance-aware and granularity-aware modeling using modality-specific biases to identify salient features without brittle cross-modal attention, and represents region features as a mixture of Gaussian distributions to capture fine-grained uncertainty.
Result: Achieves state-of-the-art performance on Flickr30K and MS-COCO datasets across various backbone architectures, significantly enhancing robustness and interpretability of fine-grained image-text alignment.
Conclusion: The proposed unified approach effectively addresses fundamental limitations in fine-grained image-text alignment through significance-aware modeling and uncertainty modeling, demonstrating superior performance and improved interpretability for multimodal applications.
Abstract: Fine-grained image-text alignment is a pivotal challenge in multimodal learning, underpinning key applications such as visual question answering, image captioning, and vision-language navigation. Unlike global alignment, fine-grained alignment requires precise correspondence between localized visual regions and textual tokens, often hindered by noisy attention mechanisms and oversimplified modeling of cross-modal relationships. In this work, we identify two fundamental limitations of existing approaches: the lack of robust intra-modal mechanisms to assess the significance of visual and textual tokens, leading to poor generalization in complex scenes; and the absence of fine-grained uncertainty modeling, which fails to capture the one-to-many and many-to-one nature of region-word correspondences. To address these issues, we propose a unified approach that incorporates significance-aware and granularity-aware modeling and region-level uncertainty modeling. Our method leverages modality-specific biases to identify salient features without relying on brittle cross-modal attention, and represents region features as a mixture of Gaussian distributions to capture fine-grained uncertainty. Extensive experiments on Flickr30K and MS-COCO demonstrate that our approach achieves state-of-the-art performance across various backbone architectures, significantly enhancing the robustness and interpretability of fine-grained image-text alignment.
[160] Deep Filament Extraction for 3D Concrete Printing
Karam Mawas, Mehdi Maboudi, Pedro Achanccaray, Markus Gerke
Main category: cs.CV
TL;DR: Automated quality control method for 3D concrete printing filaments that works with both extrusion-based and shotcrete techniques, using various sensors and applicable to fresh or cured materials.
Details
Motivation: The AEC industry needs sustainable construction methods, and 3D concrete printing requires precise quality control of filaments/layers since they define the printed structure's geometry.Method: Developed an automated QC procedure that works with multiple deposition methods (extrusion-based and SC3DP), is sensor-agnostic (camera, SLS, TLS), and handles both fresh and cured materials.
Result: Created a workflow for filament geometry quality control that supports both online (fresh material) and post-printing (cured material) QC, independent of specific data acquisition sensors.
Conclusion: The automated QC method enables reliable quality assessment for 3D concrete printing filaments across different printing techniques and material states, supporting both real-time and post-construction quality control.
Abstract: The architecture, engineering and construction (AEC) industry is constantly evolving to meet the demand for sustainable and effective design and construction of the built environment. In the literature, two primary deposition techniques for large-scale 3D concrete printing (3DCP) have been described, namely extrusion-based (Contour Crafting-CC) and shotcrete 3D printing (SC3DP) methods. The deposition methods use a digitally controlled nozzle to print material layer by layer. The continuous flow of concrete material used to create the printed structure is called a filament or layer. As these filaments are the essential structure defining the printed object, the filaments’ geometry quality control is crucial. This paper presents an automated procedure for quality control (QC) of filaments in extrusion-based and SC3DP printing methods. The paper also describes a workflow that is independent of the sensor used for data acquisition, such as a camera, a structured light system (SLS) or a terrestrial laser scanner (TLS). This method can be used with materials in either the fresh or cured state. Thus, it can be used for online and post-printing QC.
[161] Pistachio: Towards Synthetic, Balanced, and Long-Form Video Anomaly Benchmarks
Jie Li, Hongyi Cai, Mingkang Dong, Muxin Pu, Shan You, Fei Wang, Tao Huang
Main category: cs.CV
TL;DR: Pistachio is a new Video Anomaly Detection/Understanding benchmark created using AI video generation to overcome limitations of existing datasets, offering controlled scenes, diverse anomalies, and complex temporal narratives.
Details
Motivation: Existing VAD benchmarks lack scene diversity, balanced anomaly coverage, and temporal complexity needed for real-world assessment. The shift toward Video Anomaly Understanding requires deeper reasoning but is hard to benchmark due to heavy manual annotation requirements.Method: Uses a controlled generation-based pipeline leveraging recent video generation models. Integrates scene-conditioned anomaly assignment, multi-step storyline generation, and temporally consistent long-form synthesis to produce coherent 41-second videos with minimal human intervention.
Result: Creates a benchmark with scale, diversity, and complexity that reveals new challenges for existing methods. Demonstrates the ability to generate controlled, bias-free datasets that better assess real-world performance.
Conclusion: Pistachio provides a superior benchmark for VAD/VAU research by overcoming limitations of traditional datasets through AI-generated content, enabling more reliable assessment and motivating future work on dynamic, multi-event anomaly understanding.
Abstract: Automatically detecting abnormal events in videos is crucial for modern autonomous systems, yet existing Video Anomaly Detection (VAD) benchmarks lack the scene diversity, balanced anomaly coverage, and temporal complexity needed to reliably assess real-world performance. Meanwhile, the community is increasingly moving toward Video Anomaly Understanding (VAU), which requires deeper semantic and causal reasoning but remains difficult to benchmark due to the heavy manual annotation effort it demands. In this paper, we introduce Pistachio, a new VAD/VAU benchmark constructed entirely through a controlled, generation-based pipeline. By leveraging recent advances in video generation models, Pistachio provides precise control over scenes, anomaly types, and temporal narratives, effectively eliminating the biases and limitations of Internet-collected datasets. Our pipeline integrates scene-conditioned anomaly assignment, multi-step storyline generation, and a temporally consistent long-form synthesis strategy that produces coherent 41-second videos with minimal human intervention. Extensive experiments demonstrate the scale, diversity, and complexity of Pistachio, revealing new challenges for existing methods and motivating future research on dynamic and multi-event anomaly understanding.
[162] Comparative Analysis of Vision Transformer, Convolutional, and Hybrid Architectures for Mental Health Classification Using Actigraphy-Derived Images
Ifeanyi Okala
Main category: cs.CV
TL;DR: CoAtNet-Tiny outperforms VGG16 and ViT-B/16 in classifying depression, schizophrenia, and healthy controls using actigraphy-derived images, showing better accuracy, stability, and performance on underrepresented classes.
Details
Motivation: To evaluate and compare different image-based deep learning methods (VGG16, ViT-B/16, CoAtNet-Tiny) for identifying mental health conditions (depression, schizophrenia, healthy controls) using actigraphy data converted to images.Method: Converted wrist-worn activity signals from Psykose and Depresjon datasets into 30x48 pixel images, then evaluated three models (VGG16, ViT-B/16, CoAtNet-Tiny) using three-fold subject-wise cross-validation on the classification task.
Result: CoAtNet-Tiny achieved the highest average accuracy, most stable performance across folds, and best precision/recall/F1-scores, especially for underrepresented depression and schizophrenia classes. VGG16 showed steady improvement but lower accuracy, while ViT-B/16 had strong but inconsistent performance.
Conclusion: CoAtNet-Tiny is the most reliable method for mental health classification using actigraphy-derived images, suggesting hybrid architectures may be particularly suitable for this application domain.
Abstract: This work examines how three different image-based methods, VGG16, ViT-B/16, and CoAtNet-Tiny, perform in identifying depression, schizophrenia, and healthy controls using daily actigraphy records. Wrist-worn activity signals from the Psykose and Depresjon datasets were converted into 30 by 48 images and evaluated through a three-fold subject-wise split. Although all methods fitted the training data well, their behaviour on unseen data differed. VGG16 improved steadily but often settled at lower accuracy. ViT-B/16 reached strong results in some runs, but its performance shifted noticeably from fold to fold. CoAtNet-Tiny stood out as the most reliable, recording the highest average accuracy and the most stable curves across folds. It also produced the strongest precision, recall, and F1-scores, particularly for the underrepresented depression and schizophrenia classes. Overall, the findings indicate that CoAtNet-Tiny performed most consistently on the actigraphy images, while VGG16 and ViT-B/16 yielded mixed results. These observations suggest that certain hybrid designs may be especially suited for mental-health work that relies on actigraphy-derived images.
[163] Hybrid Synthetic Data Generation with Domain Randomization Enables Zero-Shot Vision-Based Part Inspection Under Extreme Class Imbalance
Ruo-Syuan Mei, Sixian Jia, Guangze Li, Soo Yeon Lee, Brian Musser, William Keller, Sreten Zakula, Jorge Arinez, Chenhui Shao
Main category: cs.CV
TL;DR: Hybrid synthetic data generation framework enables zero-shot learning for industrial quality inspection without manual annotation, achieving 90-91% balanced accuracy under severe class imbalance.
Details
Motivation: Real-world manufacturing faces data constraints: expensive labeled data collection, rare defective samples causing severe class imbalance, and limited adoption of ML-based quality inspection methods.Method: Hybrid SDG framework combining simulation-based rendering, domain randomization, and real background compositing. Two-stage architecture with YOLOv8n for detection and MobileNetV3-small for classification, trained exclusively on synthetic data.
Result: Achieves mAP@0.5 of 0.995 for detection, 96% classification accuracy, and 90.1% balanced accuracy. Outperforms few-shot real-data baselines (90-91% vs 50% balanced accuracy under severe imbalance).
Conclusion: The proposed SDG-based approach enables annotation-free, scalable, and robust quality inspection for real-world manufacturing applications, overcoming data constraints through synthetic data generation.
Abstract: Machine learning, particularly deep learning, is transforming industrial quality inspection. Yet, training robust machine learning models typically requires large volumes of high-quality labeled data, which are expensive, time-consuming, and labor-intensive to obtain in manufacturing. Moreover, defective samples are intrinsically rare, leading to severe class imbalance that degrades model performance. These data constraints hinder the widespread adoption of machine learning-based quality inspection methods in real production environments. Synthetic data generation (SDG) offers a promising solution by enabling the creation of large, balanced, and fully annotated datasets in an efficient, cost-effective, and scalable manner. This paper presents a hybrid SDG framework that integrates simulation-based rendering, domain randomization, and real background compositing to enable zero-shot learning for computer vision-based industrial part inspection without manual annotation. The SDG pipeline generates 12,960 labeled images in one hour by varying part geometry, lighting, and surface properties, and then compositing synthetic parts onto real image backgrounds. A two-stage architecture utilizing a YOLOv8n backbone for object detection and MobileNetV3-small for quality classification is trained exclusively on synthetic data and evaluated on 300 real industrial parts. The proposed approach achieves an mAP@0.5 of 0.995 for detection, 96% classification accuracy, and 90.1% balanced accuracy. Comparative evaluation against few-shot real-data baseline approaches demonstrates significant improvement. The proposed SDG-based approach achieves 90-91% balanced accuracy under severe class imbalance, while the baselines reach only 50% accuracy. These results demonstrate that the proposed method enables annotation-free, scalable, and robust quality inspection for real-world manufacturing applications.
[164] Analysis of Incursive Breast Cancer in Mammograms Using YOLO, Explainability, and Domain Adaptation
Jayan Adhikari, Prativa Joshi, Susish Baral
Main category: cs.CV
TL;DR: A deep learning framework combining ResNet50-based OOD filtering with YOLO architectures for reliable breast cancer detection from mammograms, achieving high accuracy while rejecting non-mammographic inputs.
Details
Motivation: Deep learning models for breast cancer detection suffer from reliability issues when presented with Out-of-Distribution (OOD) inputs like other imaging modalities or equipment variations, leading to unreliable detection and potential misdiagnosis in clinical settings.Method: Integrated approach using ResNet50-based OOD filtering (selected after 12 CNN architecture searches) to reject non-mammographic inputs via cosine similarity, combined with YOLO architectures (YOLOv8, YOLOv11, YOLOv12) for breast cancer detection. The system establishes an in-domain gallery to ensure only domain-associated images enter the detection pipeline.
Result: OOD detection achieves 99.77% general accuracy with perfect 100% accuracy on OOD test sets. The joint framework achieves high detection performance (mAP@0.5: 0.947) with enhanced interpretability via Grad-CAM visualizations. OOD filtering significantly improves system reliability by preventing false alarms on out-of-distribution inputs.
Conclusion: The study provides a fundamental foundation for deploying reliable AI-based breast cancer detection systems in diverse clinical environments with inherent data heterogeneity, addressing the critical OOD problem while maintaining high detection accuracy on mammographic data.
Abstract: Deep learning models for breast cancer detection from mammographic images have significant reliability problems when presented with Out-of-Distribution (OOD) inputs such as other imaging modalities (CT, MRI, X-ray) or equipment variations, leading to unreliable detection and misdiagnosis. The current research mitigates the fundamental OOD issue through a comprehensive approach integrating ResNet50-based OOD filtering with YOLO architectures (YOLOv8, YOLOv11, YOLOv12) for accurate detection of breast cancer. Our strategy establishes an in-domain gallery via cosine similarity to rigidly reject non-mammographic inputs prior to processing, ensuring that only domain-associated images supply the detection pipeline. The OOD detection component achieves 99.77% general accuracy with immaculate 100% accuracy on OOD test sets, effectively eliminating irrelevant imaging modalities. ResNet50 was selected as the optimum backbone after 12 CNN architecture searches. The joint framework unites OOD robustness with high detection performance (mAP@0.5: 0.947) and enhanced interpretability through Grad-CAM visualizations. Experimental validation establishes that OOD filtering significantly improves system reliability by preventing false alarms on out-of-distribution inputs while maintaining higher detection accuracy on mammographic data. The present study offers a fundamental foundation for the deployment of reliable AI-based breast cancer detection systems in diverse clinical environments with inherent data heterogeneity.
[165] Local and Global Context-and-Object-part-Aware Superpixel-based Data Augmentation for Deep Visual Recognition
Fadi Dornaika, Danyang Sun
Main category: cs.CV
TL;DR: LGCOAMix is a novel CutMix-based data augmentation method that uses superpixel attention and grid blending to better preserve local discriminative features and object parts, addressing limitations of existing rectangular region methods.
Details
Motivation: Existing CutMix methods focus too much on global semantics, losing attention to discriminative local context and object parts. They use inefficient rectangular regions and require double forward propagation or external networks for label consistency.Method: Proposes LGCOAMix: an efficient context-aware and object-part-aware superpixel-based grid blending method. Uses superpixel attention approach for label mixing, learns local features from discriminative superpixel regions and cross-image superpixel contrasts.
Result: Outperforms state-of-the-art CutMix methods on classification tasks and weakly supervised object location on CUB200-2011. Effective for both CNN and Transformer networks.
Conclusion: LGCOAMix successfully addresses limitations of existing CutMix methods by preserving local discriminative features through superpixel attention, achieving better performance while being more efficient.
Abstract: Cutmix-based data augmentation, which uses a cut-and-paste strategy, has shown remarkable generalization capabilities in deep learning. However, existing methods primarily consider global semantics with image-level constraints, which excessively reduces attention to the discriminative local context of the class and leads to a performance improvement bottleneck. Moreover, existing methods for generating augmented samples usually involve cutting and pasting rectangular or square regions, resulting in a loss of object part information. To mitigate the problem of inconsistency between the augmented image and the generated mixed label, existing methods usually require double forward propagation or rely on an external pre-trained network for object centering, which is inefficient. To overcome the above limitations, we propose LGCOAMix, an efficient context-aware and object-part-aware superpixel-based grid blending method for data augmentation. To the best of our knowledge, this is the first time that a label mixing strategy using a superpixel attention approach has been proposed for cutmix-based data augmentation. It is the first instance of learning local features from discriminative superpixel-wise regions and cross-image superpixel contrasts. Extensive experiments on various benchmark datasets show that LGCOAMix outperforms state-of-the-art cutmix-based data augmentation methods on classification tasks, {and weakly supervised object location on CUB200-2011.} We have demonstrated the effectiveness of LGCOAMix not only for CNN networks, but also for Transformer networks. Source codes are available at https://github.com/DanielaPlusPlus/LGCOAMix.
[166] Mammo-FM: Breast-specific foundational model for Integrated Mammographic Diagnosis, Prognosis, and Reporting
Shantanu Ghosh, Vedant Parthesh Joshi, Rayan Syed, Aya Kassem, Abhishek Varshney, Payel Basak, Weicheng Dai, Judy Wawira Gichoya, Hari M. Trivedi, Imon Banerjee, Shyam Visweswaran, Clare B. Poynton, Kayhan Batmanghelich
Main category: cs.CV
TL;DR: Mammo-FM is the first mammography-specific foundation model pretrained on 821,326 mammograms from 140,677 patients across 4 US institutions, outperforming generalist models with fewer parameters across diagnosis, prognosis, and report generation tasks.
Details
Motivation: Breast cancer is a leading cause of death among women worldwide, creating a need for specialized AI models in mammography. Current generalist foundation models lack domain-specific optimization for breast imaging tasks and clinical interpretability requirements.Method: Developed Mammo-FM, a domain-specific foundation model pretrained on the largest and most diverse mammography dataset to date (821,326 mammograms from 140,677 patients across 4 US institutions). The model aligns images with text for visual and textual interpretability, enabling unified handling of diagnosis, pathology localization, report generation, and risk prognosis within a single framework.
Result: Mammo-FM consistently outperforms state-of-the-art generalist foundation models across multiple public and private benchmarks for diagnosis, prognosis, and report generation, despite using only one-third of the parameters and operating on native-resolution mammograms. The model demonstrates superior performance in both in-distribution and out-of-distribution datasets.
Conclusion: Domain-specific foundation models like Mammo-FM designed around the full spectrum of clinical tasks offer greater efficiency and value than generalist models. The results emphasize the importance of domain-aligned evaluation and specialized models for real-world clinical adoption, particularly for improving transparency and auditability in breast cancer screening.
Abstract: Breast cancer is one of the leading causes of death among women worldwide. We introduce Mammo-FM, the first foundation model specifically for mammography, pretrained on the largest and most diverse dataset to date - 140,677 patients (821,326 mammograms) across four U.S. institutions. Mammo-FM provides a unified foundation for core clinical tasks in breast imaging, including cancer diagnosis, pathology localization, structured report generation, and cancer risk prognosis within a single framework. Its alignment between images and text enables both visual and textual interpretability, improving transparency and clinical auditability, which are essential for real-world adoption. We rigorously evaluate Mammo-FM across diagnosis, prognosis, and report-generation tasks in in- and out-of-distribution datasets. Despite operating on native-resolution mammograms and using only one-third of the parameters of state-of-the-art generalist FMs, Mammo-FM consistently outperforms them across multiple public and private benchmarks. These results highlight the efficiency and value of domain-specific foundation models designed around the full spectrum of tasks within a clinical domain and emphasize the importance of rigorous, domain-aligned evaluation.
[167] ReactionMamba: Generating Short &Long Human Reaction Sequences
Hajra Anwar Beg, Baptiste Chopin, Hao Tang, Mohamed Daoudi
Main category: cs.CV
TL;DR: ReactionMamba is a new framework for generating long 3D human reaction motions using motion VAE encoding and Mamba-based state-space models for decoding, achieving competitive performance with faster inference.
Details
Motivation: The paper aims to address the challenge of generating long, temporally consistent 3D human reaction motions, particularly for complex activities like dance and martial arts, where existing methods may struggle with sequence length and computational efficiency.Method: ReactionMamba combines a motion VAE for efficient motion encoding with Mamba-based state-space models for decoding, enabling generation of both short simple motions and long complex motion sequences while maintaining temporal consistency.
Result: The framework demonstrates competitive performance on NTU120-AS, Lindy Hop, and InterX datasets in terms of realism, diversity, and long-sequence generation compared to InterFormer, ReMoS, and Ready-to-React, with substantial improvements in inference speed.
Conclusion: ReactionMamba provides an effective solution for generating long 3D human reaction motions with improved computational efficiency, making it suitable for applications requiring complex motion generation like dance and martial arts.
Abstract: We present ReactionMamba, a novel framework for generating long 3D human reaction motions. Reaction-Mamba integrates a motion VAE for efficient motion encoding with Mamba-based state-space models to decode temporally consistent reactions. This design enables ReactionMamba to generate both short sequences of simple motions and long sequences of complex motions, such as dance and martial arts. We evaluate ReactionMamba on three datasets–NTU120-AS, Lindy Hop, and InterX–and demonstrate competitive performance in terms of realism, diversity, and long-sequence generation compared to previous methods, including InterFormer, ReMoS, and Ready-to-React, while achieving substantial improvements in inference speed.
[168] DenseScan: Advancing 3D Scene Understanding with 2D Dense Annotation
Zirui Wang, Tao Zhang
Main category: cs.CV
TL;DR: DenseScan is a novel 3D scene understanding dataset with detailed multi-level descriptions generated automatically using multi-view 2D images and multimodal LLMs, enabling dense captioning and scenario-based question generation for enhanced visual-language tasks.
Details
Motivation: Current 3D scene understanding datasets lack rich semantic annotations needed for nuanced visual-language tasks, limiting their utility for real-world AI applications that require comprehensive object-level descriptions and contextual understanding.Method: Automated pipeline leveraging multi-view 2D images and multimodal large language models (MLLMs) to generate dense captions of scene elements, followed by scenario-based question generation that integrates object properties, spatial relationships, and scene context.
Result: DenseScan significantly enhances object-level understanding and question-answering performance in 3D environments compared to traditional annotation pipelines, providing comprehensive annotations that couple geometric detail with semantic richness.
Conclusion: DenseScan broadens the range of downstream 3D understanding tasks and aims to catalyze new research avenues in robotics, augmented reality, and other applications by providing richer, more contextually aware annotations for real-world environments.
Abstract: 3D understanding is a key capability for real-world AI assistance. High-quality data plays an important role in driving the development of the 3D understanding community. Current 3D scene understanding datasets often provide geometric and instance-level information, yet they lack the rich semantic annotations necessary for nuanced visual-language tasks.In this work, we introduce DenseScan, a novel dataset with detailed multi-level descriptions generated by an automated pipeline leveraging multi-view 2D images and multimodal large language models (MLLMs). Our approach enables dense captioning of scene elements, ensuring comprehensive object-level descriptions that capture context-sensitive details. Furthermore, we extend these annotations through scenario-based question generation, producing high-level queries that integrate object properties, spatial relationships, and scene context. By coupling geometric detail with semantic richness, DenseScan broadens the range of downstream tasks, from detailed visual-language navigation to interactive question answering. Experimental results demonstrate that our method significantly enhances object-level understanding and question-answering performance in 3D environments compared to traditional annotation pipelines. We release both the annotated dataset and our annotation pipeline to facilitate future research and applications in robotics, augmented reality, and beyond. Through DenseScan, we aim to catalyze new avenues in 3D scene understanding, allowing researchers and practitioners to tackle the complexities of real-world environments with richer, more contextually aware annotations.
[169] Relightable Holoported Characters: Capturing and Relighting Dynamic Human Performance from Sparse Views
Kunwar Maheep Singh, Jianchun Chen, Vladislav Golyanik, Stephan J. Garbin, Thabo Beeler, Rishabh Dabral, Marc Habermann, Christian Theobalt
Main category: cs.CV
TL;DR: RHC is a person-specific method for free-view rendering and relighting of dynamic humans from sparse-view RGB videos, using a transformer-based RelightNet that predicts relit appearance in a single pass without OLAT capture.
Details
Motivation: Classical OLAT-based human relighting requires costly one-light-at-a-time basis capture and generation. The authors aim to enable efficient relighting of dynamic humans from sparse views without OLAT capture.Method: 1) New capture strategy with alternating random environment maps and uniformly lit tracking frames. 2) Physics-informed features encoding geometry, albedo, shading, and camera view from coarse mesh proxy. 3) Transformer-based RelightNet uses cross-attention with novel lighting conditions to regress relit appearance as texel-aligned 3D Gaussian splats.
Result: Superior visual fidelity and lighting reproduction compared to state-of-the-art approaches, enabling efficient single-pass relighting of highly dynamic humans.
Conclusion: RHC demonstrates a practical approach for free-view rendering and relighting of dynamic humans from sparse views, eliminating the need for costly OLAT capture while maintaining high-quality results.
Abstract: We present Relightable Holoported Characters (RHC), a novel person-specific method for free-view rendering and relighting of full-body and highly dynamic humans solely observed from sparse-view RGB videos at inference. In contrast to classical one-light-at-a-time (OLAT)-based human relighting, our transformer-based RelightNet predicts relit appearance within a single network pass, avoiding costly OLAT-basis capture and generation. For training such a model, we introduce a new capture strategy and dataset recorded in a multi-view lightstage, where we alternate frames lit by random environment maps with uniformly lit tracking frames, simultaneously enabling accurate motion tracking and diverse illumination as well as dynamics coverage. Inspired by the rendering equation, we derive physics-informed features that encode geometry, albedo, shading, and the virtual camera view from a coarse human mesh proxy and the input views. Our RelightNet then takes these features as input and cross-attends them with a novel lighting condition, and regresses the relit appearance in the form of texel-aligned 3D Gaussian splats attached to the coarse mesh proxy. Consequently, our RelightNet implicitly learns to efficiently compute the rendering equation for novel lighting conditions within a single feed-forward pass. Experiments demonstrate our method’s superior visual fidelity and lighting reproduction compared to state-of-the-art approaches. Project page: https://vcai.mpi-inf.mpg.de/projects/RHC/
[170] UniDiff: Parameter-Efficient Adaptation of Diffusion Models for Land Cover Classification with Multi-Modal Remotely Sensed Imagery and Sparse Annotations
Yuzhen Hu, Saurabh Prasad
Main category: cs.CV
TL;DR: UniDiff: Parameter-efficient adaptation of ImageNet-pretrained diffusion model to multiple remote sensing modalities (HSI, SAR) using only target-domain data, enabling effective multimodal fusion under sparse annotations.
Details
Motivation: Sparse annotations constrain multimodal remote sensing applications; even state-of-the-art supervised methods like MSFMamba are limited by labeled data availability. ImageNet-pretrained models offer rich visual representations but adapting them to heterogeneous modalities (HSI, SAR) without large labeled datasets remains challenging.Method: UniDiff framework combines: 1) FiLM-based timestep-modality conditioning, 2) Parameter-efficient adaptation of ~5% of parameters, and 3) Pseudo-RGB anchoring to preserve pre-trained representations and prevent catastrophic forgetting. Adapts a single ImageNet-pretrained diffusion model to multiple sensing modalities using only target-domain data.
Result: Demonstrated on two established multi-modal benchmarking datasets. Shows that unsupervised adaptation of pre-trained diffusion model effectively mitigates annotation constraints and achieves effective fusion of multi-modal remotely sensed data.
Conclusion: UniDiff enables effective feature extraction from remote sensing data under sparse annotations by adapting pre-trained diffusion models to heterogeneous modalities without requiring large labeled datasets, addressing practical deployment limitations of supervised methods.
Abstract: Sparse annotations fundamentally constrain multimodal remote sensing: even recent state-of-the-art supervised methods such as MSFMamba are limited by the availability of labeled data, restricting their practical deployment despite architectural advances. ImageNet-pretrained models provide rich visual representations, but adapting them to heterogeneous modalities such as hyperspectral imaging (HSI) and synthetic aperture radar (SAR) without large labeled datasets remains challenging. We propose UniDiff, a parameter-efficient framework that adapts a single ImageNet-pretrained diffusion model to multiple sensing modalities using only target-domain data. UniDiff combines FiLM-based timestep-modality conditioning, parameter-efficient adaptation of approximately 5% of parameters, and pseudo-RGB anchoring to preserve pre-trained representations and prevent catastrophic forgetting. This design enables effective feature extraction from remote sensing data under sparse annotations. Our results with two established multi-modal benchmarking datasets demonstrate that unsupervised adaptation of a pre-trained diffusion model effectively mitigates annotation constraints and achieves effective fusion of multi-modal remotely sensed data.
[171] HeartFormer: Semantic-Aware Dual-Structure Transformers for 3D Four-Chamber Cardiac Point Cloud Reconstruction
Zhengda Ma, Abhirup Banerjee
Main category: cs.CV
TL;DR: HeartFormer: First geometric deep learning framework using point clouds for 3D four-chamber cardiac reconstruction from 2D cine MRI, with new multi-class point cloud completion approach and large-scale dataset.
Details
Motivation: Conventional cine MRI only provides 2D slice images, limiting comprehensive understanding of cardiac morphology and physiology in both healthy and pathological conditions. There's a need for 3D reconstruction from sparse 2D data.Method: HeartFormer: Novel point cloud completion network with two components: 1) Semantic-Aware Dual-Structure Transformer Network (SA-DSTNet) generates initial coarse point cloud with global and substructure geometry features, 2) Semantic-Aware Geometry Feature Refinement Transformer Network (SA-GFRTNet) progressively refines coarse output using semantic-geometry representations.
Result: HeartFormer achieves robust, accurate, and generalizable performance, consistently surpassing state-of-the-art methods in cross-domain experiments on HeartCompv1 (new dataset) and UK Biobank. Created HeartCompv1 - first publicly available large-scale dataset with 17,000 high-resolution 3D multi-class cardiac meshes and point-clouds.
Conclusion: HeartFormer provides the first geometric deep learning framework for 3D cardiac reconstruction from cine MRI, addressing limitations of 2D imaging and establishing a benchmark for this emerging research direction with promising clinical applications.
Abstract: We present the first geometric deep learning framework based on point cloud representation for 3D four-chamber cardiac reconstruction from cine MRI data. This work addresses a long-standing limitation in conventional cine MRI, which typically provides only 2D slice images of the heart, thereby restricting a comprehensive understanding of cardiac morphology and physiological mechanisms in both healthy and pathological conditions. To overcome this, we propose \textbf{HeartFormer}, a novel point cloud completion network that extends traditional single-class point cloud completion to the multi-class. HeartFormer consists of two key components: a Semantic-Aware Dual-Structure Transformer Network (SA-DSTNet) and a Semantic-Aware Geometry Feature Refinement Transformer Network (SA-GFRTNet). SA-DSTNet generates an initial coarse point cloud with both global geometry features and substructure geometry features. Guided by these semantic-geometry representations, SA-GFRTNet progressively refines the coarse output, effectively leveraging both global and substructure geometric priors to produce high-fidelity and geometrically consistent reconstructions. We further construct \textbf{HeartCompv1}, the first publicly available large-scale dataset with 17,000 high-resolution 3D multi-class cardiac meshes and point-clouds, to establish a general benchmark for this emerging research direction. Extensive cross-domain experiments on HeartCompv1 and UK Biobank demonstrate that HeartFormer achieves robust, accurate, and generalizable performance, consistently surpassing state-of-the-art (SOTA) methods. Code and dataset will be released upon acceptance at: https://github.com/10Darren/HeartFormer.
[172] USB: Unified Synthetic Brain Framework for Bidirectional Pathology-Healthy Generation and Editing
Jun Wang, Peirong Liu
Main category: cs.CV
TL;DR: USB is the first unified framework for bidirectional generation and editing of pathological and healthy brain images using paired diffusion and consistency guidance.
Details
Motivation: Paired pathological-healthy brain data is extremely difficult to obtain due to clinical constraints, while existing methods are domain-specific and treat pathological and healthy image modeling independently.Method: USB uses a paired diffusion mechanism to model the joint distribution of lesions and brain anatomy, with a consistency guidance algorithm to preserve anatomical consistency and lesion correspondence during bidirectional editing.
Result: Extensive experiments on six public brain MRI datasets (healthy controls, stroke, Alzheimer’s patients) demonstrate USB’s ability to produce diverse and realistic results.
Conclusion: USB establishes the first unified benchmark for brain image generation and editing, opening opportunities for scalable dataset creation and robust neuroimaging analysis.
Abstract: Understanding the relationship between pathological and healthy brain structures is fundamental to neuroimaging, connecting disease diagnosis and detection with modeling, prediction, and treatment planning. However, paired pathological-healthy data are extremely difficult to obtain, as they rely on pre- and post-treatment imaging, constrained by clinical outcomes and longitudinal data availability. Consequently, most existing brain image generation and editing methods focus on visual quality yet remain domain-specific, treating pathological and healthy image modeling independently. We introduce USB (Unified Synthetic Brain), the first end-to-end framework that unifies bidirectional generation and editing of pathological and healthy brain images. USB models the joint distribution of lesions and brain anatomy through a paired diffusion mechanism and achieves both pathological and healthy image generation. A consistency guidance algorithm further preserves anatomical consistency and lesion correspondence during bidirectional pathology-healthy editing. Extensive experiments on six public brain MRI datasets including healthy controls, stroke, and Alzheimer’s patients, demonstrate USB’s ability to produce diverse and realistic results. By establishing the first unified benchmark for brain image generation and editing, USB opens opportunities for scalable dataset creation and robust neuroimaging analysis. Code is available at https://github.com/jhuldr/USB.
[173] HIMOSA: Efficient Remote Sensing Image Super-Resolution with Hierarchical Mixture of Sparse Attention
Yi Liu, Yi Wan, Xinyi Liu, Qiong Wu, Panwang Xia, Xuejun Huang, Yongjun Zhang
Main category: cs.CV
TL;DR: HIMOSA: A lightweight super-resolution framework for remote sensing imagery using content-aware sparse attention and hierarchical window expansion to balance performance and efficiency.
Details
Motivation: Remote sensing applications like disaster detection require real-time efficiency and lightweight models. Existing super-resolution methods struggle to balance model performance with computational efficiency, creating a critical trade-off problem.Method: Proposes HIMOSA framework that leverages inherent redundancy in remote sensing imagery with content-aware sparse attention mechanism. Uses hierarchical window expansion to exploit multi-scale repetitive patterns while reducing computational complexity through attention sparsity adjustment.
Result: Extensive experiments on multiple remote sensing datasets show the method achieves state-of-the-art performance while maintaining computational efficiency.
Conclusion: HIMOSA successfully addresses the performance-efficiency trade-off in remote sensing super-resolution through innovative sparse attention and hierarchical processing, enabling fast inference with strong reconstruction capabilities.
Abstract: In remote sensing applications, such as disaster detection and response, real-time efficiency and model lightweighting are of critical importance. Consequently, existing remote sensing image super-resolution methods often face a trade-off between model performance and computational efficiency. In this paper, we propose a lightweight super-resolution framework for remote sensing imagery, named HIMOSA. Specifically, HIMOSA leverages the inherent redundancy in remote sensing imagery and introduces a content-aware sparse attention mechanism, enabling the model to achieve fast inference while maintaining strong reconstruction performance. Furthermore, to effectively leverage the multi-scale repetitive patterns found in remote sensing imagery, we introduce a hierarchical window expansion and reduce the computational complexity by adjusting the sparsity of the attention. Extensive experiments on multiple remote sensing datasets demonstrate that our method achieves state-of-the-art performance while maintaining computational efficiency.
[174] U-FaceBP: Uncertainty-aware Bayesian Ensemble Deep Learning for Face Video-based Blood Pressure Measurement
Yusuke Akamatsu, Akinori F. Ebihara, Terumi Umematsu
Main category: cs.CV
TL;DR: U-FaceBP: Uncertainty-aware Bayesian ensemble deep learning method for face video-based blood pressure measurement using remote photoplethysmography.
Details
Motivation: Remote photoplethysmography (rPPG) from face videos offers convenient BP measurement without specialized medical devices, but existing methods suffer from uncertainties that limit performance and reliability.Method: U-FaceBP uses Bayesian neural networks to model aleatoric and epistemic uncertainties, and employs ensemble learning to estimate BP from rPPG signals, derived PPG signals, and face images using multiple BNNs.
Result: Outperforms state-of-the-art BP estimation methods on two datasets with 1197 subjects from diverse racial groups. Uncertainty estimates are informative for modality fusion, reliability assessment, and racial group analysis.
Conclusion: U-FaceBP provides a reliable, uncertainty-aware approach for face video-based BP measurement with practical applications in healthcare monitoring.
Abstract: Blood pressure (BP) measurement is crucial for daily health assessment. Remote photoplethysmography (rPPG), which extracts pulse waves from face videos captured by a camera, has the potential to enable convenient BP measurement without specialized medical devices. However, there are various uncertainties in BP estimation using rPPG, leading to limited estimation performance and reliability. In this paper, we propose U-FaceBP, an uncertainty-aware Bayesian ensemble deep learning method for face video-based BP measurement. U-FaceBP models aleatoric and epistemic uncertainties in face video-based BP estimation with a Bayesian neural network (BNN). Additionally, we design U-FaceBP as an ensemble method, estimating BP from rPPG signals, PPG signals derived from face videos, and face images using multiple BNNs. Large-scale experiments on two datasets involving 1197 subjects from diverse racial groups demonstrate that U-FaceBP outperforms state-of-the-art BP estimation methods. Furthermore, we show that the uncertainty estimates provided by U-FaceBP are informative and useful for guiding modality fusion, assessing prediction reliability, and analyzing performance across racial groups.
[175] Rethinking Lung Cancer Screening: AI Nodule Detection and Diagnosis Outperforms Radiologists, Leading Models, and Standards Beyond Size and Growth
Sylvain Bodard, Pierre Baudot, Benjamin Renoust, Charles Voyton, Gwendoline De Bie, Ezequiel Geremia, Van-Khoa Le, Danny Francis, Pierre-Henri Siot, Yousra Haddou, Vincent Bobin, Jean-Christophe Brisset, Carey C. Thomson, Valerie Bourdes, Benoit Huet
Main category: cs.CV
TL;DR: AI system for lung cancer screening that detects and diagnoses malignancy at nodule level on CT scans, outperforming radiologists and existing AI models with high accuracy across all nodule sizes and stages.
Details
Motivation: Current lung cancer screening depends on nodule size and growth monitoring, which inherently delays diagnosis. There's a need for more immediate malignancy assessment at the nodule level to enable earlier detection and treatment.Method: Developed an ensemble system combining shallow deep learning models with feature-based specialized models. Trained on 25,709 CT scans with 69,449 annotated nodules to address dataset scale and explainability limitations.
Result: Achieved AUC of 0.98 internally and 0.945 on independent cohort. Outperformed radiologists, Lung-RADS, and leading AI models (Sybil, Brock, Google, Kaggle) with 0.5 false positives per scan at 99.3% sensitivity. Excelled across all nodule sizes, stages (especially stage 1), and growth metrics including Volume-Doubling Time.
Conclusion: The AI system addresses key barriers to AI adoption in lung cancer screening by providing earlier and more accurate malignancy diagnosis than current methods, potentially enabling earlier intervention and improved patient outcomes.
Abstract: Early detection of malignant lung nodules is critical, but its dependence on size and growth in screening inherently delays diagnosis. We present an AI system that redefines lung cancer screening by performing both detection and malignancy diagnosis directly at the nodule level on low-dose CT scans. To address limitations in dataset scale and explainability, we designed an ensemble of shallow deep learning and feature-based specialized models. Trained and evaluated on 25,709 scans with 69,449 annotated nodules, the system outperforms radiologists, Lung-RADS, and leading AI models (Sybil, Brock, Google, Kaggle). It achieves an area under the receiver operating characteristic curve (AUC) of 0.98 internally and 0.945 on an independent cohort. With 0.5 false positives per scan at 99.3% sensitivity, it addresses key barriers to AI adoption. Critically, it outperforms radiologists across all nodule sizes and stages, excelling in stage 1 cancers, and all growth-based metrics, including the least accurate: Volume-Doubling Time. It also surpasses radiologists by up to one year in diagnosing indeterminate and slow-growing nodules.
[176] AgriPotential: A Novel Multi-Spectral and Multi-Temporal Remote Sensing Dataset for Agricultural Potentials
Mohammad El Sakka, Caroline De Pourtales, Lotfi Chaari, Josiane Mothe
Main category: cs.CV
TL;DR: AgriPotential is a new benchmark dataset of Sentinel-2 satellite imagery with pixel-level annotations for agricultural potential prediction of three crop types across five ordinal classes.
Details
Motivation: Remote sensing is crucial for large-scale Earth monitoring and land management, but there's a lack of public datasets specifically designed for agricultural potential prediction to support sustainable land use planning.Method: Created a novel benchmark dataset composed of Sentinel-2 satellite imagery captured over multiple months, with pixel-level annotations of agricultural potentials for three major crop types (viticulture, market gardening, field crops) across five ordinal classes.
Result: AgriPotential is the first public dataset specifically designed for agricultural potential prediction, covering diverse areas in Southern France with rich spectral information, supporting multiple machine learning tasks including ordinal regression, multi-label classification, and spatio-temporal modeling.
Conclusion: AgriPotential aims to improve data-driven approaches to sustainable land use planning by providing a comprehensive benchmark for agricultural potential prediction, with the dataset and code freely accessible to the research community.
Abstract: Remote sensing has emerged as a critical tool for large-scale Earth monitoring and land management. In this paper, we introduce AgriPotential, a novel benchmark dataset composed of Sentinel-2 satellite imagery captured over multiple months. The dataset provides pixel-level annotations of agricultural potentials for three major crop types - viticulture, market gardening, and field crops - across five ordinal classes. AgriPotential supports a broad range of machine learning tasks, including ordinal regression, multi-label classification, and spatio-temporal modeling. The data cover diverse areas in Southern France, offering rich spectral information. AgriPotential is the first public dataset designed specifically for agricultural potential prediction, aiming to improve data-driven approaches to sustainable land use planning. The dataset and the code are freely accessible at: https://zenodo.org/records/15551829
[177] Words into World: A Task-Adaptive Agent for Language-Guided Spatial Retrieval in AR
Lixing Guo, Tobias Höllerer
Main category: cs.CV
TL;DR: A modular AR agent system that integrates MLLMs with grounded vision models for open-vocabulary spatial reasoning and language-conditioned retrieval in physical environments, enabling complex relational understanding and meter-accurate 3D localization.
Details
Motivation: Traditional AR systems rely on fixed class detectors or fiducial markers, limiting their ability to handle complex, open-vocabulary natural language queries and understand relational reasoning in physical spaces.Method: A modular AR agent system that coordinates multimodal large language models (MLLMs) with coordinate-aware perception tools. It constructs dynamic AR scene graphs encoding nine typed relations (spatial, structural-semantic, causal-functional), uses task-adaptive region-of-interest highlighting, and supports human-in-the-loop refinement. The system dynamically invokes coordinate-aware tools for complex queries like selection, measurement, comparison, and actuation.
Result: The system enables relational reasoning in space and language-conditioned spatial retrieval, handles varying query complexities from simple object identification to multi-object relational reasoning, and returns meter-accurate 3D anchors. It also introduces GroundedAR-Bench, an evaluation framework for language-driven real-world localization and relation grounding.
Conclusion: The modular architecture establishes AR agents as intermediaries that augment MLLMs with real-world spatial intelligence for interactive scene understanding, supporting plug-and-use vision-language models without retraining, bridging language understanding with physical operations.
Abstract: Traditional augmented reality (AR) systems predominantly rely on fixed class detectors or fiducial markers, limiting their ability to interpret complex, open-vocabulary natural language queries. We present a modular AR agent system that integrates multimodal large language models (MLLMs) with grounded vision models to enable relational reasoning in space and language-conditioned spatial retrieval in physical environments. Our adaptive task agent coordinates MLLMs and coordinate-aware perception tools to address varying query complexities, ranging from simple object identification to multi-object relational reasoning, while returning meter-accurate 3D anchors. It constructs dynamic AR scene graphs encoding nine typed relations (spatial, structural-semantic, causal-functional), enabling MLLMs to understand not just what objects exist, but how they relate and interact in 3D space. Through task-adaptive region-of-interest highlighting and contextual spatial retrieval, the system guides human attention to information-dense areas while supporting human-in-the-loop refinement. The agent dynamically invokes coordinate-aware tools for complex queries-selection, measurement, comparison, and actuation-grounding language understanding in physical operations. The modular architecture supports plug-and-use vision-language models without retraining, establishing AR agents as intermediaries that augment MLLMs with real-world spatial intelligence for interactive scene understanding. We also introduce GroundedAR-Bench, an evaluation framework for language-driven real world localization and relation grounding across diverse environments.
[178] Multigranular Evaluation for Brain Visual Decoding
Weihao Xia, Cengiz Oztireli
Main category: cs.CV
TL;DR: BASIC is a unified multigranular evaluation framework for brain visual decoding that assesses structural fidelity, inferential alignment, and contextual coherence using hierarchical segmentation metrics and multimodal scene representations.
Details
Motivation: Existing evaluation protocols for brain visual decoding rely on coarse metrics that obscure inter-model differences, lack neuroscientific foundation, and fail to capture fine-grained visual distinctions.Method: Introduces BASIC framework with: 1) hierarchical segmentation-based metrics (foreground, semantic, instance, component masks) for structural fidelity, 2) structured scene representations (objects, attributes, relationships) using multimodal LLMs for semantic alignment, and 3) benchmarking across multiple stimulus-neuroimaging datasets.
Result: The framework provides more discriminative, interpretable, and comprehensive evaluation of brain visual decoding methods, enabling detailed comparisons across different granularities of visual information.
Conclusion: BASIC establishes a unified, multigranular evaluation foundation that addresses limitations of existing protocols and enables more meaningful assessment of brain visual decoding methods.
Abstract: Existing evaluation protocols for brain visual decoding predominantly rely on coarse metrics that obscure inter-model differences, lack neuroscientific foundation, and fail to capture fine-grained visual distinctions. To address these limitations, we introduce BASIC, a unified, multigranular evaluation framework that jointly quantifies structural fidelity, inferential alignment, and contextual coherence between decoded and ground-truth images. For the structural level, we introduce a hierarchical suite of segmentation-based metrics, including foreground, semantic, instance, and component masks, anchored in granularity-aware correspondence across mask structures. For the semantic level, we extract structured scene representations encompassing objects, attributes, and relationships using multimodal large language models, enabling detailed, scalable, and context-rich comparisons with ground-truth stimuli. We benchmark a diverse set of visual decoding methods across multiple stimulus-neuroimaging datasets within this unified evaluation framework. Together, these criteria provide a more discriminative, interpretable, and comprehensive foundation for evaluating brain visual decoding methods.
[179] TGSFormer: Scalable Temporal Gaussian Splatting for Embodied Semantic Scene Completion
Rui Qian, Haozhi Cao, Tianchen Deng, Tianxin Hu, Weixiang Guo, Shenghai Yuan, Lihua Xie
Main category: cs.CV
TL;DR: TGSFormer is a scalable Temporal Gaussian Splatting framework for embodied 3D Semantic Scene Completion that achieves state-of-the-art results with fewer primitives and better scalability than existing methods.
Details
Motivation: Existing Gaussian-based methods for embodied 3D Semantic Scene Completion suffer from redundancy, poor scalability to unbounded scenes, and latency/memory issues as scene scale increases. Current approaches either use random initialization within predefined bounds or local depth-guided methods that don't scale well.Method: TGSFormer maintains a persistent Gaussian memory for temporal prediction without relying on image coherence or frame caches. It uses a Dual Temporal Encoder with confidence-aware cross-attention to process current and historical Gaussian features, followed by a Confidence-aware Voxel Fusion module that merges overlapping primitives into voxel-aligned representations while regulating density and maintaining compactness.
Result: Extensive experiments show TGSFormer achieves state-of-the-art results on both local and embodied SSC benchmarks, offering superior accuracy and scalability with significantly fewer primitives while maintaining consistent long-term scene integrity.
Conclusion: TGSFormer provides an effective solution for scalable embodied 3D Semantic Scene Completion that overcomes the limitations of existing Gaussian-based methods, achieving better performance with reduced computational overhead and improved handling of unbounded scenes.
Abstract: Embodied 3D Semantic Scene Completion (SSC) infers dense geometry and semantics from continuous egocentric observations. Most existing Gaussian-based methods rely on random initialization of many primitives within predefined spatial bounds, resulting in redundancy and poor scalability to unbounded scenes. Recent depth-guided approach alleviates this issue but remains local, suffering from latency and memory overhead as scale increases. To overcome these challenges, we propose TGSFormer, a scalable Temporal Gaussian Splatting framework for embodied SSC. It maintains a persistent Gaussian memory for temporal prediction, without relying on image coherence or frame caches. For temporal fusion, a Dual Temporal Encoder jointly processes current and historical Gaussian features through confidence-aware cross-attention. Subsequently, a Confidence-aware Voxel Fusion module merges overlapping primitives into voxel-aligned representations, regulating density and maintaining compactness. Extensive experiments demonstrate that TGSFormer achieves state-of-the-art results on both local and embodied SSC benchmarks, offering superior accuracy and scalability with significantly fewer primitives while maintaining consistent long-term scene integrity. The code will be released upon acceptance.
[180] Optimizing Distributional Geometry Alignment with Optimal Transport for Generative Dataset Distillation
Xiao Cui, Yulei Qin, Wengang Zhou, Hongsheng Li, Houqiang Li
Main category: cs.CV
TL;DR: Proposes a novel dataset distillation method using Optimal Transport to align both global and instance-level distributions, achieving better generalization than previous methods.
Details
Motivation: Existing dataset distillation methods for large-scale datasets focus only on matching global statistics (mean/variance), ignoring critical instance-level characteristics and intraclass variations, leading to suboptimal generalization performance.Method: Reformulates dataset distillation as an Optimal Transport distance minimization problem with three components: (1) OT-guided diffusion sampling for latent distribution alignment, (2) label-image-aligned soft relabeling for adaptive label distributions, and (3) OT-based logit matching for output distribution alignment.
Result: Extensive experiments show consistent outperformance over state-of-the-art methods across diverse architectures and large-scale datasets, achieving at least 4% accuracy improvement under IPC=10 settings for each architecture on ImageNet-1K.
Conclusion: Optimal Transport provides a geometrically faithful framework for dataset distillation that effectively preserves local modes, intra-class patterns, and fine-grained variations, leading to superior generalization performance in an efficient manner.
Abstract: Dataset distillation seeks to synthesize a compact distilled dataset, enabling models trained on it to achieve performance comparable to models trained on the full dataset. Recent methods for large-scale datasets focus on matching global distributional statistics (e.g., mean and variance), but overlook critical instance-level characteristics and intraclass variations, leading to suboptimal generalization. We address this limitation by reformulating dataset distillation as an Optimal Transport (OT) distance minimization problem, enabling fine-grained alignment at both global and instance levels throughout the pipeline. OT offers a geometrically faithful framework for distribution matching. It effectively preserves local modes, intra-class patterns, and fine-grained variations that characterize the geometry of complex, high-dimensional distributions. Our method comprises three components tailored for preserving distributional geometry: (1) OT-guided diffusion sampling, which aligns latent distributions of real and distilled images; (2) label-image-aligned soft relabeling, which adapts label distributions based on the complexity of distilled image distributions; and (3) OT-based logit matching, which aligns the output of student models with soft-label distributions. Extensive experiments across diverse architectures and large-scale datasets demonstrate that our method consistently outperforms state-of-the-art approaches in an efficient manner, achieving at least 4% accuracy improvement under IPC=10 settings for each architecture on ImageNet-1K.
[181] ART-ASyn: Anatomy-aware Realistic Texture-based Anomaly Synthesis Framework for Chest X-Rays
Qinyi Cao, Jianan Fan, Weidong Cai
Main category: cs.CV
TL;DR: ART-ASyn generates realistic chest X-ray anomalies using texture-based augmentation guided by lung segmentation, enabling supervised anomaly segmentation and zero-shot generalization.
Details
Motivation: Existing synthetic anomaly methods produce visually distinct anomalies that ignore anatomical structure, limiting their realism and effectiveness for medical anomaly detection.Method: Progressive Binary Thresholding Segmentation (PBTSeg) for lung segmentation, then Anatomy-aware Realistic Texture-based Anomaly Synthesis (ART-ASyn) that generates anatomically consistent lung opacity anomalies using texture-based augmentation.
Result: Generates paired synthetic anomalies with precise pixel-level masks for explicit segmentation supervision, enabling both one-class classification and zero-shot anomaly segmentation on unseen datasets.
Conclusion: ART-ASyn produces realistic, anatomically consistent anomalies that enable supervised anomaly segmentation and demonstrates strong zero-shot generalization capabilities for medical imaging.
Abstract: Unsupervised anomaly detection aims to identify anomalies without pixel-level annotations. Synthetic anomaly-based methods exhibit a unique capacity to introduce controllable irregularities with known masks, enabling explicit supervision during training. However, existing methods often produce synthetic anomalies that are visually distinct from real pathological patterns and ignore anatomical structure. This paper presents a novel Anatomy-aware Realistic Texture-based Anomaly Synthesis framework (ART-ASyn) for chest X-rays that generates realistic and anatomically consistent lung opacity related anomalies using texture-based augmentation guided by our proposed Progressive Binary Thresholding Segmentation method (PBTSeg) for lung segmentation. The generated paired samples of synthetic anomalies and their corresponding precise pixel-level anomaly mask for each normal sample enable explicit segmentation supervision. In contrast to prior work limited to one-class classification, ART-ASyn is further evaluated for zero-shot anomaly segmentation, demonstrating generalizability on an unseen dataset without target-domain annotations. Code availability is available at https://github.com/angelacao-hub/ART-ASyn.
[182] Odometry Without Correspondence from Inertially Constrained Ruled Surfaces
Chenqi Zhu, Levi Burner, Yiannis Aloimonos
Main category: cs.CV
TL;DR: Novel visual odometry method using ruled surfaces from line features instead of point correspondences, constrained by IMU data for improved efficiency and accuracy.
Details
Motivation: Traditional visual odometry relies on costly and inaccurate point-to-point correspondences between frames. While some approaches use line features or sensor fusion, they still heavily depend on correspondence. The paper aims to overcome these limitations by exploiting the geometric properties of line features as they move through image-space time.Method: Uses ruled surfaces formed by straight lines sweeping through image-space time as the camera moves. Instead of point correspondences, the method analyzes the shape of these ruled surfaces for odometry information. The approach requires only differentially computed updates from point-to-line associations. The method is inspired by event cameras’ edge detection capabilities and is further constrained by IMU measurements to reduce solution space dimensionality.
Result: The paper presents a novel algorithm that reconstructs 3D scenes and estimates visual odometry from ruled surfaces. By leveraging IMU constraints, the method achieves reduced computational complexity and improved robustness compared to traditional point-based correspondence methods.
Conclusion: The proposed approach offers an efficient alternative to traditional visual odometry by avoiding costly point correspondences and instead using ruled surfaces from line features with IMU constraints, potentially enabling more robust and computationally efficient motion estimation.
Abstract: Visual odometry techniques typically rely on feature extraction from a sequence of images and subsequent computation of optical flow. This point-to-point correspondence between two consecutive frames can be costly to compute and suffers from varying accuracy, which affects the odometry estimate’s quality. Attempts have been made to bypass the difficulties originating from the correspondence problem by adopting line features and fusing other sensors (event camera, IMU) to improve performance, many of which still heavily rely on correspondence. If the camera observes a straight line as it moves, the image of the line sweeps a smooth surface in image-space time. It is a ruled surface and analyzing its shape gives information about odometry. Further, its estimation requires only differentially computed updates from point-to-line associations. Inspired by event cameras’ propensity for edge detection, this research presents a novel algorithm to reconstruct 3D scenes and visual odometry from these ruled surfaces. By constraining the surfaces with the inertia measurements from an onboard IMU sensor, the dimensionality of the solution space is greatly reduced.
[183] Face-MakeUpV2: Facial Consistency Learning for Controllable Text-to-Image Generation
Dawei Dai, Yinxiu Zhou, Chenghang Li, Guolai Jiang, Chengfang Zhang
Main category: cs.CV
TL;DR: Face-MakeUpV2 is a facial image generation model that maintains face ID and physical consistency with reference images while responding to local semantic instructions, addressing attribute leakage and physical inconsistency issues in current text-to-image models.
Details
Motivation: Current text-to-image models for facial generation suffer from facial attribute leakage (unwanted attribute transfer) and insufficient physical consistency when responding to local semantic instructions, limiting their reliability for facial editing applications.Method: 1) Constructed FaceCaptionMask-1M dataset with 1M image-text-mask pairs for precise spatial supervision; 2) Used pretrained text-to-image model as backbone with two facial injection channels: 3D facial rendering for physical characteristics and global facial feature channel; 3) Two optimization objectives: semantic alignment in embedding space to reduce attribute leakage, and perceptual loss on facial images to preserve ID consistency.
Result: Extensive experiments show Face-MakeUpV2 achieves best overall performance in preserving face ID and maintaining physical consistency of reference images compared to other methods.
Conclusion: Face-MakeUpV2 demonstrates practical potential for reliable and controllable facial editing in diverse applications by effectively addressing attribute leakage and physical inconsistency problems.
Abstract: In facial image generation, current text-to-image models often suffer from facial attribute leakage and insufficient physical consistency when responding to local semantic instructions. In this study, we propose Face-MakeUpV2, a facial image generation model that aims to maintain the consistency of face ID and physical characteristics with the reference image. First, we constructed a large-scale dataset FaceCaptionMask-1M comprising approximately one million image-text-masks pairs that provide precise spatial supervision for the local semantic instructions. Second, we employed a general text-to-image pretrained model as the backbone and introduced two complementary facial information injection channels: a 3D facial rendering channel to incorporate the physical characteristics of the image and a global facial feature channel. Third, we formulated two optimization objectives for the supervised learning of our model: semantic alignment in the model’s embedding space to mitigate the attribute leakage problem and perceptual loss on facial images to preserve ID consistency. Extensive experiments demonstrated that our Face-MakeUpV2 achieves best overall performance in terms of preserving face ID and maintaining physical consistency of the reference images. These results highlight the practical potential of Face-MakeUpV2 for reliable and controllable facial editing in diverse applications.
[184] MVAD : A Comprehensive Multimodal Video-Audio Dataset for AIGC Detection
Mengxue Hu, Yunfeng Diao, Changtao Miao, Jianshu Li, Zhe Li, Joey Tianyi Zhou
Main category: cs.CV
TL;DR: MVAD is the first comprehensive multimodal video-audio dataset for detecting AI-generated content, addressing limitations of existing datasets that focus only on visual modality or facial deepfakes.
Details
Motivation: Existing synthetic video datasets focus only on visual modality or are limited to facial deepfakes, which fails to address the expanding landscape of general multimodal AI-generated content and impedes development of trustworthy detection systems.Method: Created MVAD dataset with three key characteristics: genuine multimodality with three realistic video-audio forgery patterns, high perceptual quality using diverse state-of-the-art generative models, and comprehensive diversity across visual styles, content categories, and multimodal data types.
Result: MVAD is introduced as the first comprehensive dataset specifically designed for detecting AI-generated multimodal video-audio content, with samples available at https://github.com/HuMengXue0104/MVAD.
Conclusion: MVAD bridges a critical gap in multimodal AI-generated content detection by providing a comprehensive dataset that addresses limitations of existing resources, enabling development of more robust detection systems.
Abstract: The rapid advancement of AI-generated multimodal video-audio content has raised significant concerns regarding information security and content authenticity. Existing synthetic video datasets predominantly focus on the visual modality alone, while the few incorporating audio are largely confined to facial deepfakes–a limitation that fails to address the expanding landscape of general multimodal AI-generated content and substantially impedes the development of trustworthy detection systems. To bridge this critical gap, we introduce the Multimodal Video-Audio Dataset (MVAD), the first comprehensive dataset specifically designed for detecting AI-generated multimodal video-audio content. Our dataset exhibits three key characteristics: (1) genuine multimodality with samples generated according to three realistic video-audio forgery patterns; (2) high perceptual quality achieved through diverse state-of-the-art generative models; and (3) comprehensive diversity spanning realistic and anime visual styles, four content categories (humans, animals, objects, and scenes), and four video-audio multimodal data types. Our dataset will be available at https://github.com/HuMengXue0104/MVAD.
[185] PANDA – Patch And Distribution-Aware Augmentation for Long-Tailed Exemplar-Free Continual Learning
Siddeshwar Raghavan, Jiangpeng He, Fengqing Zhu
Main category: cs.CV
TL;DR: PANDA is a patch-and-distribution-aware augmentation framework that addresses dual-level data imbalances in exemplar-free continual learning with pre-trained models, improving accuracy and reducing catastrophic forgetting.
Details
Motivation: Real-world data streams for continual learning exhibit dual-level imbalances: dataset-level distributions combined with extreme or reversed skews within individual tasks, creating both intra-task and inter-task disparities that hinder effective learning and generalization in exemplar-free continual learning settings.Method: PANDA uses a CLIP encoder to identify representative regions of low-frequency classes and transplants those patches into frequent-class samples within each task. It also incorporates an adaptive balancing strategy that leverages prior task distributions to smooth inter-task imbalances, reducing the overall gap between average samples across tasks.
Result: Extensive experiments and ablation studies demonstrate PANDA’s capability to work with existing PTM-based continual learning methods, improving accuracy and reducing catastrophic forgetting.
Conclusion: PANDA effectively addresses dual-level data imbalances in exemplar-free continual learning by combining intra-task patch augmentation with inter-task distribution smoothing, enabling fairer learning with frozen pre-trained models.
Abstract: Exemplar-Free Continual Learning (EFCL) restricts the storage of previous task data and is highly susceptible to catastrophic forgetting. While pre-trained models (PTMs) are increasingly leveraged for EFCL, existing methods often overlook the inherent imbalance of real-world data distributions. We discovered that real-world data streams commonly exhibit dual-level imbalances, dataset-level distributions combined with extreme or reversed skews within individual tasks, creating both intra-task and inter-task disparities that hinder effective learning and generalization. To address these challenges, we propose PANDA, a Patch-and-Distribution-Aware Augmentation framework that integrates seamlessly with existing PTM-based EFCL methods. PANDA amplifies low-frequency classes by using a CLIP encoder to identify representative regions and transplanting those into frequent-class samples within each task. Furthermore, PANDA incorporates an adaptive balancing strategy that leverages prior task distributions to smooth inter-task imbalances, reducing the overall gap between average samples across tasks and enabling fairer learning with frozen PTMs. Extensive experiments and ablation studies demonstrate PANDA’s capability to work with existing PTM-based CL methods, improving accuracy and reducing catastrophic forgetting.
[186] Assimilation Matters: Model-level Backdoor Detection in Vision-Language Pretrained Models
Zhongqi Wang, Jie Zhang, Shiguang Shan, Xilin Chen
Main category: cs.CV
TL;DR: AMDET is a novel model-level backdoor detection framework for vision-language pretrained models that operates without prior knowledge of training data, triggers, or targets by exploiting feature assimilation in backdoored text encoders.
Details
Motivation: Existing backdoor detection methods for VLPs like CLIP require impractical prior knowledge (training dataset, trigger patterns, target classes, or downstream classifiers), making them unsuitable for real-world applications where such information is unavailable.Method: AMDET leverages the feature assimilation property where all tokens in backdoored samples show high similarity due to attention concentration on trigger tokens. It performs gradient-based inversion on token embeddings to recover implicit features that activate backdoor behaviors, then filters out natural backdoor features by analyzing loss landscapes.
Result: Extensive experiments on 3,600 backdoored and benign-finetuned models with two attack paradigms and three VLP structures show AMDET achieves 89.90% F1 score, completes detection in ~5 minutes on RTX 4090 GPU, and demonstrates strong robustness against adaptive attacks.
Conclusion: AMDET provides an effective, efficient, and practical solution for backdoor detection in VLPs without requiring prior knowledge, successfully identifying both intentionally injected backdoors and natural backdoor features in official CLIP models.
Abstract: Vision-language pretrained models (VLPs) such as CLIP have achieved remarkable success, but are also highly vulnerable to backdoor attacks. Given a model fine-tuned by an untrusted third party, determining whether the model has been injected with a backdoor is a critical and challenging problem. Existing detection methods usually rely on prior knowledge of training dataset, backdoor triggers and targets, or downstream classifiers, which may be impractical for real-world applications. To address this, To address this challenge, we introduce Assimilation Matters in DETection (AMDET), a novel model-level detection framework that operates without any such prior knowledge. Specifically, we first reveal the feature assimilation property in backdoored text encoders: the representations of all tokens within a backdoor sample exhibit a high similarity. Further analysis attributes this effect to the concentration of attention weights on the trigger token. Leveraging this insight, AMDET scans a model by performing gradient-based inversion on token embeddings to recover implicit features that capable of activating backdoor behaviors. Furthermore, we identify the natural backdoor feature in the OpenAI’s official CLIP model, which are not intentionally injected but still exhibit backdoor-like behaviors. We then filter them out from real injected backdoor by analyzing their loss landscapes. Extensive experiments on 3,600 backdoored and benign-finetuned models with two attack paradigms and three VLP model structures show that AMDET detects backdoors with an F1 score of 89.90%. Besides, it achieves one complete detection in approximately 5 minutes on a RTX 4090 GPU and exhibits strong robustness against adaptive attacks. Code is available at: https://github.com/Robin-WZQ/AMDET
[187] mmPred: Radar-based Human Motion Prediction in the Dark
Junqiao Fan, Haocong Rao, Jiarui Zhang, Jianfei Yang, Lihua Xie
Main category: cs.CV
TL;DR: First diffusion-based framework for radar-based human motion prediction that addresses radar signal artifacts through dual-domain historical motion representation and global skeleton modeling.
Details
Motivation: RGB-D cameras for human motion prediction have limitations: sensitive to lighting conditions and raise privacy concerns. Millimeter-wave radar offers robustness and privacy preservation but suffers from specular reflections and multipath effects causing noisy measurements.Method: Proposes mmPred, a diffusion-based framework with dual-domain historical motion representation: Time-domain Pose Refinement (TPR) for fine-grained details and Frequency-domain Dominant Motion (FDM) for global trends. Uses Global Skeleton-relational Transformer (GST) as diffusion backbone to model global inter-joint cooperation.
Result: Achieves state-of-the-art performance, outperforming existing methods by 8.6% on mmBody dataset and 22% on mm-Fi dataset.
Conclusion: mmPred successfully addresses radar-specific artifacts in human motion prediction, demonstrating the viability of radar as a novel sensing modality with better performance than existing approaches.
Abstract: Existing Human Motion Prediction (HMP) methods based on RGB-D cameras are sensitive to lighting conditions and raise privacy concerns, limiting their real-world applications such as firefighting and healthcare. Motivated by the robustness and privacy-preserving nature of millimeter-wave (mmWave) radar, this work introduces radar as a novel sensing modality for HMP, for the first time. Nevertheless, radar signals often suffer from specular reflections and multipath effects, resulting in noisy and temporally inconsistent measurements, such as body-part miss-detection. To address these radar-specific artifacts, we propose mmPred, the first diffusion-based framework tailored for radar-based HMP. mmPred introduces a dual-domain historical motion representation to guide the generation process, combining a Time-domain Pose Refinement (TPR) branch for learning fine-grained details and a Frequency-domain Dominant Motion (FDM) branch for capturing global motion trends and suppressing frame-level inconsistency. Furthermore, we design a Global Skeleton-relational Transformer (GST) as the diffusion backbone to model global inter-joint cooperation, enabling corrupted joints to dynamically aggregate information from others. Extensive experiments show that mmPred achieves state-of-the-art performance, outperforming existing methods by 8.6% on mmBody and 22% on mm-Fi.
[188] SMamDiff: Spatial Mamba for Stochastic Human Motion Prediction
Junqiao Fan, Pengfei Liu, Haocong Rao
Main category: cs.CV
TL;DR: SMamDiff: A single-stage diffusion model for human motion prediction using Spatial Mamba architecture with residual-DCT encoding and stickman-drawing spatial processing for better spatial-temporal coherence.
Details
Motivation: Existing human motion prediction methods either produce deterministic forecasts ignoring uncertainty or use probabilistic models sacrificing kinematic plausibility. Diffusion models improve accuracy-diversity trade-off but rely on multi-stage pipelines that are costly for edge deployment. Need to ensure spatial-temporal coherence within a single-stage diffusion model.Method: Introduces SMamDiff with two novel designs: 1) residual-DCT motion encoding that subtracts last observed pose before temporal DCT to reduce DC component dominance and highlight higher-frequency cues, 2) stickman-drawing spatial-mamba module that processes joints in ordered, joint-by-joint manner to induce long-range cross-joint dependencies.
Result: Achieves state-of-the-art results among single-stage probabilistic HMP methods on Human3.6M and HumanEva datasets, while using less latency and memory than multi-stage diffusion baselines.
Conclusion: The proposed coherence mechanisms in SMamDiff enable effective spatial-temporal modeling for human motion prediction within a single-stage diffusion framework, making it suitable for edge deployment with improved efficiency and performance.
Abstract: With intelligent room-side sensing and service robots widely deployed, human motion prediction (HMP) is essential for safe, proactive assistance. However, many existing HMP methods either produce a single, deterministic forecast that ignores uncertainty or rely on probabilistic models that sacrifice kinematic plausibility. Diffusion models improve the accuracy-diversity trade-off but often depend on multi-stage pipelines that are costly for edge deployment. This work focuses on how to ensure spatial-temporal coherence within a single-stage diffusion model for HMP. We introduce SMamDiff, a Spatial Mamba-based Diffusion model with two novel designs: (i) a residual-DCT motion encoding that subtracts the last observed pose before a temporal DCT, reducing the first DC component ($f=0$) dominance and highlighting informative higher-frequency cues so the model learns how joints move rather than where they are; and (ii) a stickman-drawing spatial-mamba module that processes joints in an ordered, joint-by-joint manner, making later joints condition on earlier ones to induce long-range, cross-joint dependencies. On Human3.6M and HumanEva, these coherence mechanisms deliver state-of-the-art results among single-stage probabilistic HMP methods while using less latency and memory than multi-stage diffusion baselines.
[189] MM-DETR: An Efficient Multimodal Detection Transformer with Mamba-Driven Dual-Granularity Fusion and Frequency-Aware Modality Adapters
Jianhong Han, Yupei Wang, Yuan Zhang, Liang Chen
Main category: cs.CV
TL;DR: MM-DETR: A lightweight multimodal object detection framework using Mamba-based fusion with linear complexity, modality completion via 2D selective scanning, and frequency-aware adapters for efficient deployment.
Details
Motivation: Existing multimodal object detection methods struggle to balance performance and lightweight design. Attention-based or deformable convolution fusion blocks are complex, while shared backbones yield suboptimal representations and dual-stream architectures double parameters, limiting practical deployment.Method: 1) Mamba-based dual granularity fusion encoder reformulates global interaction as channel-wise dynamic gating with 1D selective scan for efficient cross-modal modeling (linear complexity). 2) Reinterprets fusion as modality completion problem with region-aware 2D selective scanning completion branch for fine-grained fusion. 3) Lightweight frequency-aware modality adapter with spatial-frequency co-expert structure and pixel-wise router for efficient feature extraction in shared backbone.
Result: Extensive experiments on four multimodal benchmark datasets demonstrate effectiveness and generalization capability of the proposed method.
Conclusion: MM-DETR provides a lightweight and efficient framework for multimodal object detection that balances performance with practical deployment requirements through innovative fusion strategies and parameter-efficient design.
Abstract: Multimodal remote sensing object detection aims to achieve more accurate and robust perception under challenging conditions by fusing complementary information from different modalities. However, existing approaches that rely on attention-based or deformable convolution fusion blocks still struggle to balance performance and lightweight design. Beyond fusion complexity, extracting modality features with shared backbones yields suboptimal representations due to insufficient modality-specific modeling, whereas dual-stream architectures nearly double the parameter count, ultimately limiting practical deployment. To this end, we propose MM-DETR, a lightweight and efficient framework for multimodal object detection. Specifically, we propose a Mamba-based dual granularity fusion encoder that reformulates global interaction as channel-wise dynamic gating and leverages a 1D selective scan for efficient cross-modal modeling with linear complexity. Following this design, we further reinterpret multimodal fusion as a modality completion problem. A region-aware 2D selective scanning completion branch is introduced to recover modality-specific cues, supporting fine-grained fusion along a bidirectional pyramid pathway with minimal overhead. To further reduce parameter redundancy while retaining strong feature extraction capability, a lightweight frequency-aware modality adapter is inserted into the shared backbone. This adapter employs a spatial-frequency co-expert structure to capture modality-specific cues, while a pixel-wise router dynamically balances expert contributions for efficient spatial-frequency fusion. Extensive experiments conducted on four multimodal benchmark datasets demonstrate the effectiveness and generalization capability of the proposed method.
[190] Towards aligned body representations in vision models
Andrey Gizdov, Andrea Procopio, Yichen Li, Daniel Harari, Tomer Ullman
Main category: cs.CV
TL;DR: Smaller segmentation models develop human-like coarse body representations, while larger models create overly detailed encodings, suggesting coarse representations emerge under limited computational resources.
Details
Motivation: To understand whether computer vision models trained for segmentation develop similar coarse body representations as humans use for physical reasoning, and to investigate how model size affects these representations.Method: Adapted a psychophysical experiment with 50 human participants to a semantic segmentation task, then tested a family of seven segmentation networks of varying sizes to analyze their internal representations.
Result: Smaller segmentation models naturally form human-like coarse body representations, while larger models tend to develop overly detailed, fine-grained encodings that deviate from human representations.
Conclusion: Coarse body representations emerge naturally under limited computational resources, and machine learning models provide a scalable approach to understanding the structure of physical reasoning in the human brain.
Abstract: Human physical reasoning relies on internal “body” representations - coarse, volumetric approximations that capture an object’s extent and support intuitive predictions about motion and physics. While psychophysical evidence suggests humans use such coarse representations, their internal structure remains largely unknown. Here we test whether vision models trained for segmentation develop comparable representations. We adapt a psychophysical experiment conducted with 50 human participants to a semantic segmentation task and test a family of seven segmentation networks, varying in size. We find that smaller models naturally form human-like coarse body representations, whereas larger models tend toward overly detailed, fine-grain encodings. Our results demonstrate that coarse representations can emerge under limited computational resources, and that machine representations can provide a scalable path toward understanding the structure of physical reasoning in the brain.
[191] THCRL: Trusted Hierarchical Contrastive Representation Learning for Multi-View Clustering
Jian Zhu
Main category: cs.CV
TL;DR: THCRL is a novel deep multi-view clustering method that addresses untrustworthy fusion through hierarchical contrastive learning with denoising mechanisms and neighbor-aware similarity computation.
Details
Motivation: Existing multi-view clustering methods suffer from untrustworthy fusion due to two main issues: 1) ignoring inherent noise within individual views, and 2) traditional contrastive learning approaches that only consider same-instance similarities across views while neglecting structural information from nearest neighbors within clusters.Method: THCRL consists of two key modules: 1) Deep Symmetry Hierarchical Fusion (DSHF) module using UNet architecture with multiple denoising mechanisms for trustworthy multi-view fusion, and 2) Average K-Nearest Neighbors Contrastive Learning (AKCL) module that aligns fused representations with view-specific representations by enhancing similarity among samples in the same cluster rather than just same samples across views.
Result: Extensive experiments demonstrate that THCRL achieves state-of-the-art performance in deep multi-view clustering tasks.
Conclusion: THCRL effectively addresses the untrustworthy fusion problem in multi-view clustering through hierarchical contrastive representation learning with denoising and neighbor-aware similarity computation, leading to superior clustering performance.
Abstract: Multi-View Clustering (MVC) has garnered increasing attention in recent years. It is capable of partitioning data samples into distinct groups by learning a consensus representation. However, a significant challenge remains: the problem of untrustworthy fusion. This problem primarily arises from two key factors: 1) Existing methods often ignore the presence of inherent noise within individual views; 2) In traditional MVC methods using Contrastive Learning (CL), similarity computations typically rely on different views of the same instance, while neglecting the structural information from nearest neighbors within the same cluster. Consequently, this leads to the wrong direction for multi-view fusion. To address this problem, we present a novel Trusted Hierarchical Contrastive Representation Learning (THCRL). It consists of two key modules. Specifically, we propose the Deep Symmetry Hierarchical Fusion (DSHF) module, which leverages the UNet architecture integrated with multiple denoising mechanisms to achieve trustworthy fusion of multi-view data. Furthermore, we present the Average K-Nearest Neighbors Contrastive Learning (AKCL) module to align the fused representation with the view-specific representation. Unlike conventional strategies, AKCL enhances representation similarity among samples belonging to the same cluster, rather than merely focusing on the same sample across views, thereby reinforcing the confidence of the fused representation. Extensive experiments demonstrate that THCRL achieves the state-of-the-art performance in deep MVC tasks.
[192] POLARIS: Projection-Orthogonal Least Squares for Robust and Adaptive Inversion in Diffusion Models
Wenshuo Chen, Haosen Li, Shaofeng Liang, Lei Wang, Haozhe Jia, Kaishen Yuan, Jieming Wu, Bowen Tian, Yutao Yue
Main category: cs.CV
TL;DR: POLARIS addresses reconstruction degradation in diffusion model inversion by minimizing noise approximation errors through step-wise optimization of guidance scale ω, improving inversion quality with minimal code changes.
Details
Motivation: The paper identifies that current Inversion-Denoising Paradigm suffers from reconstruction degradation due to approximate noise error - the error from approximating noise at step t with prediction at step t-1, which accumulates throughout inversion and harms downstream tasks.Method: POLARIS (Projection-Orthogonal Least Squares for Robust and Adaptive Inversion) reformulates inversion as an error-origin problem rather than error-compensation. It treats guidance scale ω as a step-wise variable and derives a mathematically grounded formula to minimize inversion error at each step, requiring just one line of code.
Result: POLARIS substantially mitigates noise approximation errors, improves inversion latent quality with negligible performance overhead, and consistently enhances accuracy of downstream tasks in image editing and restoration.
Conclusion: POLARIS provides an efficient, mathematically sound solution to the critical problem of noise approximation error accumulation in diffusion model inversion, offering significant improvements with minimal implementation complexity.
Abstract: The Inversion-Denoising Paradigm, which is based on diffusion models, excels in diverse image editing and restoration tasks. We revisit its mechanism and reveal a critical, overlooked factor in reconstruction degradation: the approximate noise error. This error stems from approximating the noise at step t with the prediction at step t-1, resulting in severe error accumulation throughout the inversion process. We introduce Projection-Orthogonal Least Squares for Robust and Adaptive Inversion (POLARIS), which reformulates inversion from an error-compensation problem into an error-origin problem. Rather than optimizing embeddings or latent codes to offset accumulated drift, POLARIS treats the guidance scale ω as a step-wise variable and derives a mathematically grounded formula to minimize inversion error at each step. Remarkably, POLARIS improves inversion latent quality with just one line of code. With negligible performance overhead, it substantially mitigates noise approximation errors and consistently improves the accuracy of downstream tasks.
[193] Pore-scale Image Patch Dataset and A Comparative Evaluation of Pore-scale Facial Features
Dong Li, HuaLiang Lin, JiaYu Li
Main category: cs.CV
TL;DR: The paper introduces PorePatch dataset for pore-scale facial image patches and shows deep learning descriptors outperform traditional ones in matching tasks but not significantly in 3D reconstruction.
Details
Motivation: Facial skin regions have weak texture, making local descriptor matching challenging for facial motion analysis and 3D reconstruction. Deep learning descriptors perform well in other domains but lack pore-scale facial datasets for development.Method: Proposed PorePatch dataset with Data-Model Co-Evolution (DMCE) framework to generate high-quality pore-scale image patches from high-resolution facial images. Trained SOTA models on this dataset and evaluated on matching and 3D reconstruction tasks.
Result: SOTA model achieved FPR95 of 1.91% on matching task, outperforming PSIFT (22.41%) by 20.5%. However, in 3D reconstruction, deep learning descriptors didn’t significantly outperform traditional descriptors, showing limitations in facial weak-texture regions.
Conclusion: Deep learning descriptors still have limitations in addressing facial weak-texture challenges despite good matching performance. More work needed in this field, and the PorePatch dataset provides a valuable benchmark for future research.
Abstract: The weak-texture nature of facial skin regions presents significant challenges for local descriptor matching in applications such as facial motion analysis and 3D face reconstruction. Although deep learning-based descriptors have demonstrated superior performance to traditional hand-crafted descriptors in many applications, the scarcity of pore-scale image patch datasets has hindered their further development in the facial domain. In this paper, we propose the PorePatch dataset, a high-quality pore-scale image patch dataset, and establish a rational evaluation benchmark. We introduce a Data-Model Co-Evolution (DMCE) framework to generate a progressively refined, high-quality dataset from high-resolution facial images. We then train existing SOTA models on our dataset and conduct extensive experiments. Our results show that the SOTA model achieves a FPR95 value of 1.91% on the matching task, outperforming PSIFT (22.41%) by a margin of 20.5%. However, its advantage is diminished in the 3D reconstruction task, where its overall performance is not significantly better than that of traditional descriptors. This indicates that deep learning descriptors still have limitations in addressing the challenges of facial weak-texture regions, and much work remains to be done in this field.
[194] EZ-SP: Fast and Lightweight Superpoint-Based 3D Segmentation
Louis Geist, Loic Landrieu, Damien Robert
Main category: cs.CV
TL;DR: EZ-SP: A GPU-accelerated superpoint partitioning algorithm that’s 13× faster than prior methods, enabling efficient 3D semantic segmentation with real-time inference and minimal memory footprint.
Details
Motivation: Superpoint-based 3D semantic segmentation pipelines are often bottlenecked by CPU-bound partition steps, limiting their efficiency and real-time applicability. Existing methods require handcrafted features and are too slow for practical deployment.Method: Proposes a learnable, fully GPU partitioning algorithm that generates geometrically and semantically coherent superpoints. The module is compact (<60k parameters), trains with a differentiable surrogate loss in under 20 minutes, and requires no handcrafted features. Combined with a lightweight superpoint classifier.
Result: Achieves 13× faster partitioning than prior methods, with 72× faster inference and 120× fewer parameters. The full pipeline fits in <2 MB VRAM, scales to multi-million-point scenes, and supports real-time inference. Matches accuracy of point-based SOTA models across indoor scans (S3DIS), autonomous driving (KITTI-360), and aerial LiDAR (DALES).
Conclusion: EZ-SP provides an efficient, GPU-accelerated superpoint partitioning solution that enables real-time 3D semantic segmentation with minimal memory footprint while maintaining state-of-the-art accuracy across diverse domains.
Abstract: Superpoint-based pipelines provide an efficient alternative to point- or voxel-based 3D semantic segmentation, but are often bottlenecked by their CPU-bound partition step. We propose a learnable, fully GPU partitioning algorithm that generates geometrically and semantically coherent superpoints 13$\times$ faster than prior methods. Our module is compact (under 60k parameters), trains in under 20 minutes with a differentiable surrogate loss, and requires no handcrafted features. Combine with a lightweight superpoint classifier, the full pipeline fits in $<$2 MB of VRAM, scales to multi-million-point scenes, and supports real-time inference. With 72$\times$ faster inference and 120$\times$ fewer parameters, EZ-SP matches the accuracy of point-based SOTA models across three domains: indoor scans (S3DIS), autonomous driving (KITTI-360), and aerial LiDAR (DALES). Code and pretrained models are accessible at github.com/drprojects/superpoint_transformer.
[195] WiseEdit: Benchmarking Cognition- and Creativity-Informed Image Editing
Kaihang Pan, Weile Chen, Haiyi Qiu, Qifan Yu, Wendong Bu, Zehan Wang, Yun Zhu, Juncheng Li, Siliang Tang
Main category: cs.CV
TL;DR: WiseEdit is a knowledge-intensive benchmark for evaluating cognition- and creativity-informed image editing models, featuring 1,220 test cases across three cognitive steps and three knowledge types.
Details
Motivation: Existing benchmarks are too narrow to holistically assess the advanced cognition and creativity capabilities of modern image editing models, which require comprehensive evaluation of knowledge-intensive editing tasks.Method: WiseEdit decomposes image editing into three cascaded cognitive steps (Awareness, Interpretation, Imagination) and incorporates three knowledge types (Declarative, Procedural, Metacognitive). It includes both step-specific tasks and complex tasks that combine multiple steps.
Result: The benchmark comprises 1,220 test cases that objectively reveal limitations of state-of-the-art image editing models in knowledge-based cognitive reasoning and creative composition capabilities.
Conclusion: WiseEdit provides a comprehensive evaluation framework for advanced image editing models, addressing the gap in existing benchmarks and enabling better assessment of cognition- and creativity-informed editing capabilities.
Abstract: Recent image editing models boast next-level intelligent capabilities, facilitating cognition- and creativity-informed image editing. Yet, existing benchmarks provide too narrow a scope for evaluation, failing to holistically assess these advanced abilities. To address this, we introduce WiseEdit, a knowledge-intensive benchmark for comprehensive evaluation of cognition- and creativity-informed image editing, featuring deep task depth and broad knowledge breadth. Drawing an analogy to human cognitive creation, WiseEdit decomposes image editing into three cascaded steps, i.e., Awareness, Interpretation, and Imagination, each corresponding to a task that poses a challenge for models to complete at the specific step. It also encompasses complex tasks, where none of the three steps can be finished easily. Furthermore, WiseEdit incorporates three fundamental types of knowledge: Declarative, Procedural, and Metacognitive knowledge. Ultimately, WiseEdit comprises 1,220 test cases, objectively revealing the limitations of SoTA image editing models in knowledge-based cognitive reasoning and creative composition capabilities. The benchmark, evaluation code, and the generated images of each model will be made publicly available soon. Project Page: https://qnancy.github.io/wiseedit_project_page/.
[196] Better, Stronger, Faster: Tackling the Trilemma in MLLM-based Segmentation with Simultaneous Textual Mask Prediction
Jiazhen Liu, Mingkuan Feng, Long Chen
Main category: cs.CV
TL;DR: STAMP resolves the trilemma in MLLMs by decoupling dialogue generation from segmentation via all-mask prediction, achieving simultaneous excellence in dialogue preservation, segmentation performance, and inference speed.
Details
Motivation: Current MLLMs face a trilemma: they must compromise between preserving dialogue ability, achieving high segmentation performance, and ensuring fast inference. Existing methods either degrade dialogue with pixel-level objectives or suffer from poor segmentation/speed trade-offs in autoregressive approaches.Method: STAMP uses all-mask prediction paradigm that decouples autoregressive dialogue generation from non-autoregressive mask prediction. After generating textual response, it predicts entire segmentation mask in single forward pass by treating it as parallel “fill-in-the-blank” task over image patches.
Result: STAMP significantly outperforms state-of-the-art methods across multiple segmentation benchmarks, providing solution that excels in dialogue, segmentation, and speed without compromise.
Conclusion: The all-mask prediction paradigm successfully resolves the core trilemma in MLLMs, enabling simultaneous preservation of dialogue ability, high segmentation performance, and fast inference through decoupled architecture design.
Abstract: Integrating segmentation into Multimodal Large Language Models (MLLMs) presents a core trilemma: simultaneously preserving dialogue ability, achieving high segmentation performance, and ensuring fast inference. Prevailing paradigms are forced into a compromise. Embedding prediction methods introduce a conflicting pixel-level objective that degrades the MLLM’s general dialogue abilities. The alternative, next-token prediction, reframes segmentation as an autoregressive task, which preserves dialogue but forces a trade-off between poor segmentation performance with sparse outputs or prohibitive inference speeds with rich ones. We resolve this trilemma with all-mask prediction, a novel paradigm that decouples autoregressive dialogue generation from non-autoregressive mask prediction. We present STAMP: Simultaneous Textual All-Mask Prediction, an MLLM that embodies this paradigm. After generating a textual response, STAMP predicts an entire segmentation mask in a single forward pass by treating it as a parallel “fill-in-the-blank” task over image patches. This design maintains the MLLM’s dialogue ability by avoiding conflicting objectives, enables high segmentation performance by leveraging rich, bidirectional spatial context for all mask tokens, and achieves exceptional speed. Extensive experiments show that STAMP significantly outperforms state-of-the-art methods across multiple segmentation benchmarks, providing a solution that excels in dialogue, segmentation, and speed without compromise.
[197] Low-Bitrate Video Compression through Semantic-Conditioned Diffusion
Lingdong Wang, Guan-Ming Su, Divya Kothandaraman, Tsung-Wei Huang, Mohammad Hajiesmaili, Ramesh K. Sitaraman
Main category: cs.CV
TL;DR: DiSCo is a semantic video compression framework that transmits only key semantic, appearance, and motion information, then uses a conditional video diffusion model to reconstruct high-quality videos from these compact representations, achieving 2-10X better perceptual quality at low bitrates.
Details
Motivation: Traditional video codecs fail at ultra-low bitrates because they optimize for pixel fidelity rather than human perception, leading to severe artifacts. There's a fundamental misalignment between pixel accuracy and what humans actually perceive as high quality.Method: Decomposes video into three compact modalities: textual descriptions (semantic), spatiotemporally degraded video (appearance), and optional sketches/poses (motion cues). Uses a conditional video diffusion model to reconstruct high-quality, temporally coherent videos from these representations. Includes temporal forward filling, token interleaving, and modality-specific codecs to improve multimodal generation and compactness.
Result: Outperforms baseline semantic and traditional codecs by 2-10X on perceptual metrics at low bitrates. The method successfully maintains video quality even at ultra-low bitrates where traditional codecs collapse.
Conclusion: Semantic video compression that focuses on transmitting meaningful information rather than pixel accuracy, combined with generative priors for detail synthesis, provides a superior approach for ultra-low bitrate video compression that aligns better with human perception.
Abstract: Traditional video codecs optimized for pixel fidelity collapse at ultra-low bitrates and produce severe artifacts. This failure arises from a fundamental misalignment between pixel accuracy and human perception. We propose a semantic video compression framework named DiSCo that transmits only the most meaningful information while relying on generative priors for detail synthesis. The source video is decomposed into three compact modalities: a textual description, a spatiotemporally degraded video, and optional sketches or poses that respectively capture semantic, appearance, and motion cues. A conditional video diffusion model then reconstructs high-quality, temporally coherent videos from these compact representations. Temporal forward filling, token interleaving, and modality-specific codecs are proposed to improve multimodal generation and modality compactness. Experiments show that our method outperforms baseline semantic and traditional codecs by 2-10X on perceptual metrics at low bitrates.
[198] SplatFont3D: Structure-Aware Text-to-3D Artistic Font Generation with Part-Level Style Control
Ji Gan, Lingxu Chen, Jiaxu Leng, Xinbo Gao
Main category: cs.CV
TL;DR: SplatFont3D: A novel text-to-3D artistic font generation framework using 3D Gaussian splatting for precise part-level style control and efficient rendering.
Details
Motivation: 3D artistic font generation (3D-AFG) is underexplored despite its potential applications in immersive 3D environments and enhancement of 2D-AFG. 3D fonts have unique challenges: precise semantics with strong structural constraints and need for fine-grained part-level style control.Method: Proposes SplatFont3D with Glyph2Cloud module that enhances 2D glyphs to produce 3D point clouds for Gaussian initialization. Uses score distillation sampling with pretrained 2D diffusion model for optimization. Implements dynamic component assignment strategy using geometric priors of 3D Gaussians for part-level control while preventing drift-induced entanglement.
Result: Outperforms existing 3D models for 3D-AFG in style-text consistency, visual quality, and rendering efficiency. Provides more explicit and effective part-level style control than NeRF with faster rendering.
Conclusion: SplatFont3D successfully addresses the challenges of 3D-AFG by enabling precise part-level style control through 3D Gaussian splatting, offering superior performance in style consistency, quality, and efficiency compared to existing methods.
Abstract: Artistic font generation (AFG) can assist human designers in creating innovative artistic fonts. However, most previous studies primarily focus on 2D artistic fonts in flat design, leaving personalized 3D-AFG largely underexplored. 3D-AFG not only enables applications in immersive 3D environments such as video games and animations, but also may enhance 2D-AFG by rendering 2D fonts of novel views. Moreover, unlike general 3D objects, 3D fonts exhibit precise semantics with strong structural constraints and also demand fine-grained part-level style control. To address these challenges, we propose SplatFont3D, a novel structure-aware text-to-3D AFG framework with 3D Gaussian splatting, which enables the creation of 3D artistic fonts from diverse style text prompts with precise part-level style control. Specifically, we first introduce a Glyph2Cloud module, which progressively enhances both the shapes and styles of 2D glyphs (or components) and produces their corresponding 3D point clouds for Gaussian initialization. The initialized 3D Gaussians are further optimized through interaction with a pretrained 2D diffusion model using score distillation sampling. To enable part-level control, we present a dynamic component assignment strategy that exploits the geometric priors of 3D Gaussians to partition components, while alleviating drift-induced entanglement during 3D Gaussian optimization. Our SplatFont3D provides more explicit and effective part-level style control than NeRF, attaining faster rendering efficiency. Experiments show that our SplatFont3D outperforms existing 3D models for 3D-AFG in style-text consistency, visual quality, and rendering efficiency.
[199] PhysGen: Physically Grounded 3D Shape Generation for Industrial Design
Yingxuan You, Chen Zhao, Hantao Zhang, Mingda Xu, Pascal Fua
Main category: cs.CV
TL;DR: A physics-based 3D shape generation pipeline that incorporates physical properties (like aerodynamic efficiency) into generative models to enhance shape realism beyond visual plausibility.
Details
Motivation: Existing 3D shape generation models lack knowledge of physical properties that are crucial for engineering-designed shapes (e.g., aerodynamic efficiency for automobiles), limiting their ability to produce physically realistic shapes.Method: Proposes a unified physics-based pipeline with: 1) A flow matching model with explicit physical guidance using alternating velocity-based updates and physics-based refinement, 2) Physics-aware regularization, and 3) A shape-and-physics variational autoencoder (SP-VAE) that jointly encodes shape and physics into a unified latent space.
Result: Experiments on three benchmarks show the synergistic formulation improves shape realism beyond mere visual plausibility by incorporating physical validity.
Conclusion: The proposed physics-guided 3D shape generation approach successfully integrates physical properties into the generative process, enhancing realism for industrial design applications where physical performance matters.
Abstract: Existing generative models for 3D shapes can synthesize high-fidelity and visually plausible shapes. For certain classes of shapes that have undergone an engineering design process, the realism of the shape is tightly coupled with the underlying physical properties, e.g., aerodynamic efficiency for automobiles. Since existing methods lack knowledge of such physics, they are unable to use this knowledge to enhance the realism of shape generation. Motivated by this, we propose a unified physics-based 3D shape generation pipeline, with a focus on industrial design applications. Specifically, we introduce a new flow matching model with explicit physical guidance, consisting of an alternating update process. We iteratively perform a velocity-based update and a physics-based refinement, progressively adjusting the latent code to align with the desired 3D shapes and physical properties. We further strengthen physical validity by incorporating a physics-aware regularization term into the velocity-based update step. To support such physics-guided updates, we build a shape-and-physics variational autoencoder (SP-VAE) that jointly encodes shape and physics information into a unified latent space. The experiments on three benchmarks show that this synergistic formulation improves shape realism beyond mere visual plausibility.
[200] Recovering Origin Destination Flows from Bus CCTV: Early Results from Nairobi and Kigali
Nthenya Kyatha, Jay Taneja
Main category: cs.CV
TL;DR: A computer vision pipeline using YOLOv12, BotSORT, OSNet, OCR, and telematics achieves high passenger counting accuracy in ideal conditions but degrades significantly under real-world stressors like overcrowding in African public transport.
Details
Motivation: Public transport in sub-Saharan Africa operates in overcrowded conditions where existing automated systems fail to capture reliable passenger flow data, creating a need for alternative solutions using existing infrastructure.Method: A baseline pipeline combining YOLOv12 detection, BotSORT tracking, OSNet embeddings, OCR-based timestamping, and telematics-based stop classification to recover bus origin-destination flows from onboard CCTV footage.
Result: High counting accuracy under low-density, well-lit conditions (recall ≈95%, precision ≈91%, F1 ≈93%) with OD matrices matching manual tallies, but sharp performance degradation under overcrowding, color-to-monochrome shifts, posture variation, and non-standard door use.
Conclusion: The system reveals deployment-specific failure modes in SSA transit, motivating the need for more robust, deployment-focused Re-ID methods that can handle real-world stressors like overcrowding and challenging environmental conditions.
Abstract: Public transport in sub-Saharan Africa (SSA) often operates in overcrowded conditions where existing automated systems fail to capture reliable passenger flow data. Leveraging onboard CCTV already deployed for security, we present a baseline pipeline that combines YOLOv12 detection, BotSORT tracking, OSNet embeddings, OCR-based timestamping, and telematics-based stop classification to recover bus origin–destination (OD) flows. On annotated CCTV segments from Nairobi and Kigali buses, the system attains high counting accuracy under low-density, well-lit conditions (recall $\approx$95%, precision $\approx$91%, F1 $\approx$93%). It produces OD matrices that closely match manual tallies. Under realistic stressors such as overcrowding, color-to-monochrome shifts, posture variation, and non-standard door use, performance degrades sharply (e.g., $\sim$40% undercount in peak-hour boarding and a $\sim$17 percentage-point drop in recall for monochrome segments), revealing deployment-specific failure modes and motivating more robust, deployment-focused Re-ID methods for SSA transit.
[201] What about gravity in video generation? Post-Training Newton’s Laws with Verifiable Rewards
Minh-Quan Le, Yuanzhi Zhu, Vicky Kalogeiton, Dimitris Samaras
Main category: cs.CV
TL;DR: NewtonRewards is a physics-grounded post-training framework that uses verifiable rewards (optical flow for velocity, appearance features for mass) to enforce Newtonian constraints in video generation, improving physical plausibility without human/VLM feedback.
Details
Motivation: Current video diffusion models produce visually compelling clips but often violate basic physical laws (objects float, inconsistent accelerations, unrealistic collisions), creating a gap between visual realism and physical realism.Method: Extracts measurable proxies from generated videos using frozen utility models: optical flow as velocity proxy, appearance features as mass proxy. Enforces Newtonian structure through two rewards: Newtonian kinematic constraint (constant-acceleration dynamics) and mass conservation reward (prevents degenerate solutions).
Result: Evaluated on NewtonBench-60K benchmark with five Newtonian Motion Primitives. Consistently improves physical plausibility, motion smoothness, and temporal coherence over prior methods. Maintains strong performance under out-of-distribution shifts in height, speed, and friction.
Conclusion: Physics-grounded verifiable rewards offer a scalable path toward physics-aware video generation, bridging the gap between visual and physical realism without relying on human or VLM feedback.
Abstract: Recent video diffusion models can synthesize visually compelling clips, yet often violate basic physical laws-objects float, accelerations drift, and collisions behave inconsistently-revealing a persistent gap between visual realism and physical realism. We propose $\texttt{NewtonRewards}$, the first physics-grounded post-training framework for video generation based on $\textit{verifiable rewards}$. Instead of relying on human or VLM feedback, $\texttt{NewtonRewards}$ extracts $\textit{measurable proxies}$ from generated videos using frozen utility models: optical flow serves as a proxy for velocity, while high-level appearance features serve as a proxy for mass. These proxies enable explicit enforcement of Newtonian structure through two complementary rewards: a Newtonian kinematic constraint enforcing constant-acceleration dynamics, and a mass conservation reward preventing trivial, degenerate solutions. We evaluate $\texttt{NewtonRewards}$ on five Newtonian Motion Primitives (free fall, horizontal/parabolic throw, and ramp sliding down/up) using our newly constructed large-scale benchmark, $\texttt{NewtonBench-60K}$. Across all primitives in visual and physics metrics, $\texttt{NewtonRewards}$ consistently improves physical plausibility, motion smoothness, and temporal coherence over prior post-training methods. It further maintains strong performance under out-of-distribution shifts in height, speed, and friction. Our results show that physics-grounded verifiable rewards offer a scalable path toward physics-aware video generation.
[202] Recognizing Pneumonia in Real-World Chest X-rays with a Classifier Trained with Images Synthetically Generated by Nano Banana
Jiachuan Peng, Kyle Lam, Jianing Qiu
Main category: cs.CV
TL;DR: A classifier trained only on synthetic chest X-ray images generated by Google’s Nano Banana AI model achieved strong performance on real-world pneumonia detection, demonstrating feasibility of synthetic data for medical AI development.
Details
Motivation: To explore whether synthetic medical imaging data generated by advanced AI models can be effectively used to train classifiers for real-world medical diagnosis, potentially addressing data scarcity and privacy concerns in medical AI development.Method: Trained a pneumonia classifier exclusively using synthetic chest X-ray images generated by Google’s Nano Banana AI model, then evaluated it on two real-world datasets (RSNA Pneumonia Detection dataset with 14,863 CXRs and Chest X-Ray dataset with 5,856 CXRs) without any real-world training data.
Result: The synthetic-data-only classifier achieved AUROC of 0.923 and AUPR of 0.900 on RSNA dataset, and AUROC of 0.824 and AUPR of 0.913 on Chest X-Ray dataset, demonstrating strong performance on real-world pneumonia detection despite never seeing real training data.
Conclusion: Synthetic medical imaging data shows promise for medical AI development, but faces challenges in prompt design diversity control, post-processing alignment, and requires substantial validation, regulatory approval, and ethical oversight before clinical use.
Abstract: We trained a classifier with synthetic chest X-ray (CXR) images generated by Nano Banana, the latest AI model for image generation and editing, released by Google. When directly applied to real-world CXRs having only been trained with synthetic data, the classifier achieved an AUROC of 0.923 (95% CI: 0.919 - 0.927), and an AUPR of 0.900 (95% CI: 0.894 - 0.907) in recognizing pneumonia in the 2018 RSNA Pneumonia Detection dataset (14,863 CXRs), and an AUROC of 0.824 (95% CI: 0.810 - 0.836), and an AUPR of 0.913 (95% CI: 0.904 - 0.922) in the Chest X-Ray dataset (5,856 CXRs). These external validation results on real-world data demonstrate the feasibility of this approach and suggest potential for synthetic data in medical AI development. Nonetheless, several limitations remain at present, including challenges in prompt design for controlling the diversity of synthetic CXR data and the requirement for post-processing to ensure alignment with real-world data. However, the growing sophistication and accessibility of medical intelligence will necessitate substantial validation, regulatory approval, and ethical oversight prior to clinical translation.
[203] FR-TTS: Test-Time Scaling for NTP-based Image Generation with Effective Filling-based Reward Signal
Hang Xu, Linjiang Huang, Feng Zhao
Main category: cs.CV
TL;DR: FR-TTS introduces a filling-based reward method to enable effective test-time scaling for next-token prediction by estimating future trajectories of incomplete sequences, overcoming the low correlation problem between intermediate and final rewards.
Details
Motivation: Test-time scaling works well for image generation but fails in next-token prediction due to low correlation between rewards of intermediate token sequences and fully generated images, making intermediate representations poor indicators for pruning decisions.Method: Proposes Filling-Based Reward (FR) that estimates approximate future trajectories by finding and applying reasonable filling schemes to complete intermediate sequences. Builds FR-TTS strategy that searches for good filling schemes and incorporates diversity reward with dynamic weighting for balanced evaluation.
Result: FR provides excellent and reliable metric for evaluating intermediate sample quality, with high correlation between intermediate and final rewards. FR-TTS outperforms multiple established benchmarks and various reward models in experimental validation.
Conclusion: FR-TTS successfully enables effective test-time scaling for next-token prediction by addressing the critical issue of low correlation between intermediate and final rewards through intelligent sequence filling and comprehensive evaluation strategies.
Abstract: Test-time scaling (TTS) has become a prevalent technique in image generation, significantly boosting output quality by expanding the number of parallel samples and filtering them using pre-trained reward models. However, applying this powerful methodology to the next-token prediction (NTP) paradigm remains challenging. The primary obstacle is the low correlation between the reward of an image decoded from an intermediate token sequence and the reward of the fully generated image. Consequently, these incomplete intermediate representations prove to be poor indicators for guiding the pruning direction, a limitation that stems from their inherent incompleteness in scale or semantic content. To effectively address this critical issue, we introduce the Filling-Based Reward (FR). This novel design estimates the approximate future trajectory of an intermediate sample by finding and applying a reasonable filling scheme to complete the sequence. Both the correlation coefficient between rewards of intermediate samples and final samples, as well as multiple intrinsic signals like token confidence, indicate that the FR provides an excellent and reliable metric for accurately evaluating the quality of intermediate samples. Building upon this foundation, we propose FR-TTS, a sophisticated scaling strategy. FR-TTS efficiently searches for good filling schemes and incorporates a diversity reward with a dynamic weighting schedule to achieve a balanced and comprehensive evaluation of intermediate samples. We experimentally validate the superiority of FR-TTS over multiple established benchmarks and various reward models. Code is available at \href{https://github.com/xuhang07/FR-TTS}{https://github.com/xuhang07/FR-TTS}.
[204] RecruitView: A Multimodal Dataset for Predicting Personality and Interview Performance for Human Resources Applications
Amit Kumar Gupta, Farhan Sheth, Hammad Shaikh, Dheeraj Kumar, Angkul Puniya, Deepak Panwar, Sandeep Chaurasia, Priya Mathur
Main category: cs.CV
TL;DR: RecruitView dataset with 2,011 video interview clips and 27,000 pairwise judgments across 12 personality/performance dimensions, plus CRMF geometric deep learning framework using hyperbolic, spherical, and Euclidean manifolds for multimodal personality assessment.
Details
Motivation: Automated personality and soft skill assessment from multimodal behavioral data is challenging due to limited datasets and methods that fail to capture the geometric structure inherent in human traits.Method: Cross-Modal Regression with Manifold Fusion (CRMF): geometric deep learning framework modeling behavioral representations across hyperbolic, spherical, and Euclidean manifolds. Uses geometry-specific expert networks for hierarchical traits, directional patterns, and continuous variations, with adaptive routing and tangent space fusion.
Result: CRMF substantially outperforms baselines with up to 11.4% improvement in Spearman correlation and 6.0% in concordance index, while training 40-50% fewer parameters than large multimodal models.
Conclusion: The RecruitView dataset and CRMF framework provide effective solutions for automated personality assessment by capturing geometric structure in human traits through multimodal behavioral data analysis.
Abstract: Automated personality and soft skill assessment from multimodal behavioral data remains challenging due to limited datasets and methods that fail to capture geometric structure inherent in human traits. We introduce RecruitView, a dataset of 2,011 naturalistic video interview clips from 300+ participants with 27,000 pairwise comparative judgments across 12 dimensions: Big Five personality traits, overall personality score, and six interview performance metrics. To leverage this data, we propose Cross-Modal Regression with Manifold Fusion (CRMF), a geometric deep learning framework that explicitly models behavioral representations across hyperbolic, spherical, and Euclidean manifolds. CRMF employs geometry-specific expert networks to capture hierarchical trait structures, directional behavioral patterns, and continuous performance variations simultaneously. An adaptive routing mechanism dynamically weights expert contributions based on input characteristics. Through principled tangent space fusion, CRMF achieves superior performance while training 40-50% fewer trainable parameters than large multimodal models. Extensive experiments demonstrate that CRMF substantially outperforms the selected baselines, achieving up to 11.4% improvement in Spearman correlation and 6.0% in concordance index. Our RecruitView dataset is publicly available at https://huggingface.co/datasets/AI4A-lab/RecruitView
[205] CausalAffect: Causal Discovery for Facial Affective Understanding
Guanyu Hu, Tangzheng Lian, Dimitrios Kollias, Oya Celiktutan, Xinyu Yang
Main category: cs.CV
TL;DR: CausalAffect is the first framework for causal graph discovery in facial affect analysis that models AU-AU and AU-Expression dependencies through a two-level causal hierarchy with counterfactual interventions, achieving state-of-the-art results without requiring annotated datasets or handcrafted priors.
Details
Motivation: Existing approaches to facial affect analysis rarely address how to infer psychologically plausible causal relations between Action Units (AUs) and expressions directly from data, despite AUs being the foundation of affective computing. There's a need for structured reasoning over latent dependencies that drive muscle activations and their expressive outcomes.Method: CausalAffect uses a two-level polarity and direction aware causal hierarchy that integrates population-level regularities with sample-adaptive structures. It includes a feature-level counterfactual intervention mechanism to enforce true causal effects while suppressing spurious correlations. The approach requires neither jointly annotated datasets nor handcrafted causal priors.
Result: The framework recovers causal structures consistent with established psychological theories while revealing novel inhibitory and previously uncharacterized dependencies. Extensive experiments across six benchmarks demonstrate state-of-the-art performance in both AU detection and expression recognition.
Conclusion: CausalAffect establishes a principled connection between causal discovery and interpretable facial behavior, advancing the field by providing the first framework for causal graph discovery in facial affect analysis with practical applications and theoretical consistency.
Abstract: Understanding human affect from facial behavior requires not only accurate recognition but also structured reasoning over the latent dependencies that drive muscle activations and their expressive outcomes. Although Action Units (AUs) have long served as the foundation of affective computing, existing approaches rarely address how to infer psychologically plausible causal relations between AUs and expressions directly from data. We propose CausalAffect, the first framework for causal graph discovery in facial affect analysis. CausalAffect models AU-AU and AU-Expression dependencies through a two-level polarity and direction aware causal hierarchy that integrates population-level regularities with sample-adaptive structures. A feature-level counterfactual intervention mechanism further enforces true causal effects while suppressing spurious correlations. Crucially, our approach requires neither jointly annotated datasets nor handcrafted causal priors, yet it recovers causal structures consistent with established psychological theories while revealing novel inhibitory and previously uncharacterized dependencies. Extensive experiments across six benchmarks demonstrate that CausalAffect advances the state of the art in both AU detection and expression recognition, establishing a principled connection between causal discovery and interpretable facial behavior. All trained models and source code will be released upon acceptance.
[206] RealGen: Photorealistic Text-to-Image Generation via Detector-Guided Rewards
Junyan Ye, Leiqi Zhu, Yuncheng Guo, Dongzhi Jiang, Zilong Huang, Yifan Zhang, Zhiyuan Yan, Haohuan Fu, Conghui He, Weijia Li
Main category: cs.CV
TL;DR: RealGen is a photorealistic text-to-image framework that addresses AI artifacts in current models by integrating prompt optimization with diffusion models and using a Detector Reward mechanism to enhance realism.
Details
Motivation: Current advanced image generation models (GPT-Image-1, Qwen-Image) achieve good text-to-image consistency but fail at photorealism, producing "fake" images with AI artifacts like overly smooth skin and oily facial sheens, missing the original goal of "indistinguishable-from-reality" generation.Method: RealGen integrates an LLM for prompt optimization and a diffusion model for image generation. It introduces a “Detector Reward” mechanism that quantifies artifacts and assesses realism using semantic-level and feature-level synthetic image detectors, then uses GRPO algorithm to optimize the entire pipeline.
Result: RealGen significantly outperforms general models (GPT-Image-1, Qwen-Image) and specialized photorealistic models (FLUX-Krea) in terms of realism, detail, and aesthetics.
Conclusion: RealGen successfully addresses photorealism in text-to-image generation through its integrated framework with detector-based reward optimization, and the proposed RealBench benchmark enables accurate, human-free evaluation of photorealism aligned with real user experience.
Abstract: With the continuous advancement of image generation technology, advanced models such as GPT-Image-1 and Qwen-Image have achieved remarkable text-to-image consistency and world knowledge However, these models still fall short in photorealistic image generation. Even on simple T2I tasks, they tend to produce " fake" images with distinct AI artifacts, often characterized by “overly smooth skin” and “oily facial sheens”. To recapture the original goal of “indistinguishable-from-reality” generation, we propose RealGen, a photorealistic text-to-image framework. RealGen integrates an LLM component for prompt optimization and a diffusion model for realistic image generation. Inspired by adversarial generation, RealGen introduces a “Detector Reward” mechanism, which quantifies artifacts and assesses realism using both semantic-level and feature-level synthetic image detectors. We leverage this reward signal with the GRPO algorithm to optimize the entire generation pipeline, significantly enhancing image realism and detail. Furthermore, we propose RealBench, an automated evaluation benchmark employing Detector-Scoring and Arena-Scoring. It enables human-free photorealism assessment, yielding results that are more accurate and aligned with real user experience. Experiments demonstrate that RealGen significantly outperforms general models like GPT-Image-1 and Qwen-Image, as well as specialized photorealistic models like FLUX-Krea, in terms of realism, detail, and aesthetics. The code is available at https://github.com/yejy53/RealGen.
[207] Structured Context Learning for Generic Event Boundary Detection
Xin Gu, Congcong Li, Xinyao Wang, Dexiang Hong, Libo Zhang, Tiejian Luo, Longyin Wen, Heng Fan
Main category: cs.CV
TL;DR: Proposes Structured Context Learning with Structured Partition of Sequence (SPoS) for Generic Event Boundary Detection, achieving linear computational complexity and state-of-the-art performance across multiple datasets.
Details
Motivation: Generic Event Boundary Detection (GEBD) aims to identify moments in videos that humans perceive as event boundaries, but existing methods lack structured context and have computational limitations.Method: Uses Structured Partition of Sequence (SPoS) to partition input frames and provide structured context, calculates group similarities to capture frame differences, employs lightweight fully convolutional network for boundary detection, and adapts Gaussian kernel to handle annotation ambiguities.
Result: Achieves superior performance over state-of-the-art methods on Kinetics-GEBD, TAPOS, and shot transition detection datasets with better speed-accuracy trade-off and linear computational complexity.
Conclusion: The proposed Structured Context Learning with SPoS provides an effective, flexible, and efficient solution for GEBD that outperforms existing methods while maintaining computational efficiency.
Abstract: Generic Event Boundary Detection (GEBD) aims to identify moments in videos that humans perceive as event boundaries. This paper proposes a novel method for addressing this task, called Structured Context Learning, which introduces the Structured Partition of Sequence (SPoS) to provide a structured context for learning temporal information. Our approach is end-to-end trainable and flexible, not restricted to specific temporal models like GRU, LSTM, and Transformers. This flexibility enables our method to achieve a better speed-accuracy trade-off. Specifically, we apply SPoS to partition the input frame sequence and provide a structured context for the subsequent temporal model. Notably, SPoS’s overall computational complexity is linear with respect to the video length. We next calculate group similarities to capture differences between frames, and a lightweight fully convolutional network is utilized to determine the event boundaries based on the grouped similarity maps. To remedy the ambiguities of boundary annotations, we adapt the Gaussian kernel to preprocess the ground-truth event boundaries. Our proposed method has been extensively evaluated on the challenging Kinetics-GEBD, TAPOS, and shot transition detection datasets, demonstrating its superiority over existing state-of-the-art methods.
[208] Learning What Helps: Task-Aligned Context Selection for Vision Tasks
Jingyu Guo, Emir Konuk, Fredrik Strand, Christos Matsoukas, Kevin Smith
Main category: cs.CV
TL;DR: TACS is a framework that teaches ViTs to select helpful context examples for better predictions, outperforming similarity-based retrieval across 18 datasets.
Details
Motivation: Humans resolve visual uncertainty by comparing images with relevant examples, but Vision Transformers (ViTs) lack the ability to identify which examples would actually improve their predictions. Current similarity-based retrieval methods often select examples that merely look similar rather than those that genuinely help with the task.Method: Task-Aligned Context Selection (TACS) jointly trains a selector network with the task model using a hybrid optimization scheme combining gradient-based supervision and reinforcement learning. This makes retrieval part of the learning objective, aligning example selection with task performance rewards rather than just visual similarity.
Result: Across 18 datasets covering fine-grained recognition, medical image classification, and medical image segmentation, TACS consistently outperforms similarity-based retrieval methods. The improvements are particularly significant in challenging or data-limited settings.
Conclusion: By aligning context selection with task rewards, TACS enables discriminative models to discover which contextual examples genuinely help improve predictions, moving beyond superficial similarity to achieve better task performance.
Abstract: Humans often resolve visual uncertainty by comparing an image with relevant examples, but ViTs lack the ability to identify which examples would improve their predictions. We present Task-Aligned Context Selection (TACS), a framework that learns to select paired examples which truly improve task performance rather than those that merely appear similar. TACS jointly trains a selector network with the task model through a hybrid optimization scheme combining gradient-based supervision and reinforcement learning, making retrieval part of the learning objective. By aligning selection with task rewards, TACS enables discriminative models to discover which contextual examples genuinely help. Across 18 datasets covering fine-grained recognition, medical image classification, and medical image segmentation, TACS consistently outperforms similarity-based retrieval, particularly in challenging or data-limited settings.
[209] CC-FMO: Camera-Conditioned Zero-Shot Single Image to 3D Scene Generation with Foundation Model Orchestration
Boshi Tang, Henry Zheng, Rui Huang, Gao Huang
Main category: cs.CV
TL;DR: CC-FMO is a zero-shot pipeline for generating coherent 3D scenes from single images by combining semantic-aware object generation with camera-conditioned layout preservation.
Details
Motivation: Existing methods struggle with coherent scene generation due to inaccurate object pose estimation and spatial inconsistency, especially when moving from instance-level to scene-level 3D generation from single images.Method: Uses a hybrid instance generator combining semantics-aware vector-set representation with detail-rich structured latent representation, plus a camera-conditioned scale-solving algorithm to apply foundational pose estimation models for scene coherence.
Result: Extensive experiments show CC-FMO consistently generates high-fidelity camera-aligned compositional scenes, outperforming all state-of-the-art methods.
Conclusion: CC-FMO successfully addresses scene generation challenges by preserving both object layout fidelity and instance quality through its novel hybrid approach and camera-conditioned pipeline.
Abstract: High-quality 3D scene generation from a single image is crucial for AR/VR and embodied AI applications. Early approaches struggle to generalize due to reliance on specialized models trained on curated small datasets. While recent advancements in large-scale 3D foundation models have significantly enhanced instance-level generation, coherent scene generation remains a challenge, where performance is limited by inaccurate per-object pose estimations and spatial inconsistency. To this end, this paper introduces CC-FMO, a zero-shot, camera-conditioned pipeline for single-image to 3D scene generation that jointly conforms to the object layout in input image and preserves instance fidelity. CC-FMO employs a hybrid instance generator that combines semantics-aware vector-set representation with detail-rich structured latent representation, yielding object geometries that are both semantically plausible and high-quality. Furthermore, CC-FMO enables the application of foundational pose estimation models in the scene generation task via a simple yet effective camera-conditioned scale-solving algorithm, to enforce scene-level coherence. Extensive experiments demonstrate that CC-FMO consistently generates high-fidelity camera-aligned compositional scenes, outperforming all state-of-the-art methods.
[210] Terrain Sensing with Smartphone Structured Light: 2D Dynamic Time Warping for Grid Pattern Matching
Tanaka Nobuaki
Main category: cs.CV
TL;DR: Smartphone-based structured-light system projects grid pattern onto ground to detect small terrain unevenness using novel 2D-DTW algorithm for grid matching.
Details
Motivation: Mobile rovers on uneven terrain need to detect small bumps/tilts that are visually imperceptible but affect locomotion stability. Current visual perception is insufficient for these subtle terrain variations.Method: Smartphone-based structured-light system projects grid pattern onto ground. Uses novel topology-constrained 2D dynamic time warping (2D-DTW) algorithm for robust grid matching under perspective distortion and occlusion. Designed for resource-limited platforms while preserving grid structure for accurate triangulation.
Result: Proposed 2D-DTW formulation enables terrain sensing by matching deformed grid patterns and also serves as general tool for structured grid pattern matching in image processing. System reconstructs local terrain unevenness from single handheld device.
Conclusion: The smartphone-based structured-light system with 2D-DTW algorithm provides practical solution for detecting subtle terrain unevenness on resource-limited platforms, with applications beyond terrain sensing to general grid pattern matching problems.
Abstract: Low-cost mobile rovers often operate on uneven terrain where small bumps or tilts are difficult to perceive visually but can significantly affect locomotion stability. To address this problem, we explore a smartphone-based structured-light system that projects a grid pattern onto the ground and reconstructs local terrain unevenness from a single handheld device. The system is inspired by face-recognition projectors, but adapted for ground sensing. A key technical challenge is robustly matching the projected grid with its deformed observation under perspective distortion and partial occlusion. Conventional one-dimensional dynamic time warping (1D-DTW) is not directly applicable to such two-dimensional grid patterns. We therefore propose a topology-constrained two-dimensional dynamic time warping (2D-DTW) algorithm that performs column-wise alignment under a global grid consistency constraint. The proposed method is designed to be simple enough to run on resource limited platforms while preserving the grid structure required for accurate triangulation. We demonstrate that our 2D-DTW formulation can be used not only for terrain sensing but also as a general tool for matching structured grid patterns in image processing scenarios. This paper describes the overall system design as well as the 2D-DTW extension that emerged from this application.
[211] Image Generation as a Visual Planner for Robotic Manipulation
Ye Pang
Main category: cs.CV
TL;DR: Image generation models can serve as visual planners for robotic manipulation with minimal adaptation, producing coherent robot videos from text or trajectory conditions.
Details
Motivation: Existing video diffusion models require large domain-specific datasets and struggle to generalize, while pretrained image generation models show strong compositionality and temporal coherence that could be leveraged for robotic planning.Method: Two-part framework: (1) text-conditioned generation using language instruction and first frame, and (2) trajectory-conditioned generation using 2D trajectory overlay and initial frame. Uses LoRA fine-tuning for adaptation.
Result: Experiments on Jaco Play, Bridge V2, and RT1 datasets show both modes produce smooth, coherent robot videos aligned with their respective conditions.
Conclusion: Pretrained image generators encode transferable temporal priors and can function as video-like robotic planners under minimal supervision.
Abstract: Generating realistic robotic manipulation videos is an important step toward unifying perception, planning, and action in embodied agents. While existing video diffusion models require large domain-specific datasets and struggle to generalize, recent image generation models trained on language-image corpora exhibit strong compositionality, including the ability to synthesize temporally coherent grid images. This suggests a latent capacity for video-like generation even without explicit temporal modeling. We explore whether such models can serve as visual planners for robots when lightly adapted using LoRA finetuning. We propose a two-part framework that includes: (1) text-conditioned generation, which uses a language instruction and the first frame, and (2) trajectory-conditioned generation, which uses a 2D trajectory overlay and the same initial frame. Experiments on the Jaco Play dataset, Bridge V2, and the RT1 dataset show that both modes produce smooth, coherent robot videos aligned with their respective conditions. Our findings indicate that pretrained image generators encode transferable temporal priors and can function as video-like robotic planners under minimal supervision. Code is released at \href{https://github.com/pangye202264690373/Image-Generation-as-a-Visual-Planner-for-Robotic-Manipulation}{https://github.com/pangye202264690373/Image-Generation-as-a-Visual-Planner-for-Robotic-Manipulation}.
[212] Cross-Temporal 3D Gaussian Splatting for Sparse-View Guided Scene Update
Zeyuan An, Yanghang Xiao, Zhiying Leng, Frederick W. B. Li, Xiaohui Liang
Main category: cs.CV
TL;DR: Cross-Temporal 3DGS: A framework for reconstructing and updating 3D scenes across time periods using sparse images and scene priors, enabling efficient scene versioning and cross-temporal digital twins.
Details
Motivation: Maintaining consistent 3D scene representations over time is challenging but crucial for applications like urban planning, disaster assessment, and historical preservation where dense scans are often unavailable. There's a need for efficient methods to update 3D scenes from sparse-view observations.Method: Three-stage framework: 1) Cross-temporal camera alignment for pose estimation across timestamps, 2) Interference-based confidence initialization to identify unchanged regions between timestamps, 3) Progressive cross-temporal optimization that iteratively integrates historical prior information into 3D scenes using 3D Gaussian Splatting.
Result: Experimental results show significant improvements over baseline methods in reconstruction quality and data efficiency. The method supports non-continuous capture, enabling scene updates with new sparse views and recovery of past scenes from limited data.
Conclusion: Cross-Temporal 3DGS is a promising solution for scene versioning, cross-temporal digital twins, and long-term spatial documentation, demonstrating potential for temporal change analysis using only sparse images that can later be reconstructed into detailed 3D representations.
Abstract: Maintaining consistent 3D scene representations over time is a significant challenge in computer vision. Updating 3D scenes from sparse-view observations is crucial for various real-world applications, including urban planning, disaster assessment, and historical site preservation, where dense scans are often unavailable or impractical. In this paper, we propose Cross-Temporal 3D Gaussian Splatting (Cross-Temporal 3DGS), a novel framework for efficiently reconstructing and updating 3D scenes across different time periods, using sparse images and previously captured scene priors. Our approach comprises three stages: 1) Cross-temporal camera alignment for estimating and aligning camera poses across different timestamps; 2) Interference-based confidence initialization to identify unchanged regions between timestamps, thereby guiding updates; and 3) Progressive cross-temporal optimization, which iteratively integrates historical prior information into the 3D scene to enhance reconstruction quality. Our method supports non-continuous capture, enabling not only updates using new sparse views to refine existing scenes, but also recovering past scenes from limited data with the help of current captures. Furthermore, we demonstrate the potential of this approach to achieve temporal changes using only sparse images, which can later be reconstructed into detailed 3D representations as needed. Experimental results show significant improvements over baseline methods in reconstruction quality and data efficiency, making this approach a promising solution for scene versioning, cross-temporal digital twins, and long-term spatial documentation.
[213] SAIDO: Generalizable Detection of AI-Generated Images via Scene-Aware and Importance-Guided Dynamic Optimization in Continual Learning
Yongkang Hu, Yu Cheng, Yushuo Zhang, Yuan Xie, Zhaoxia Yin
Main category: cs.CV
TL;DR: SAIDO: A continual learning framework for AI-generated image detection that uses scene-aware expert modules and importance-guided optimization to improve generalization across emerging generative methods and content types.
Details
Motivation: The widespread misuse of image generation technologies creates security concerns, but existing detection methods struggle with generalization to new generative methods and content types in real-world scenarios.Method: Proposes SAIDO framework with two key components: 1) Scene-Awareness-Based Expert Module (SAEM) that dynamically identifies new scenes using VLLMs and allocates independent expert modules for each scene, 2) Importance-Guided Dynamic Optimization Mechanism (IDOM) that optimizes neurons through importance-guided gradient projection to balance plasticity and stability while mitigating catastrophic forgetting.
Result: Outperforms current SOTA in continual learning tasks with 44.22% relative reduction in average detection error rate and 40.57% reduction in forgetting rate. On open-world datasets, improves average detection accuracy by 9.47% compared to current SOTA.
Conclusion: SAIDO effectively addresses the generalization challenge in AI-generated image detection by combining scene-aware expert allocation with importance-guided optimization, achieving superior performance in both stability and plasticity for continual learning scenarios.
Abstract: The widespread misuse of image generation technologies has raised security concerns, driving the development of AI-generated image detection methods. However, generalization has become a key challenge and open problem: existing approaches struggle to adapt to emerging generative methods and content types in real-world scenarios. To address this issue, we propose a Scene-Aware and Importance-Guided Dynamic Optimization detection framework with continual learning (SAIDO). Specifically, we design Scene-Awareness-Based Expert Module (SAEM) that dynamically identifies and incorporates new scenes using VLLMs. For each scene, independent expert modules are dynamically allocated, enabling the framework to capture scene-specific forgery features better and enhance cross-scene generalization. To mitigate catastrophic forgetting when learning from multiple image generative methods, we introduce Importance-Guided Dynamic Optimization Mechanism (IDOM), which optimizes each neuron through an importance-guided gradient projection strategy, thereby achieving an effective balance between model plasticity and stability. Extensive experiments on continual learning tasks demonstrate that our method outperforms the current SOTA method in both stability and plasticity, achieving 44.22% and 40.57% relative reductions in average detection error rate and forgetting rate, respectively. On open-world datasets, it improves the average detection accuracy by 9.47% compared to the current SOTA method.
[214] Asset-Driven Sematic Reconstruction of Dynamic Scene with Multi-Human-Object Interactions
Sandika Biswas, Qianyi Wu, Biplab Banerjee, Hamid Rezatofighi
Main category: cs.CV
TL;DR: A hybrid approach combining 3D generative models, semantic-aware deformation, and Gaussian Splatting optimization for reconstructing multi-human, multi-object dynamic scenes from monocular video.
Details
Motivation: Real-world environments are dynamic with complex human-object interactions, but 3D geometry modeling of such scenes remains challenging due to diverse motion patterns, frequent occlusions, and the difficulty of maintaining structural consistency in monocular setups.Method: Proposes a hybrid approach: 1) Uses 3D generative models to create high-fidelity meshes of scene elements, 2) Applies semantic-aware deformation (rigid transformations for objects, LBS-based deformation for humans), and 3) Performs Gaussian Splatting optimization to refine element alignments in the dynamic scene.
Result: Outperforms state-of-the-art methods on the HOI-M3 dataset (the only dataset with multi-human, multi-object interactions), producing better surface reconstruction that maintains object structures under severe occlusion with multi-view and temporal consistency.
Conclusion: The hybrid approach effectively addresses challenges in multi-human, multi-object dynamic scene reconstruction by combining strengths of different techniques, enabling robust geometry modeling even under severe occlusion conditions.
Abstract: Real-world human-built environments are highly dynamic, involving multiple humans and their complex interactions with surrounding objects. While 3D geometry modeling of such scenes is crucial for applications like AR/VR, gaming, and embodied AI, it remains underexplored due to challenges like diverse motion patterns and frequent occlusions. Beyond novel view rendering, 3D Gaussian Splatting (GS) has demonstrated remarkable progress in producing detailed, high-quality surface geometry with fast optimization of the underlying structure. However, very few GS-based methods address multihuman, multiobject scenarios, primarily due to the above-mentioned inherent challenges. In a monocular setup, these challenges are further amplified, as maintaining structural consistency under severe occlusion becomes difficult when the scene is optimized solely based on GS-based rendering loss. To tackle the challenges of such a multihuman, multiobject dynamic scene, we propose a hybrid approach that effectively combines the advantages of 1) 3D generative models for generating high-fidelity meshes of the scene elements, 2) Semantic-aware deformation, \ie rigid transformation of the rigid objects and LBS-based deformation of the humans, and mapping of the deformed high-fidelity meshes in the dynamic scene, and 3) GS-based optimization of the individual elements for further refining their alignments in the scene. Such a hybrid approach helps maintain the object structures even under severe occlusion and can produce multiview and temporally consistent geometry. We choose HOI-M3 for evaluation, as, to the best of our knowledge, this is the only dataset featuring multihuman, multiobject interactions in a dynamic scene. Our method outperforms the state-of-the-art method in producing better surface reconstruction of such scenes.
[215] NeuroVolve: Evolving Visual Stimuli toward Programmable Neural Objectives
Haomiao Chen, Keith W Jamison, Mert R. Sabuncu, Amy Kuceyeski
Main category: cs.CV
TL;DR: NeuroVolve is a generative framework that synthesizes images guided by neural activity patterns from brain regions, using optimization in vision-language model embedding space to create stimuli that satisfy single or multiple region constraints.
Details
Motivation: Prior work focused on replicating known category selectivity in isolated brain regions (like faces in FFA), but offered limited insight into how different brain regions interact during complex, naturalistic vision. There's a need to understand how distributed neural patterns combine to create representations.Method: NeuroVolve optimizes a neural objective function in the embedding space of a pretrained vision-language model. Images are generated by programming neural objectives that activate/deactivate single or multiple brain regions together. The framework tracks optimization steps to reveal semantic trajectories through embedding space.
Result: The framework successfully recovers known selectivity for individual brain regions and synthesizes coherent scenes satisfying complex multi-region constraints. It generates both low-level and semantic feature-specific stimuli for single ROIs, as well as stimuli aligned to curated neural objectives including co-activation and decorrelation between regions. It captures subject-specific preferences for personalized brain-driven synthesis.
Conclusion: NeuroVolve unifies brain-guided image editing and preferred stimulus generation in a single process, exposing cooperative and antagonistic tuning relationships between brain regions. It offers interpretable constraints for mapping, analyzing, and probing neural representations of visual information, supporting personalized brain-driven synthesis.
Abstract: What visual information is encoded in individual brain regions, and how do distributed patterns combine to create their neural representations? Prior work has used generative models to replicate known category selectivity in isolated regions (e.g., faces in FFA), but these approaches offer limited insight into how regions interact during complex, naturalistic vision. We introduce NeuroVolve, a generative framework that provides brain-guided image synthesis via optimization of a neural objective function in the embedding space of a pretrained vision-language model. Images are generated under the guidance of a programmable neural objective, i.e., activating or deactivating single regions or multiple regions together. NeuroVolve is validated by recovering known selectivity for individual brain regions, while expanding to synthesize coherent scenes that satisfy complex, multi-region constraints. By tracking optimization steps, it reveals semantic trajectories through embedding space, unifying brain-guided image editing and preferred stimulus generation in a single process. We show that NeuroVolve can generate both low-level and semantic feature-specific stimuli for single ROIs, as well as stimuli aligned to curated neural objectives. These include co-activation and decorrelation between regions, exposing cooperative and antagonistic tuning relationships. Notably, the framework captures subject-specific preferences, supporting personalized brain-driven synthesis and offering interpretable constraints for mapping, analyzing, and probing neural representations of visual information.
[216] Describe Anything Anywhere At Any Moment
Nicolas Gorlo, Lukas Schmid, Luca Carlone
Main category: cs.CV
TL;DR: DAAAM is a real-time 4D scene understanding framework that builds hierarchical 4D scene graphs with detailed semantic descriptions while maintaining real-time performance, achieving state-of-the-art results on spatio-temporal QA and task grounding benchmarks.
Details
Motivation: Existing methods face a tradeoff between producing rich open-vocabulary descriptions and real-time performance when grounding these descriptions in 3D. Applications like augmented reality and robot autonomy require spatio-temporal memory frameworks that capture both geometric structure and semantic detail.Method: DAAAM introduces an optimization-based frontend that infers detailed semantic descriptions from localized captioning models (like DAM) using batch processing to speed up inference. It builds hierarchical 4D scene graphs that serve as globally consistent spatio-temporal memory representations.
Result: DAAAM achieves state-of-the-art results: improves OC-NaVQA question accuracy by 53.6%, position errors by 21.9%, temporal errors by 21.6%, and SG3D task grounding accuracy by 27.8% over competitive baselines.
Conclusion: DAAAM provides a novel spatio-temporal memory framework that enables real-time 4D scene understanding with detailed semantic descriptions, effectively bridging the gap between rich open-vocabulary scene understanding and real-time performance for robotics and vision applications.
Abstract: Computer vision and robotics applications ranging from augmented reality to robot autonomy in large-scale environments require spatio-temporal memory frameworks that capture both geometric structure for accurate language-grounding as well as semantic detail. Existing methods face a tradeoff, where producing rich open-vocabulary descriptions comes at the expense of real-time performance when these descriptions have to be grounded in 3D. To address these challenges, we propose Describe Anything, Anywhere, at Any Moment (DAAAM), a novel spatio-temporal memory framework for large-scale and real-time 4D scene understanding. DAAAM introduces a novel optimization-based frontend to infer detailed semantic descriptions from localized captioning models, such as the Describe Anything Model (DAM), leveraging batch processing to speed up inference by an order of magnitude for online processing. It leverages such semantic understanding to build a hierarchical 4D scene graph (SG), which acts as an effective globally spatially and temporally consistent memory representation. DAAAM constructs 4D SGs with detailed, geometrically grounded descriptions while maintaining real-time performance. We show that DAAAM’s 4D SG interfaces well with a tool-calling agent for inference and reasoning. We thoroughly evaluate DAAAM in the complex task of spatio-temporal question answering on the NaVQA benchmark and show its generalization capabilities for sequential task grounding on the SG3D benchmark. We further curate an extended OC-NaVQA benchmark for large-scale and long-time evaluations. DAAAM achieves state-of-the-art results in both tasks, improving OC-NaVQA question accuracy by 53.6%, position errors by 21.9%, temporal errors by 21.6%, and SG3D task grounding accuracy by 27.8% over the most competitive baselines, respectively. We release our data and code open-source.
[217] Integrating Skeleton Based Representations for Robust Yoga Pose Classification Using Deep Learning Models
Mohammed Mohiuddin, Syed Mohammod Minhaz Hossain, Sumaiya Khanam, Prionkar Barua, Aparup Barua, MD Tamim Hossain
Main category: cs.CV
TL;DR: This paper introduces Yoga-16 dataset and benchmarks deep learning models for yoga pose classification, finding skeleton-based representations outperform raw images with VGG16+MediaPipe achieving 96.09% accuracy.
Details
Motivation: Yoga is popular but incorrect postures can cause injuries. Automated yoga pose classification is needed to reduce reliance on expert practitioners, but systematic benchmarking for yoga pose recognition remains limited with prior works focusing only on raw images or single pose extraction models.Method: Created curated ‘Yoga-16’ dataset addressing limitations of existing datasets. Systematically evaluated three deep learning architectures (VGG16, ResNet50, Xception) using three input modalities: direct images, MediaPipe Pose skeleton images, and YOLOv8 Pose skeleton images. Used Grad-CAM for interpretability analysis and cross validation.
Result: Skeleton-based representations outperform raw image inputs. Highest accuracy of 96.09% achieved by VGG16 with MediaPipe Pose skeleton input. Provided interpretability insights through Grad-CAM analysis.
Conclusion: Skeleton-based approaches are superior for yoga pose classification, with VGG16+MediaPipe being the most effective combination. The Yoga-16 dataset and systematic benchmarking provide valuable resources for future research in automated yoga pose assessment.
Abstract: Yoga is a popular form of exercise worldwide due to its spiritual and physical health benefits, but incorrect postures can lead to injuries. Automated yoga pose classification has therefore gained importance to reduce reliance on expert practitioners. While human pose keypoint extraction models have shown high potential in action recognition, systematic benchmarking for yoga pose recognition remains limited, as prior works often focus solely on raw images or a single pose extraction model. In this study, we introduce a curated dataset, ‘Yoga-16’, which addresses limitations of existing datasets, and systematically evaluate three deep learning architectures (VGG16, ResNet50, and Xception) using three input modalities (direct images, MediaPipe Pose skeleton images, and YOLOv8 Pose skeleton images). Our experiments demonstrate that skeleton-based representations outperform raw image inputs, with the highest accuracy of 96.09% achieved by VGG16 with MediaPipe Pose skeleton input. Additionally, we provide interpretability analysis using Grad-CAM, offering insights into model decision-making for yoga pose classification with cross validation analysis.
[218] SatireDecoder: Visual Cascaded Decoupling for Enhancing Satirical Image Comprehension
Yue Jiang, Haiwei Xue, Minghao Han, Mingcheng Li, Xiaolu Hou, Dingkang Yang, Lihua Zhang, Xu Zheng
Main category: cs.CV
TL;DR: SatireDecoder: A training-free framework using multi-agent visual cascaded decoupling and chain-of-thought reasoning to improve visual satire comprehension in vision-language models.
Details
Motivation: Visual satire comprehension is challenging for current vision-language models, requiring detection of satire, deciphering nuanced meaning, and identifying implicated entities. Existing models fail to integrate local entity relationships with global context, leading to misinterpretation, biases, and hallucinations.Method: Proposes SatireDecoder framework with: 1) Multi-agent system performing visual cascaded decoupling to decompose images into fine-grained local and global semantic representations; 2) Chain-of-thought reasoning guided by uncertainty analysis to break down satire comprehension into sequential subtasks with minimized uncertainty.
Result: SatireDecoder significantly improves interpretive accuracy while reducing hallucinations. Experimental results show it outperforms existing baselines in comprehending visual satire.
Conclusion: The framework offers a promising direction for vision-language reasoning in nuanced, high-level semantic tasks like satire comprehension, addressing limitations of current models through training-free, structured decomposition and reasoning approaches.
Abstract: Satire, a form of artistic expression combining humor with implicit critique, holds significant social value by illuminating societal issues. Despite its cultural and societal significance, satire comprehension, particularly in purely visual forms, remains a challenging task for current vision-language models. This task requires not only detecting satire but also deciphering its nuanced meaning and identifying the implicated entities. Existing models often fail to effectively integrate local entity relationships with global context, leading to misinterpretation, comprehension biases, and hallucinations. To address these limitations, we propose SatireDecoder, a training-free framework designed to enhance satirical image comprehension. Our approach proposes a multi-agent system performing visual cascaded decoupling to decompose images into fine-grained local and global semantic representations. In addition, we introduce a chain-of-thought reasoning strategy guided by uncertainty analysis, which breaks down the complex satire comprehension process into sequential subtasks with minimized uncertainty. Our method significantly improves interpretive accuracy while reducing hallucinations. Experimental results validate that SatireDecoder outperforms existing baselines in comprehending visual satire, offering a promising direction for vision-language reasoning in nuanced, high-level semantic tasks.
[219] Scaling Down to Scale Up: Towards Operationally-Efficient and Deployable Clinical Models via Cross-Modal Low-Rank Adaptation for Medical Vision-Language Models
Thuraya Alzubaidi, Farhad R. Nezami, Muzammil Behzad
Main category: cs.CV
TL;DR: MedCT-VLM adapts CT-CLIP foundation model for zero-shot medical CT classification using LoRA, achieving significant performance gains with only 0.38% parameter tuning.
Details
Motivation: Foundation models show strong zero-shot capabilities but have limited application to volumetric medical imaging. There's a need to adapt large-scale CT foundation models for downstream clinical tasks efficiently, especially when labeled data is scarce.Method: MedCT-VLM uses parameter-efficient adaptation of CT-CLIP (trained on 25,692 chest CT volumes) via Low-Rank Adaptation (LoRA). Instead of fine-tuning all 440M parameters, it inserts low-rank decomposition matrices into attention layers of both vision and text encoders, training only 1.67M parameters (0.38% of total).
Result: LoRA fine-tuning improves mean AUROC from 61.3% to 68.9% (+7.6 pp), accuracy from 67.2% to 73.6% (+6.4 pp), and macro-F1 from 32.1% to 36.9% (+4.8 pp) on zero-shot classification across 18 thoracic pathologies.
Conclusion: Parameter-efficient methods like LoRA can effectively transfer large-scale pretraining to downstream medical imaging tasks, particularly valuable for zero-shot scenarios where labeled data is scarce.
Abstract: Foundation models trained via vision-language pretraining have demonstrated strong zero-shot capabilities across diverse image domains, yet their application to volumetric medical imaging remains limited. We introduce MedCT-VLM: Medical CT Vision-Language Model, a parameter-efficient vision-language framework designed to adapt large-scale CT foundation models for downstream clinical tasks. MedCT-VLM uses a parameter-efficient approach to adapt CT-CLIP, a contrastive vision-language model trained on 25,692 chest CT volumes, for multi-label pathology classification using Low-Rank Adaptation (LoRA). Rather than fine-tuning the model’s 440 M parameters directly, we insert low-rank decomposition matrices into attention layers of both vision and text encoders, training only 1.67M parameters (0.38% of total). We evaluate on zero-shot classification across 18 thoracic pathologies, where the model must align CT embeddings with unseen text prompts at inference without task-specific training. LoRA fine-tuning improves mean AUROC from 61.3% to 68.9% (+7.6 pp), accuracy from 67.2% to 73.6% (+6.4 pp), and macro-F1 from 32.1% to 36.9% (+4.8 pp). These results demonstrate that parameter-efficient methods can effectively transfer large-scale pretraining to downstream medical imaging tasks, particularly for zero-shot scenarios where labeled data is scarce.
[220] Automatic Pith Detection in Tree Cross-Section Images Using Deep Learning
Tzu-I Liao, Mahmoud Fakhry, Jibin Yesudas Varghese
Main category: cs.CV
TL;DR: Deep learning models (YOLOv9, U-Net, Swin Transformer, DeepLabV3, Mask R-CNN) were evaluated for automated pith detection in tree cross-sections, with Swin Transformer achieving highest accuracy (0.94) and Mask R-CNN showing significant improvement with NMS.
Details
Motivation: Pith detection in tree cross-sections is currently manual and error-prone, creating a need for automated solutions to improve forestry and wood quality analysis efficiency and accuracy.Method: Used 582 labeled images with dynamic augmentation to train and evaluate five deep learning models. Tested generalizability on an oak dataset from Oregon State University, and conducted exploratory analysis by training worst-performing model on additional 64 images.
Result: Swin Transformer achieved highest accuracy (0.94) with excellent fine segmentation. YOLOv9 performed well for bounding boxes but struggled with boundaries. Mask R-CNN improved from 0.45 to 0.80 IoU with NMS. Generalizability testing showed varying performance across models.
Conclusion: Deep learning shows strong potential for automated pith detection, with model selection depending on dataset characteristics and application needs. Swin Transformer performed best overall, while techniques like NMS can significantly improve underperforming models.
Abstract: Pith detection in tree cross-sections is essential for forestry and wood quality analysis but remains a manual, error-prone task. This study evaluates deep learning models – YOLOv9, U-Net, Swin Transformer, DeepLabV3, and Mask R-CNN – to automate the process efficiently. A dataset of 582 labeled images was dynamically augmented to improve generalization. Swin Transformer achieved the highest accuracy (0.94), excelling in fine segmentation. YOLOv9 performed well for bounding box detection but struggled with boundary precision. U-Net was effective for structured patterns, while DeepLabV3 captured multi-scale features with slight boundary imprecision. Mask R-CNN initially underperformed due to overlapping detections, but applying Non-Maximum Suppression (NMS) improved its IoU from 0.45 to 0.80. Generalizability was next tested using an oak dataset of 11 images from Oregon State University’s Tree Ring Lab. Additionally, for exploratory analysis purposes, an additional dataset of 64 labeled tree cross-sections was used to train the worst-performing model to see if this would improve its performance generalizing to the unseen oak dataset. Key challenges included tensor mismatches and boundary inconsistencies, addressed through hyperparameter tuning and augmentation. Our results highlight deep learning’s potential for tree cross-section pith detection, with model choice depending on dataset characteristics and application needs.
[221] XAI-Driven Skin Disease Classification: Leveraging GANs to Augment ResNet-50 Performance
Kim Gerard A. Villanueva, Priyanka Kumar
Main category: cs.CV
TL;DR: Proposed a trustworthy CAD system for multi-class skin lesion diagnosis using DCGANs for data augmentation, fine-tuned ResNet-50 classifier, and XAI techniques (LIME/SHAP) for interpretability, achieving 92.50% accuracy and 98.82% Macro-AUC.
Details
Motivation: Address limitations in skin lesion diagnosis: subjective methods, data imbalance in datasets like HAM10000, and the "black box" nature of DL models that hinder clinical trust and deployment.Method: 1) Use DCGANs for per-class data augmentation to resolve class imbalance; 2) Train fine-tuned ResNet-50 classifier on augmented dataset for 7-class classification; 3) Integrate LIME and SHAP XAI techniques to provide transparency and verify predictions are based on clinically relevant features.
Result: Achieved 92.50% overall accuracy and 98.82% Macro-AUC, outperforming prior benchmark architectures. Successfully validated a verifiable framework combining high performance with clinical interpretability.
Conclusion: The work demonstrates a trustworthy CAD system that addresses both performance and interpretability challenges. Future research should focus on enhancing discrimination for critical categories like Melanoma NOS (F1-Score: 0.8602) to further improve clinical utility.
Abstract: Accurate and timely diagnosis of multi-class skin lesions is hampered by subjective methods, inherent data imbalance in datasets like HAM10000, and the “black box” nature of Deep Learning (DL) models. This study proposes a trustworthy and highly accurate Computer-Aided Diagnosis (CAD) system to overcome these limitations. The approach utilizes Deep Convolutional Generative Adversarial Networks (DCGANs) for per class data augmentation to resolve the critical class imbalance problem. A fine-tuned ResNet-50 classifier is then trained on the augmented dataset to classify seven skin disease categories. Crucially, LIME and SHAP Explainable AI (XAI) techniques are integrated to provide transparency by confirming that predictions are based on clinically relevant features like irregular morphology. The system achieved a high overall Accuracy of 92.50 % and a Macro-AUC of 98.82 %, successfully outperforming various prior benchmarked architectures. This work successfully validates a verifiable framework that combines high performance with the essential clinical interpretability required for safe diagnostic deployment. Future research should prioritize enhancing discrimination for critical categories, such as Melanoma NOS (F1-Score is 0.8602).
[222] Doppler-Enhanced Deep Learning: Improving Thyroid Nodule Segmentation with YOLOv5 Instance Segmentation
Mahmoud El Hussieni
Main category: cs.CV
TL;DR: YOLOv5 instance segmentation for thyroid nodules in ultrasound images, showing doppler images significantly improve performance across all model variants.
Details
Motivation: Thyroid cancer prevalence is increasing globally, requiring accurate segmentation of thyroid nodules as a critical first step for AI-assisted clinical decision support systems.Method: Used YOLOv5 instance segmentation algorithms (Nano, Small, Medium, Large, XLarge variants) on ultrasound images, evaluated across two dataset versions with and without doppler images.
Result: YOLOv5-Large achieved highest performance with 91% dice score and 0.87 mAP on dataset including doppler images. Doppler images significantly improved segmentation performance across all models (e.g., YOLOv5-Small improved from 79% to higher scores).
Conclusion: Instance segmentation with YOLOv5 provides effective real-time approach for thyroid nodule detection, with potential clinical applications in automated diagnostic systems, and doppler images (typically excluded by physicians) can significantly improve performance.
Abstract: The increasing prevalence of thyroid cancer globally has led to the development of various computer-aided detection methods. Accurate segmentation of thyroid nodules is a critical first step in the development of AI-assisted clinical decision support systems. This study focuses on instance segmentation of thyroid nodules using YOLOv5 algorithms on ultrasound images. We evaluated multiple YOLOv5 variants (Nano, Small, Medium, Large, and XLarge) across two dataset versions, with and without doppler images. The YOLOv5-Large algorithm achieved the highest performance with a dice score of 91% and mAP of 0.87 on the dataset including doppler images. Notably, our results demonstrate that doppler images, typically excluded by physicians, can significantly improve segmentation performance. The YOLOv5-Small model achieved 79% dice score when doppler images were excluded, while including them improved performance across all model variants. These findings suggest that instance segmentation with YOLOv5 provides an effective real-time approach for thyroid nodule detection, with potential clinical applications in automated diagnostic systems.
[223] Graph-Attention Network with Adversarial Domain Alignment for Robust Cross-Domain Facial Expression Recognition
Razieh Ghaedi, AmirReza BabaAhmadi, Reyer Zwiggelaar, Xinqi Fan, Nashid Alam
Main category: cs.CV
TL;DR: GAT-ADA is a hybrid framework for cross-domain facial expression recognition that combines ResNet-50 backbone with Graph Attention Network and adversarial domain alignment to address domain shift issues.
Details
Motivation: Cross-domain facial expression recognition is challenging due to severe domain shift between training and deployment data, requiring methods that can effectively adapt to new domains without labeled target data.Method: Proposes GAT-ADA: ResNet-50 backbone + batch-level Graph Attention Network (GAT) to model inter-sample relations via sparse ring graphs, combined with adversarial learning (Gradient Reversal Layer) and statistical alignment (CORAL and MMD) for domain adaptation.
Result: Achieves 74.39% mean cross-domain accuracy across multiple datasets. Specifically, reaches 98.0% accuracy on RAF-DB to FER2013 adaptation, representing ~36-point improvement over best baseline with same backbone and preprocessing.
Conclusion: GAT-ADA effectively addresses domain shift in facial expression recognition by combining graph attention for sample relations with adversarial and statistical domain alignment, achieving state-of-the-art performance.
Abstract: Cross-domain facial expression recognition (CD-FER) remains difficult due to severe domain shift between training and deployment data. We propose Graph-Attention Network with Adversarial Domain Alignment (GAT-ADA), a hybrid framework that couples a ResNet-50 as backbone with a batch-level Graph Attention Network (GAT) to model inter-sample relations under shift. Each mini-batch is cast as a sparse ring graph so that attention aggregates cross-sample cues that are informative for adaptation. To align distributions, GAT-ADA combines adversarial learning via a Gradient Reversal Layer (GRL) with statistical alignment using CORAL and MMD. GAT-ADA is evaluated under a standard unsupervised domain adaptation protocol: training on one labeled source (RAF-DB) and adapting to multiple unlabeled targets (CK+, JAFFE, SFEW 2.0, FER2013, and ExpW). GAT-ADA attains 74.39% mean cross-domain accuracy. On RAF-DB to FER2013, it reaches 98.0% accuracy, corresponding to approximately a 36-point improvement over the best baseline we re-implemented with the same backbone and preprocessing.
[224] MambaScope: Coarse-to-Fine Scoping for Efficient Vision Mamba
Shanhui Liu, Rui Xu, Yunke Wang
Main category: cs.CV
TL;DR: CF-ViM is an adaptive coarse-to-fine framework for Vision Mamba that reduces computation by processing simple images at coarse resolution and only refining complex regions when needed, achieving better accuracy-efficiency trade-off than existing token reduction methods.
Details
Motivation: Vision Mamba's efficiency is constrained by input token count. Existing token reduction methods (pruning/merging) cause information loss by uniformly applying fine-grained processing to all images, regardless of visual complexity. Simple images can be handled coarsely while only complex ones need refinement.Method: CF-ViM first performs coarse-grained inference using large patches to reduce token length. When prediction confidence is low, selected regions are re-processed at finer resolution to recover visual details. This dynamic resolution assignment adapts computation to image complexity.
Result: Experiments on ImageNet show CF-ViM outperforms both baseline Vision Mamba and state-of-the-art token reduction techniques in accuracy and efficiency.
Conclusion: CF-ViM provides an adaptive framework that allocates computation according to image complexity, ensuring efficient processing without compromising essential visual information, making Vision Mamba more practical for real-world applications.
Abstract: Vision Mamba has emerged as a promising and efficient alternative to Vision Transformers, yet its efficiency remains fundamentally constrained by the number of input tokens. Existing token reduction approaches typically adopt token pruning or merging to reduce computation. However, they inherently lead to information loss, as they discard or compress token representations. This problem is exacerbated when applied uniformly to fine-grained token representations across all images, regardless of visual complexity. We observe that not all inputs require fine-grained processing. Simple images can be effectively handled at coarse resolution, while only complex ones may warrant refinement. Based on this insight, we propose \textit{Coarse-to-Fine Vision Mamba (CF-ViM)}, an adaptive framework for efficient inference. CF-ViM first performs coarse-grained inference by dividing the input image into large patches, significantly reducing the token length and computation. When the model’s prediction confidence is low, selected regions are re-processed at a finer resolution to recover critical visual details with minimal additional cost. This dynamic resolution assignment strategy allows CF-ViM to allocate computation adaptively according to image complexity, ensuring efficient processing without compromising essential visual information. Experiments on ImageNet demonstrate that CF-ViM outperforms both the baseline Vision Mamba and state-of-the-art token reduction techniques in terms of accuracy and efficiency.
[225] Realistic Handwritten Multi-Digit Writer (MDW) Number Recognition Challenges
Kiri L. Wagstaff
Main category: cs.CV
TL;DR: Researchers create realistic multi-digit writer benchmarks using NIST data to better evaluate number recognition systems for real-world applications like ZIP codes and check amounts.
Details
Motivation: Current isolated digit classification benchmarks don't reflect real-world scenarios where numbers appear as multiple digits written by the same person (e.g., ZIP codes, check amounts, appointment times). There's a need for more realistic evaluation data sets.Method: Leverage writer information from NIST digit images to create multi-digit writer (MDW) benchmark data sets that simulate real-world scenarios where multiple digits come from the same writer.
Result: Classifiers that perform well on isolated digits often perform poorly on multi-digit number recognition. The MDW benchmarks reveal this performance gap and provide task-specific metrics that better align with real-world impact.
Conclusion: Additional advances beyond isolated digit classification are needed for real number recognition problems. MDW benchmarks enable development of methods that leverage task-specific knowledge to achieve better performance than individual digit classification approaches.
Abstract: Isolated digit classification has served as a motivating problem for decades of machine learning research. In real settings, numbers often occur as multiple digits, all written by the same person. Examples include ZIP Codes, handwritten check amounts, and appointment times. In this work, we leverage knowledge about the writers of NIST digit images to create more realistic benchmark multi-digit writer (MDW) data sets. As expected, we find that classifiers may perform well on isolated digits yet do poorly on multi-digit number recognition. If we want to solve real number recognition problems, additional advances are needed. The MDW benchmarks come with task-specific performance metrics that go beyond typical error calculations to more closely align with real-world impact. They also create opportunities to develop methods that can leverage task-specific knowledge to improve performance well beyond that of individual digit classification methods.
[226] Dynamic-eDiTor: Training-Free Text-Driven 4D Scene Editing with Multimodal Diffusion Transformer
Dong In Lee, Hyungjun Doh, Seunggeun Chi, Runlin Duan, Sangpil Kim, Karthik Ramani
Main category: cs.CV
TL;DR: Dynamic-eDiTor: A training-free text-driven 4D scene editing framework using MM-DiT and 4DGS that achieves consistent multi-view video editing without additional training.
Details
Motivation: Text-driven 4D scene editing remains challenging due to difficulties in maintaining multi-view and temporal consistency across space and time. Existing methods using 2D diffusion models edit frames independently, leading to motion distortion, geometric drift, and incomplete editing.Method: Leverages Multimodal Diffusion Transformer (MM-DiT) and 4D Gaussian Splatting (4DGS). Uses Spatio-Temporal Sub-Grid Attention (STGA) for locally consistent cross-view and temporal fusion, and Context Token Propagation (CTP) for global propagation via token inheritance and optical-flow-guided token replacement. Directly optimizes pre-trained source 4DGS without additional training.
Result: Extensive experiments on multi-view video dataset DyNeRF demonstrate superior editing fidelity and both multi-view and temporal consistency compared to prior approaches.
Conclusion: Dynamic-eDiTor enables seamless, globally consistent text-driven 4D scene editing without requiring additional training, addressing key challenges in maintaining consistency across space and time during editing.
Abstract: Recent progress in 4D representations, such as Dynamic NeRF and 4D Gaussian Splatting (4DGS), has enabled dynamic 4D scene reconstruction. However, text-driven 4D scene editing remains under-explored due to the challenge of ensuring both multi-view and temporal consistency across space and time during editing. Existing studies rely on 2D diffusion models that edit frames independently, often causing motion distortion, geometric drift, and incomplete editing. We introduce Dynamic-eDiTor, a training-free text-driven 4D editing framework leveraging Multimodal Diffusion Transformer (MM-DiT) and 4DGS. This mechanism consists of Spatio-Temporal Sub-Grid Attention (STGA) for locally consistent cross-view and temporal fusion, and Context Token Propagation (CTP) for global propagation via token inheritance and optical-flow-guided token replacement. Together, these components allow Dynamic-eDiTor to perform seamless, globally consistent multi-view video without additional training and directly optimize pre-trained source 4DGS. Extensive experiments on multi-view video dataset DyNeRF demonstrate that our method achieves superior editing fidelity and both multi-view and temporal consistency prior approaches. Project page for results and code: https://di-lee.github.io/dynamic-eDiTor/
[227] Silhouette-based Gait Foundation Model
Dingqiang Ye, Chao Fan, Kartik Narayan, Bingzhe Wu, Chengwen Luo, Jianqiang Li, Vishal M. Patel
Main category: cs.CV
TL;DR: FoundationGait is the first scalable, self-supervised pretraining framework for gait understanding, achieving state-of-the-art performance across diverse gait tasks through large-scale pretraining on 12 datasets with 2M+ walking sequences.
Details
Motivation: Current gait models are small, narrowly designed, and fail to scale or generalize. The paper aims to overcome two barriers: scalability (why gait models don't follow scaling laws) and generalization (whether one model can serve diverse gait tasks traditionally studied in isolation).Method: Introduces FoundationGait, a scalable self-supervised pretraining framework with largest version having 0.13B parameters, pretrained on 12 public gait datasets comprising over 2M walking sequences. Uses unified framework for diverse gait tasks.
Result: Achieves 48.0% zero-shot rank-1 accuracy on challenging in-the-wild Gait3D dataset (1,000 test subjects) and 64.5% on largest in-the-lab OU-MVLP dataset (5,000+ test subjects). Performs robustly across wide spectrum of gait datasets, conditions, tasks, and input modalities.
Conclusion: FoundationGait sets a new milestone in robust gait recognition, demonstrating that scalable, unified foundation models can effectively address both scalability and generalization challenges in gait understanding.
Abstract: Gait patterns play a critical role in human identification and healthcare analytics, yet current progress remains constrained by small, narrowly designed models that fail to scale or generalize. Building a unified gait foundation model requires addressing two longstanding barriers: (a) Scalability. Why have gait models historically failed to follow scaling laws? (b) Generalization. Can one model serve the diverse gait tasks that have traditionally been studied in isolation? We introduce FoundationGait, the first scalable, self-supervised pretraining framework for gait understanding. Its largest version has nearly 0.13 billion parameters and is pretrained on 12 public gait datasets comprising over 2 million walking sequences. Extensive experiments demonstrate that FoundationGait, with or without fine-tuning, performs robustly across a wide spectrum of gait datasets, conditions, tasks (e.g., human identification, scoliosis screening, depression prediction, and attribute estimation), and even input modality. Notably, it achieves 48.0% zero-shot rank-1 accuracy on the challenging in-the-wild Gait3D dataset (1,000 test subjects) and 64.5% on the largest in-the-lab OU-MVLP dataset (5,000+ test subjects), setting a new milestone in robust gait recognition. Coming code and model: https://github.com/ShiqiYu/OpenGait.
[228] Affordance-First Decomposition for Continual Learning in Video-Language Understanding
Mengzhu Xu, Hanzhi Liu, Ningkang Peng, Qianyu Chen, Canran Xiao
Main category: cs.CV
TL;DR: AFD introduces an affordance-first decomposition approach for continual video-language learning that explicitly separates stable affordance tokens from query-specific adaptation, achieving SOTA with minimal forgetting while respecting memory/privacy constraints.
Details
Motivation: Current continual learning approaches for video-language understanding fail to properly distinguish what should remain stable versus what should adapt, use static routing/capacity, or require replaying past videos, which violates memory and privacy constraints.Method: AFD maps videos to slowly varying affordance tokens forming a shared time-aligned substrate, stabilized via weak alignment and teacher consistency. A lightweight query-routed, conflict-aware scheduler concentrates adaptation and grows capacity only when needed, using question-only replay for training.
Result: Achieves state-of-the-art across protocols: 51.6% average accuracy with -1.8% forgetting on domain-incremental VideoQA, ViLCo R@1@0.5 of 29.6% (MQ) and 20.7% (NLQ) with 18.4% stAP@0.25 (VQ), and 39.5% accuracy with -1.6% forgetting on time-incremental iVQA.
Conclusion: AFD provides an explicit, interpretable split between a stable interaction-centered substrate and targeted adaptation, offering a principled solution for continual video-language learning under realistic constraints.
Abstract: Continual learning for video–language understanding is increasingly important as models face non-stationary data, domains, and query styles, yet prevailing solutions blur what should stay stable versus what should adapt, rely on static routing/capacity, or require replaying past videos. We aim to explicitly specify where stability lives and where plasticity should be focused under realistic memory and privacy constraints. We introduce Affordance-First Decomposition (AFD): videos are mapped to slowly varying affordance tokens that form a shared, time-aligned substrate, while a lightweight, query-routed, conflict-aware scheduler concentrates adaptation and grows capacity only when needed. The substrate is stabilized via weak alignment and teacher consistency, and training uses question-only replay. AFD achieves state-of-the-art across protocols: 51.6% average accuracy with -1.8% forgetting on domain-incremental VideoQA, ViLCo R@1@0.5 of 29.6% (MQ) and 20.7% (NLQ) with 18.4% stAP@0.25 (VQ), and 39.5% accuracy with -1.6% forgetting on time-incremental iVQA. Overall, AFD offers an explicit, interpretable split between a stable interaction-centered substrate and targeted adaptation.
[229] CAR-Net: A Cascade Refinement Network for Rotational Motion Deblurring under Angle Information Uncertainty
Ka Chung Lai, Ahmet Cetinkaya
Main category: cs.CV
TL;DR: CAR-net is a cascade refinement network for deblurring images with rotational motion blur, designed for semi-blind scenarios where only noisy blur angle information is available.
Details
Motivation: The paper addresses the challenge of deblurring images affected by rotational motion blur, particularly in semi-blind scenarios where only noisy information about the blur angle is available, requiring robust handling of parameter uncertainty.Method: The architecture uses a progressive refinement process starting with frequency-domain inversion for initial deblurring, followed by multiple refinement stages that predict and apply residual corrections to suppress artifacts and restore details. It includes an optional angle detection module for end-to-end training.
Result: The method demonstrates efficiency through experiments on both synthetic and real-life images, with code, models, and dataset links made publicly available.
Conclusion: CAR-net provides an effective solution for rotational motion blur deblurring in semi-blind scenarios, combining frequency-domain initialization with cascade refinement and optional angle detection for robust performance.
Abstract: We propose a new neural network architecture called CAR-net (CAscade Refinement Network) to deblur images that are subject to rotational motion blur. Our architecture is specifically designed for the semi-blind scenarios where only noisy information of the rotational motion blur angle is available. The core of our approach is progressive refinement process that starts with an initial deblurred estimate obtained from frequency-domain inversion; A series of refinement stages take the current deblurred image to predict and apply residual correction to the current estimate, progressively suppressing artifacts and restoring fine details. To handle parameter uncertainty, our architecture accommodates an optional angle detection module which can be trained end-to-end with refinement modules. We provide a detailed description of our architecture and illustrate its efficiency through experiments using both synthetic and real-life images. Our code and model as well as the links to the datasets are available at https://github.com/tony123105/CAR-Net
[230] Optimizing LVLMs with On-Policy Data for Effective Hallucination Mitigation
Chengzhi Yu, Yifan Xu, Yifan Chen, Wenyi Zhang
Main category: cs.CV
TL;DR: The paper proposes a new method to reduce hallucinations in large vision-language models by using on-policy data with binary hallucination annotations and iterative DPO with dynamic sample reweighting.
Details
Motivation: Current LVLMs suffer from hallucination problems, and existing annotation methods for preference data introduce additional hallucinations into training samples, which can actually reinforce the model's hallucination patterns rather than mitigate them.Method: 1) Use on-policy data instead of off-policy data for better hallucination mitigation; 2) Train a binary hallucination classifier to provide clean annotations; 3) Implement iterative Direct Preference Optimization (DPO) with dynamic sample reweighting scheme.
Result: Reduces hallucination rate of LLaVA-1.5-7B on MMHalBench by 50.8% and average hallucination rate on Object HalBench by 79.5%. Enables LLaVA-1.5-13B to surpass GPT-4V performance.
Conclusion: The proposed approach effectively mitigates hallucinations in LVLMs by leveraging clean on-policy data with binary annotations and robust iterative DPO, demonstrating superior performance over existing methods.
Abstract: Recently, large vision-language models (LVLMs) have risen to be a promising approach for multimodal tasks. However, principled hallucination mitigation remains a critical challenge.In this work, we first analyze the data generation process in LVLM hallucination mitigation and affirm that on-policy data significantly outperforms off-policy data, which thus calls for efficient and reliable preference annotation of on-policy data. We then point out that, existing annotation methods introduce additional hallucination in training samples, which may enhance the model’s hallucination patterns, to address this problem, we propose training a hallucination classifier giving binary annotations, which guarantee clean chosen samples for the subsequent alignment. To further harness of the power of on-policy data, we design a robust iterative direct preference optimization (DPO) algorithm adopting a dynamic sample reweighting scheme. We conduct comprehensive experiments on three benchmarks with comparison to 8 state-of-the-art baselines. In particular, our approach reduces the hallucination rate of LLaVA-1.5-7B on MMHalBench by 50.8% and the average hallucination rate on Object HalBench by 79.5%; more significantly, our method fully taps into the potential of open-source models, enabling LLaVA-1.5-13B to even surpass the performance of GPT-4V.
[231] Deep Learning-Based Computer Vision Models for Early Cancer Detection Using Multimodal Medical Imaging and Radiogenomic Integration Frameworks
Emmanuella Avwerosuoghene Oghenekaro
Main category: cs.CV
TL;DR: Deep learning models (CNNs, transformers, hybrid attention) analyze multimodal medical imaging to detect subtle cancer patterns, surpassing human radiological assessment. Integration with radiogenomics enables non-invasive prediction of tumor characteristics and treatment response.
Details
Motivation: Early cancer detection is critical for improving survival outcomes, as delayed diagnosis significantly reduces patient prognosis. Traditional radiological assessment has limitations in identifying subtle tissue abnormalities and tumor microenvironment variations.Method: Utilizes deep learning-based computer vision models including convolutional neural networks (CNNs), transformers, and hybrid attention architectures to automatically extract complex spatial, morphological, and temporal patterns from multimodal imaging data (MRI, CT, PET, mammography, histopathology, ultrasound). Integrates multimodal imaging with radiogenomics to link quantitative imaging features with genomics, transcriptomics, and epigenetic biomarkers.
Result: Models surpass traditional radiological assessment by identifying subtle tissue abnormalities and tumor microenvironment variations invisible to the human eye. Radiogenomic fusion enables prediction of tumor genotype, immune response, molecular subtypes, and treatment resistance without invasive biopsies.
Conclusion: Deep learning and radiogenomics represent a transformative paradigm for personalized oncology, enabling early cancer detection and non-invasive characterization of tumors through advanced analysis of multimodal imaging data.
Abstract: Early cancer detection remains one of the most critical challenges in modern healthcare, where delayed diagnosis significantly reduces survival outcomes. Recent advancements in artificial intelligence, particularly deep learning, have enabled transformative progress in medical imaging analysis. Deep learning-based computer vision models, such as convolutional neural networks (CNNs), transformers, and hybrid attention architectures, can automatically extract complex spatial, morphological, and temporal patterns from multimodal imaging data including MRI, CT, PET, mammography, histopathology, and ultrasound. These models surpass traditional radiological assessment by identifying subtle tissue abnormalities and tumor microenvironment variations invisible to the human eye. At a broader scale, the integration of multimodal imaging with radiogenomics linking quantitative imaging features with genomics, transcriptomics, and epigenetic biomarkers has introduced a new paradigm for personalized oncology. This radiogenomic fusion allows the prediction of tumor genotype, immune response, molecular subtypes, and treatment resistance without invasive biopsies.
[232] RS-ISRefiner: Towards Better Adapting Vision Foundation Models for Interactive Segmentation of Remote Sensing Images
Deliang Wang, Peng Liu
Main category: cs.CV
TL;DR: RS-ISRefiner is a click-based interactive image segmentation framework specifically designed for remote sensing images, addressing challenges like scale variations, irregular boundaries, and complex backgrounds through adapter-based tuning and hybrid attention mechanisms.
Details
Motivation: Existing interactive image segmentation methods designed for natural images struggle with remote sensing imagery due to scale variations, irregular boundaries, complex backgrounds, limited annotated data, and computational overhead. There's a need for specialized IIS frameworks for remote sensing applications.Method: Proposes RS-ISRefiner framework with: 1) Adapter-based tuning strategy to preserve Vision Foundation Model representations while learning remote sensing-specific features, 2) Hybrid attention mechanism combining convolutional local modeling with Transformer-based global reasoning for scale diversity and scene complexity, 3) Improved probability map modulation scheme incorporating historical user interactions for stable iterative refinement.
Result: Comprehensive experiments on six remote sensing datasets (iSAID, ISPRS Potsdam, SandBar, NWPU, LoveDA Urban, WHUBuilding) show RS-ISRefiner consistently outperforms state-of-the-art IIS methods in segmentation accuracy, efficiency, and interaction cost.
Conclusion: RS-ISRefiner demonstrates effectiveness and generalizability for high-quality instance segmentation in practical remote sensing scenarios, making it suitable for precise annotation tasks in remote sensing imagery.
Abstract: Interactive image segmentation(IIS) plays a critical role in generating precise annotations for remote sensing imagery, where objects often exhibit scale variations, irregular boundaries and complex backgrounds. However, existing IIS methods, primarily designed for natural images, struggle to generalize to remote sensing domains due to limited annotated data and computational overhead. To address these challenges, we proposed RS-ISRefiner, a novel click-based IIS framework tailored for remote sensing images. The framework employs an adapter-based tuning strategy that preserves the general representations of Vision Foundation Models while enabling efficient learning of remote sensing-specific spatial and boundary characteristics. A hybrid attention mechanism integrating convolutional local modeling with Transformer-based global reasoning enhances robustness against scale diversity and scene complexity. Furthermore, an improved probability map modulation scheme effectively incorporates historical user interactions, yielding more stable iterative refinement and higher boundary fidelity. Comprehensive experiments on six remote sensing datasets, including iSAID, ISPRS Potsdam, SandBar, NWPU, LoveDA Urban and WHUBuilding, demonstrate that RS-ISRefiner consistently outperforms state-of-the-art IIS methods in terms of segmentation accuracy, efficiency and interaction cost. These results confirm the effectiveness and generalizability of our framework, making it highly suitable for high-quality instance segmentation in practical remote sensing scenarios.
[233] Generalized Medical Phrase Grounding
Wenjun Zhang, Shekhar S. Chandra, Aaron Nicolson
Main category: cs.CV
TL;DR: The paper introduces MedGrounder, a generalized medical phrase grounding model that maps radiology report sentences to zero, one, or multiple scored image regions, addressing limitations of existing single-bounding-box approaches.
Details
Motivation: Existing medical phrase grounding systems follow the referring expression comprehension paradigm and return exactly one bounding box per phrase, but real radiology reports often contain multi-region findings, non-diagnostic text, non-groundable phrases (negations, normal anatomy descriptions), violating this assumption.Method: Introduces MedGrounder, the first generalized medical phrase grounding model with two-stage training: pre-training on report sentence-anatomy box alignment datasets, then fine-tuning on report sentence-human annotated box datasets.
Result: MedGrounder achieves strong zero-shot transfer and outperforms REC-style and grounded report generation baselines on multi-region and non-groundable phrases, while using far fewer human box annotations. Can be composed with existing report generators to produce grounded reports without retraining.
Conclusion: The paper successfully reformulates medical phrase grounding as a generalized task (GMPG) and demonstrates MedGrounder’s effectiveness in handling real-world radiology report complexities with improved performance and reduced annotation requirements.
Abstract: Medical phrase grounding (MPG) maps textual descriptions of radiological findings to corresponding image regions. These grounded reports are easier to interpret, especially for non-experts. Existing MPG systems mostly follow the referring expression comprehension (REC) paradigm and return exactly one bounding box per phrase. Real reports often violate this assumption. They contain multi-region findings, non-diagnostic text, and non-groundable phrases, such as negations or descriptions of normal anatomy. Motivated by this, we reformulate the task as generalised medical phrase grounding (GMPG), where each sentence is mapped to zero, one, or multiple scored regions. To realise this formulation, we introduce the first GMPG model: MedGrounder. We adopted a two-stage training regime: pre-training on report sentence–anatomy box alignment datasets and fine-tuning on report sentence–human annotated box datasets. Experiments on PadChest-GR and MS-CXR show that MedGrounder achieves strong zero-shot transfer and outperforms REC-style and grounded report generation baselines on multi-region and non-groundable phrases, while using far fewer human box annotations. Finally, we show that MedGrounder can be composed with existing report generators to produce grounded reports without retraining the generator.
[234] TrajDiff: End-to-end Autonomous Driving without Perception Annotation
Xingtai Gui, Jianbo Zhao, Wencheng Han, Jikai Wang, Jiahao Gong, Feiyang Tan, Cheng-zhong Xu, Jianbing Shen
Main category: cs.CV
TL;DR: TrajDiff is a perception annotation-free end-to-end autonomous driving framework that uses trajectory-oriented BEV conditioned diffusion to generate driving policies directly from raw sensor inputs without manual perception labels.
Details
Motivation: Current end-to-end autonomous driving systems rely on expensive manual perception annotations. The authors aim to develop a fully perception annotation-free planning paradigm to reduce annotation costs while maintaining performance.Method: Proposes TrajDiff: 1) Uses only raw sensor inputs and future trajectories to construct Gaussian BEV heatmap targets capturing driving modalities; 2) Designs trajectory-oriented BEV encoder without perceptual supervision; 3) Introduces Trajectory-oriented BEV Diffusion Transformer (TB-DiT) that leverages ego-state and predicted TrajBEV features to generate diverse trajectories without handcrafted motion priors.
Result: Achieves 87.5 PDMS on NAVSIM benchmark, state-of-the-art among annotation-free methods. With data scaling, improves to 88.5 PDMS, comparable to advanced perception-based approaches.
Conclusion: TrajDiff successfully establishes a perception annotation-free generative method for end-to-end autonomous driving that achieves competitive performance while enabling exploration of data scaling benefits in annotation-free settings.
Abstract: End-to-end autonomous driving systems directly generate driving policies from raw sensor inputs. While these systems can extract effective environmental features for planning, relying on auxiliary perception tasks, developing perception annotation-free planning paradigms has become increasingly critical due to the high cost of manual perception annotation. In this work, we propose TrajDiff, a Trajectory-oriented BEV Conditioned Diffusion framework that establishes a fully perception annotation-free generative method for end-to-end autonomous driving. TrajDiff requires only raw sensor inputs and future trajectory, constructing Gaussian BEV heatmap targets that inherently capture driving modalities. We design a simple yet effective trajectory-oriented BEV encoder to extract the TrajBEV feature without perceptual supervision. Furthermore, we introduce Trajectory-oriented BEV Diffusion Transformer (TB-DiT), which leverages ego-state information and the predicted TrajBEV features to directly generate diverse yet plausible trajectories, eliminating the need for handcrafted motion priors. Beyond architectural innovations, TrajDiff enables exploration of data scaling benefits in the annotation-free setting. Evaluated on the NAVSIM benchmark, TrajDiff achieves 87.5 PDMS, establishing state-of-the-art performance among all annotation-free methods. With data scaling, it further improves to 88.5 PDMS, which is comparable to advanced perception-based approaches. Our code and model will be made publicly available.
[235] Multi-GRPO: Multi-Group Advantage Estimation for Text-to-Image Generation with Tree-Based Trajectories and Multiple Rewards
Qiang Lyu, Zicong Chen, Chongxiao Wang, Haolin Shi, Shibo Gao, Ran Piao, Youwei Zeng, Jianlou Si, Fei Ding, Jing Li, Chun Pong Lau, Weiqiang Wang
Main category: cs.CV
TL;DR: Multi-GRPO improves text-to-image model alignment by addressing credit assignment and reward-mixing issues in Group Relative Policy Optimization through temporal and reward-based grouping mechanisms.
Details
Motivation: Existing GRPO-based methods for aligning text-to-image models suffer from two critical limitations: (1) shared credit assignment where trajectory-level advantages from sparse terminal rewards are uniformly applied across timesteps, failing to accurately estimate early denoising steps' potential, and (2) reward-mixing where predefined weights for combining multi-objective rewards with mismatched scales lead to unstable gradients and conflicting updates.Method: Proposes Multi-GRPO with two orthogonal grouping mechanisms: (1) tree-based trajectories inspired by Monte Carlo Tree Search that create temporal groups by branching at early denoising steps, enabling accurate advantage estimation for early steps via descendant leaves while amortizing computation through shared prefixes, and (2) reward-based grouping that computes advantages for each reward function independently before aggregation to disentangle conflicting signals. Also introduces OCR-Color-10 dataset for multi-objective evaluation.
Result: Multi-GRPO achieves superior stability and alignment performance across single-reward PickScore-25k and multi-objective OCR-Color-10 benchmarks, effectively balancing conflicting objectives.
Conclusion: Multi-GRPO addresses critical limitations of existing GRPO methods through innovative temporal and reward-based grouping mechanisms, demonstrating improved stability and alignment performance for text-to-image models across both single and multi-objective settings.
Abstract: Recently, Group Relative Policy Optimization (GRPO) has shown promising potential for aligning text-to-image (T2I) models, yet existing GRPO-based methods suffer from two critical limitations. (1) \textit{Shared credit assignment}: trajectory-level advantages derived from group-normalized sparse terminal rewards are uniformly applied across timesteps, failing to accurately estimate the potential of early denoising steps with vast exploration spaces. (2) \textit{Reward-mixing}: predefined weights for combining multi-objective rewards (e.g., text accuracy, visual quality, text color)–which have mismatched scales and variances–lead to unstable gradients and conflicting updates. To address these issues, we propose \textbf{Multi-GRPO}, a multi-group advantage estimation framework with two orthogonal grouping mechanisms. For better credit assignment, we introduce tree-based trajectories inspired by Monte Carlo Tree Search: branching trajectories at selected early denoising steps naturally forms \emph{temporal groups}, enabling accurate advantage estimation for early steps via descendant leaves while amortizing computation through shared prefixes. For multi-objective optimization, we introduce \emph{reward-based grouping} to compute advantages for each reward function \textit{independently} before aggregation, disentangling conflicting signals. To facilitate evaluation of multiple objective alignment, we curate \textit{OCR-Color-10}, a visual text rendering dataset with explicit color constraints. Across the single-reward \textit{PickScore-25k} and multi-objective \textit{OCR-Color-10} benchmarks, Multi-GRPO achieves superior stability and alignment performance, effectively balancing conflicting objectives. Code will be publicly available at \href{https://github.com/fikry102/Multi-GRPO}{https://github.com/fikry102/Multi-GRPO}.
[236] Joint Multi-scale Gated Transformer and Prior-guided Convolutional Network for Learned Image Compression
Zhengxin Chen, Xiaohai He, Tingrong Zhang, Shuhua Xiong, Chao Ren
Main category: cs.CV
TL;DR: Proposes MGTPCN, a learned image compression method combining novel prior-guided convolution (PGConv) for local features and multi-scale gated transformer (MGT) for non-local features, achieving state-of-the-art performance with better complexity trade-off.
Details
Motivation: Learned image compression methods have surpassed traditional codecs like VVC, largely due to powerful nonlinear transform coding. The representation capabilities of convolutional layers and Swin-T blocks are crucial for this, but vanilla convolution needs better local feature extraction and Swin-T blocks need improved non-local feature extraction.Method: 1) PGConv: Uses asymmetric convolutions to strengthen skeleton elements and difference convolutions to extract high-frequency information, with re-parameterization to reduce complexity. 2) MGT: Uses dilated window-based multi-head self-attention with different dilation rates and depth-wise convolutions with different kernel sizes for multi-scale features, plus gate mechanism for enhanced non-linearity. 3) MGTPCN: Combines both components into a joint network for learned image compression.
Result: Experimental results show MGTPCN surpasses state-of-the-art algorithms with a better trade-off between performance and computational complexity.
Conclusion: The proposed MGTPCN framework effectively improves learned image compression by enhancing both local feature extraction (via PGConv) and non-local feature extraction (via MGT), achieving superior performance with manageable complexity.
Abstract: Recently, learned image compression methods have made remarkable achievements, some of which have outperformed the traditional image codec VVC. The advantages of learned image compression methods over traditional image codecs can be largely attributed to their powerful nonlinear transform coding. Convolutional layers and shifted window transformer (Swin-T) blocks are the basic units of neural networks, and their representation capabilities play an important role in nonlinear transform coding. In this paper, to improve the ability of the vanilla convolution to extract local features, we propose a novel prior-guided convolution (PGConv), where asymmetric convolutions (AConvs) and difference convolutions (DConvs) are introduced to strengthen skeleton elements and extract high-frequency information, respectively. A re-parameterization strategy is also used to reduce the computational complexity of PGConv. Moreover, to improve the ability of the Swin-T block to extract non-local features, we propose a novel multi-scale gated transformer (MGT), where dilated window-based multi-head self-attention blocks with different dilation rates and depth-wise convolution layers with different kernel sizes are used to extract multi-scale features, and a gate mechanism is introduced to enhance non-linearity. Finally, we propose a novel joint Multi-scale Gated Transformer and Prior-guided Convolutional Network (MGTPCN) for learned image compression. Experimental results show that our MGTPCN surpasses state-of-the-art algorithms with a better trade-off between performance and complexity.
[237] Generative Adversarial Gumbel MCTS for Abstract Visual Composition Generation
Zirui Zhao, Boye Niu, David Hsu, Wee Sun Lee
Main category: cs.CV
TL;DR: A constraint-guided framework combining geometric reasoning with neural semantics for abstract visual composition, using AlphaGo-style search and adversarial reward refinement to generate feasible, semantically-aligned geometric structures.
Details
Motivation: Abstract visual composition with geometric primitives is challenging due to combinatorial placement choices, limited data, and discrete feasibility constraints, creating sparse solution manifolds ill-suited to purely statistical pixel-space generators.Method: Combines explicit geometric reasoning with neural semantics using AlphaGo-style search to enforce feasibility, with a fine-tuned vision-language model scoring semantic alignment as reward signals. Uses policy network as heuristic in Monte-Carlo Tree Search, fine-tuned via search-generated plans, and adversarial reward refinement inspired by GANs.
Result: In the Tangram Assembly task, the approach yields higher validity and semantic fidelity than diffusion and auto-regressive baselines, especially as constraints tighten.
Conclusion: The constraint-guided framework successfully addresses the challenges of abstract visual composition by integrating geometric reasoning with neural semantics, outperforming purely statistical methods in generating feasible and semantically-aligned geometric structures.
Abstract: We study abstract visual composition, in which identity is primarily determined by the spatial configuration and relations among a small set of geometric primitives (e.g., parts, symmetry, topology). They are invariant primarily to texture and photorealistic detail. Composing such structures from fixed components under geometric constraints and vague goal specification (such as text) is non-trivial due to combinatorial placement choices, limited data, and discrete feasibility (overlap-free, allowable orientations), which create a sparse solution manifold ill-suited to purely statistical pixel-space generators. We propose a constraint-guided framework that combines explicit geometric reasoning with neural semantics. An AlphaGo-style search enforces feasibility, while a fine-tuned vision-language model scores semantic alignment as reward signals. Our algorithm uses a policy network as a heuristic in Monte-Carlo Tree Search and fine-tunes the network via search-generated plans. Inspired by the Generative Adversarial Network, we use the generated instances for adversarial reward refinement. Over time, the generation should approach the actual data more closely when the reward model cannot distinguish between generated instances and ground-truth. In the Tangram Assembly task, our approach yields higher validity and semantic fidelity than diffusion and auto-regressive baselines, especially as constraints tighten.
[238] Probabilistic Modeling of Multi-rater Medical Image Segmentation for Diversity and Personalization
Ke Liu, Shangde Gao, Yichao Fu, Shangqi Gao, Chunhua Shen
Main category: cs.CV
TL;DR: ProSeg is a probabilistic model for multi-rater medical image segmentation that simultaneously achieves diversification and personalization by modeling expert preferences and boundary ambiguity through latent variables and variational inference.
Details
Motivation: Medical image segmentation faces data uncertainty from ambiguous boundaries and inter-observer variability. Existing models either produce diverse segmentations lacking expert specificity or personalized outputs that merely replicate individual annotators, but not both simultaneously.Method: Introduces two latent variables to model expert annotation preferences and image boundary ambiguity. Uses variational inference to obtain conditional probabilistic distributions, allowing segmentation outputs to be generated by sampling from these distributions.
Result: Extensive experiments on nasopharyngeal carcinoma (NPC) and lung nodule (LIDC-IDRI) datasets demonstrate state-of-the-art performance, providing segmentation results that are both diverse and expert-personalized.
Conclusion: ProSeg successfully addresses the limitations of previous multi-rater segmentation models by enabling simultaneous diversification and personalization through probabilistic modeling of expert preferences and boundary ambiguity.
Abstract: Medical image segmentation is inherently influenced by data uncertainty, arising from ambiguous boundaries in medical scans and inter-observer variability in diagnosis. To address this challenge, previous works formulated the multi-rater medical image segmentation task, where multiple experts provide separate annotations for each image. However, existing models are typically constrained to either generate diverse segmentation that lacks expert specificity or to produce personalized outputs that merely replicate individual annotators. We propose Probabilistic modeling of multi-rater medical image Segmentation (ProSeg) that simultaneously enables both diversification and personalization. Specifically, we introduce two latent variables to model expert annotation preferences and image boundary ambiguity. Their conditional probabilistic distributions are then obtained through variational inference, allowing segmentation outputs to be generated by sampling from these distributions. Extensive experiments on both the nasopharyngeal carcinoma dataset (NPC) and the lung nodule dataset (LIDC-IDRI) demonstrate that our ProSeg achieves a new state-of-the-art performance, providing segmentation results that are both diverse and expert-personalized. Code can be found in https://github.com/AI4MOL/ProSeg.
[239] Charts Are Not Images: On the Challenges of Scientific Chart Editing
Shawn Li, Ryan Rossi, Sungchul Kim, Sunav Choudhary, Franck Dernoncourt, Puneet Mathur, Zhengzhong Tu, Yue Zhao
Main category: cs.CV
TL;DR: FigEdit is a large-scale benchmark for scientific figure editing that addresses the fundamental mismatch between pixel-level image editing approaches and the structured nature of charts, revealing limitations of current generative models.
Details
Motivation: Current generative models treat chart editing as pixel manipulation, but charts are visual representations of structured data governed by graphical grammar. This fundamental mismatch requires a structure-aware approach to chart editing.Method: The authors introduce FigEdit, a benchmark with over 30,000 samples covering 10 chart types and complex editing instructions. It’s organized into five progressively challenging tasks: single edits, multi edits, conversational edits, visual-guidance-based edits, and style transfer.
Result: Evaluation of state-of-the-art models shows poor performance on scientific figures, as they fail to handle structured transformations. Traditional metrics (SSIM, PSNR) are inadequate for capturing semantic correctness of chart edits.
Conclusion: FigEdit demonstrates limitations of pixel-level manipulation and provides a foundation for developing structure-aware models. The benchmark enables systematic progress in figure editing and encourages research on models that understand both visual and semantic layers of charts.
Abstract: Generative models, such as diffusion and autoregressive approaches, have demonstrated impressive capabilities in editing natural images. However, applying these tools to scientific charts rests on a flawed assumption: a chart is not merely an arrangement of pixels but a visual representation of structured data governed by a graphical grammar. Consequently, chart editing is not a pixel-manipulation task but a structured transformation problem. To address this fundamental mismatch, we introduce \textit{FigEdit}, a large-scale benchmark for scientific figure editing comprising over 30,000 samples. Grounded in real-world data, our benchmark is distinguished by its diversity, covering 10 distinct chart types and a rich vocabulary of complex editing instructions. The benchmark is organized into five distinct and progressively challenging tasks: single edits, multi edits, conversational edits, visual-guidance-based edits, and style transfer. Our evaluation of a range of state-of-the-art models on this benchmark reveals their poor performance on scientific figures, as they consistently fail to handle the underlying structured transformations required for valid edits. Furthermore, our analysis indicates that traditional evaluation metrics (e.g., SSIM, PSNR) have limitations in capturing the semantic correctness of chart edits. Our benchmark demonstrates the profound limitations of pixel-level manipulation and provides a robust foundation for developing and evaluating future structure-aware models. By releasing \textit{FigEdit} (https://github.com/adobe-research/figure-editing), we aim to enable systematic progress in structure-aware figure editing, provide a common ground for fair comparison, and encourage future research on models that understand both the visual and semantic layers of scientific charts.
[240] Seeing the Wind from a Falling Leaf
Zhiyuan Gao, Jiageng Mao, Hong-Xing Yu, Haozhe Lou, Emily Yue-Ting Jia, Jernej Barbic, Jiajun Wu, Yue Wang
Main category: cs.CV
TL;DR: The paper presents an end-to-end differentiable inverse graphics framework that recovers invisible physical forces (like wind fields) from video observations by jointly modeling object geometry, physical properties, and interactions.
Details
Motivation: Current computer vision focuses on modeling motions from videos but ignores the underlying physical interactions that cause those motions. The paper aims to bridge the gap between vision and physics by recovering the invisible forces behind observed motions.Method: An end-to-end differentiable inverse graphics framework that jointly models object geometry, physical properties, and interactions directly from videos. Through backpropagation, the approach enables recovery of force representations from object motions.
Result: The method successfully infers plausible force fields from both synthetic and real-world videos, demonstrating the ability to recover invisible physical forces from visual observations.
Conclusion: The approach shows potential for physics-based video generation and editing, and represents a step toward understanding and modeling the physical processes behind pixels, bridging vision and physics.
Abstract: A longstanding goal in computer vision is to model motions from videos, while the representations behind motions, i.e. the invisible physical interactions that cause objects to deform and move, remain largely unexplored. In this paper, we study how to recover the invisible forces from visual observations, e.g., estimating the wind field by observing a leaf falling to the ground. Our key innovation is an end-to-end differentiable inverse graphics framework, which jointly models object geometry, physical properties, and interactions directly from videos. Through backpropagation, our approach enables the recovery of force representations from object motions. We validate our method on both synthetic and real-world scenarios, and the results demonstrate its ability to infer plausible force fields from videos. Furthermore, we show the potential applications of our approach, including physics-based video generation and editing. We hope our approach sheds light on understanding and modeling the physical process behind pixels, bridging the gap between vision and physics. Please check more video results in our \href{https://chaoren2357.github.io/seeingthewind/}{project page}.
[241] The Outline of Deception: Physical Adversarial Attacks on Traffic Signs Using Edge Patches
Haojie Jia, Te Hu, Haowen Li, Long Jin, Chongshi Xin, Yuchi Yao, Jiarui Xiao
Main category: cs.CV
TL;DR: TESP-Attack: A stealth-aware adversarial patch method for traffic signs that uses edge-aligned masks and U-Net generator with color/texture constraints for seamless integration, achieving high attack success rates while maintaining visual concealment.
Details
Motivation: Current physical adversarial attacks on traffic signs lack stealth - they apply perturbations to central regions creating visually salient patterns easily detectable by humans, limiting real-world practicality. These attacks can cause misclassification leading to safety issues and can propagate through V2X networks causing cascading failures.Method: Uses instance segmentation to generate edge-aligned masks conforming to traffic sign shapes. Employs a U-Net generator to craft adversarial patches optimized through color/texture constraints and frequency domain analysis for seamless background integration. Focuses on peripheral areas where human attention is lower.
Result: Achieves over 90% attack success rate across various traffic sign classification models under limited query budgets. Shows strong cross-model transferability and maintains robust real-world performance stable under varying angles and distances.
Conclusion: TESP-Attack provides a highly effective stealth-aware adversarial attack method for traffic signs that balances attack effectiveness with visual concealment, addressing the key limitation of current physical attacks while maintaining practical applicability in real-world scenarios.
Abstract: Intelligent driving systems are vulnerable to physical adversarial attacks on traffic signs. These attacks can cause misclassification, leading to erroneous driving decisions that compromise road safety. Moreover, within V2X networks, such misinterpretations can propagate, inducing cascading failures that disrupt overall traffic flow and system stability. However, a key limitation of current physical attacks is their lack of stealth. Most methods apply perturbations to central regions of the sign, resulting in visually salient patterns that are easily detectable by human observers, thereby limiting their real-world practicality. This study proposes TESP-Attack, a novel stealth-aware adversarial patch method for traffic sign classification. Based on the observation that human visual attention primarily focuses on the central regions of traffic signs, we employ instance segmentation to generate edge-aligned masks that conform to the shape characteristics of the signs. A U-Net generator is utilized to craft adversarial patches, which are then optimized through color and texture constraints along with frequency domain analysis to achieve seamless integration with the background environment, resulting in highly effective visual concealment. The proposed method demonstrates outstanding attack success rates across traffic sign classification models with varied architectures, achieving over 90% under limited query budgets. It also exhibits strong cross-model transferability and maintains robust real-world performance that remains stable under varying angles and distances.
[242] EAG3R: Event-Augmented 3D Geometry Estimation for Dynamic and Extreme-Lighting Scenes
Xiaoshan Wu, Yifei Yu, Xiaoyang Lyu, Yihua Huang, Bo Wang, Baoheng Zhang, Zhongrui Wang, Xiaojuan Qi
Main category: cs.CV
TL;DR: EAG3R is a robust 3D geometry estimation framework that combines RGB images with asynchronous event streams to handle challenging dynamic low-light scenes without needing retraining on night-time data.
Details
Motivation: Existing RGB-only 3D geometry estimation methods struggle under real-world conditions with dynamic objects and extreme illumination due to limitations of conventional cameras. Event cameras offer complementary information but aren't effectively integrated with modern pointmap-based reconstruction approaches.Method: Built on MonST3R backbone, EAG3R introduces: 1) Retinex-inspired image enhancement module and lightweight event adapter with SNR-aware fusion that adaptively combines RGB and event features based on local reliability; 2) Novel event-based photometric consistency loss that reinforces spatiotemporal coherence during global optimization.
Result: EAG3R significantly outperforms state-of-the-art RGB-only baselines across monocular depth estimation, camera pose tracking, and dynamic reconstruction tasks in challenging dynamic low-light scenes.
Conclusion: Augmenting pointmap-based reconstruction with asynchronous event streams enables robust 3D geometry estimation in challenging real-world conditions, overcoming limitations of RGB-only approaches without requiring specialized retraining for night-time scenarios.
Abstract: Robust 3D geometry estimation from videos is critical for applications such as autonomous navigation, SLAM, and 3D scene reconstruction. Recent methods like DUSt3R demonstrate that regressing dense pointmaps from image pairs enables accurate and efficient pose-free reconstruction. However, existing RGB-only approaches struggle under real-world conditions involving dynamic objects and extreme illumination, due to the inherent limitations of conventional cameras. In this paper, we propose EAG3R, a novel geometry estimation framework that augments pointmap-based reconstruction with asynchronous event streams. Built upon the MonST3R backbone, EAG3R introduces two key innovations: (1) a retinex-inspired image enhancement module and a lightweight event adapter with SNR-aware fusion mechanism that adaptively combines RGB and event features based on local reliability; and (2) a novel event-based photometric consistency loss that reinforces spatiotemporal coherence during global optimization. Our method enables robust geometry estimation in challenging dynamic low-light scenes without requiring retraining on night-time data. Extensive experiments demonstrate that EAG3R significantly outperforms state-of-the-art RGB-only baselines across monocular depth estimation, camera pose tracking, and dynamic reconstruction tasks.
[243] DEJIMA: A Novel Large-scale Japanese Dataset for Image Captioning and Visual Question Answering
Toshiki Katsube, Taiga Fukuhara, Kenichiro Ando, Yusuke Mukuta, Kohei Uehara, Tatsuya Harada
Main category: cs.CV
TL;DR: Created DEJIMA, the largest Japanese V&L dataset (3.88M pairs) using scalable pipeline with web collection, filtering, object detection, and LLM refinement, achieving superior Japaneseness and cultural relevance.
Details
Motivation: Addresses scarcity of high-quality, large-scale Japanese Vision-and-Language resources, as existing datasets are limited in size and often rely on translation or manual annotation which lack authentic Japanese cultural and linguistic representation.Method: Developed scalable pipeline combining: 1) large-scale web collection, 2) rigorous filtering/deduplication, 3) object-detection-driven evidence extraction, 4) LLM-based refinement under grounding constraints. Built two resources: DEJIMA-Cap (image-caption) and DEJIMA-VQA (VQA).
Result: Created 3.88M image-text pairs, far exceeding existing Japanese V&L datasets. Human evaluation shows substantially higher Japaneseness and linguistic naturalness than translation/manual datasets, with factual correctness comparable to human-annotated corpora. Models trained on DEJIMA show consistent improvements across Japanese multimodal benchmarks.
Conclusion: Culturally grounded, large-scale resources are crucial for enhancing Japanese V&L model performance. DEJIMA dataset and pipeline are publicly released with commercial-friendly licensing to advance research and industrial applications in Japanese multimodal AI.
Abstract: This work addresses the scarcity of high-quality, large-scale resources for Japanese Vision-and-Language (V&L) modeling. We present a scalable and reproducible pipeline that integrates large-scale web collection with rigorous filtering/deduplication, object-detection-driven evidence extraction, and Large Language Model (LLM)-based refinement under grounding constraints. Using this pipeline, we build two resources: an image-caption dataset (DEJIMA-Cap) and a VQA dataset (DEJIMA-VQA), each containing 3.88M image-text pairs, far exceeding the size of existing Japanese V&L datasets. Human evaluations demonstrate that DEJIMA achieves substantially higher Japaneseness and linguistic naturalness than datasets constructed via translation or manual annotation, while maintaining factual correctness at a level comparable to human-annotated corpora. Quantitative analyses of image feature distributions further confirm that DEJIMA broadly covers diverse visual domains characteristic of Japan, complementing its linguistic and cultural representativeness. Models trained on DEJIMA exhibit consistent improvements across multiple Japanese multimodal benchmarks, confirming that culturally grounded, large-scale resources play a key role in enhancing model performance. All data sources and modules in our pipeline are licensed for commercial use, and we publicly release the resulting dataset and metadata to encourage further research and industrial applications in Japanese V&L modeling.
[244] PolarGS: Polarimetric Cues for Ambiguity-Free Gaussian Splatting with Accurate Geometry Recovery
Bo Guo, Sijia Wen, Yifan Zhao, Jia Li, Zhiming Zheng
Main category: cs.CV
TL;DR: PolarGS enhances 3D Gaussian Splatting by using polarization cues to handle photometrically ambiguous regions like reflective and textureless surfaces, improving geometric accuracy.
Details
Motivation: Existing surface reconstruction methods for 3DGS degrade in photometrically ambiguous regions (reflective/textureless surfaces) where unreliable photometric cues disrupt geometry estimation. Polarization provides complementary optical information that can resolve these ambiguities.Method: Two complementary modules: 1) Polarization-guided photometric correction - identifies reflective regions via Degree of Linear Polarization (DoLP) and refines reflective Gaussians with Color Refinement Maps; 2) Polarization-enhanced Gaussian densification - integrates Angle and Degree of Linear Polarization (A/DoLP) into PatchMatch-based depth completion for textureless areas, enabling back-projection and fusion of new Gaussians.
Result: PolarGS achieves superior geometric accuracy compared to state-of-the-art methods and is framework-agnostic.
Conclusion: Leveraging polarization as an optical prior effectively resolves photometric ambiguities in 3DGS, enabling more complete and accurate surface reconstruction, particularly for challenging reflective and textureless surfaces.
Abstract: Recent advances in surface reconstruction for 3D Gaussian Splatting (3DGS) have enabled remarkable geometric accuracy. However, their performance degrades in photometrically ambiguous regions such as reflective and textureless surfaces, where unreliable cues disrupt photometric consistency and hinder accurate geometry estimation. Reflected light is often partially polarized in a manner that reveals surface orientation, making polarization an optic complement to photometric cues in resolving such ambiguities. Therefore, we propose PolarGS, an optics-aware extension of RGB-based 3DGS that leverages polarization as an optical prior to resolve photometric ambiguities and enhance reconstruction accuracy. Specifically, we introduce two complementary modules: a polarization-guided photometric correction strategy, which ensures photometric consistency by identifying reflective regions via the Degree of Linear Polarization (DoLP) and refining reflective Gaussians with Color Refinement Maps; and a polarization-enhanced Gaussian densification mechanism for textureless area geometry recovery, which integrates both Angle and Degree of Linear Polarization (A/DoLP) into a PatchMatch-based depth completion process. This enables the back-projection and fusion of new Gaussians, leading to more complete reconstruction. PolarGS is framework-agnostic and achieves superior geometric accuracy compared to state-of-the-art methods.
[245] CircleFlow: Flow-Guided Camera Blur Estimation using a Circle Grid Target
Jiajian He, Enjie Hu, Shiqi Chen, Tianchen Qiu, Huajun Feng, Zhihai Xu, Yueting Chen
Main category: cs.CV
TL;DR: CircleFlow is a high-fidelity PSF estimation framework that uses flow-guided edge localization and implicit neural representations to accurately estimate camera blur from circle grid targets.
Details
Motivation: Accurate PSF estimation is crucial for optical characterization and computational vision, but remains challenging due to the inherent ambiguity and ill-posed nature of intensity-based deconvolution.Method: Uses structured capture with circle grid targets to encode anisotropic PSFs, leverages binary luminance prior, employs flow-guided subpixel alignment for sharp image reconstruction, models PSF as energy-constrained implicit neural representation, and jointly optimizes both components in a demosaicing-aware differentiable framework.
Result: Extensive experiments on simulated and real-world data demonstrate state-of-the-art accuracy and reliability, validating effectiveness for practical PSF calibration.
Conclusion: CircleFlow provides a robust framework for high-fidelity PSF estimation through accurate edge localization and physically consistent modeling, enabling practical camera calibration applications.
Abstract: The point spread function (PSF) serves as a fundamental descriptor linking the real-world scene to the captured signal, manifesting as camera blur. Accurate PSF estimation is crucial for both optical characterization and computational vision, yet remains challenging due to the inherent ambiguity and the ill-posed nature of intensity-based deconvolution. We introduce CircleFlow, a high-fidelity PSF estimation framework that employs flow-guided edge localization for precise blur characterization. CircleFlow begins with a structured capture that encodes locally anisotropic and spatially varying PSFs by imaging a circle grid target, while leveraging the target’s binary luminance prior to decouple image and kernel estimation. The latent sharp image is then reconstructed through subpixel alignment of an initialized binary structure guided by optical flow, whereas the PSF is modeled as an energy-constrained implicit neural representation. Both components are jointly optimized within a demosaicing-aware differentiable framework, ensuring physically consistent and robust PSF estimation enabled by accurate edge localization. Extensive experiments on simulated and real-world data demonstrate that CircleFlow achieves state-of-the-art accuracy and reliability, validating its effectiveness for practical PSF calibration.
[246] Thinking with Drafts: Speculative Temporal Reasoning for Efficient Long Video Understanding
Pengfei Hu, Meng Cao, Yingyao Wang, Yi Wang, Jiahua Dong, Jun Song, Yu Cheng, Bo Zheng, Xiaodan Liang
Main category: cs.CV
TL;DR: SpecTemp is a reinforcement learning framework that uses dual MLLMs to accelerate long video understanding by having a lightweight draft model propose salient frames and a powerful target model verify and refine them, mimicking human brain collaboration.
Details
Motivation: Current thinking-with-frames approaches for video MLLMs suffer from efficiency bottlenecks due to growing redundant multi-modal context, despite advancing temporal reasoning capabilities for long video understanding.Method: A reinforcement learning-based speculative temporal reasoning framework with dual-model design: lightweight draft MLLM rapidly explores/proposes salient frames from dense temporal regions, while powerful target MLLM focuses on temporal reasoning and verifies proposals, iteratively refining attention until convergence.
Result: SpecTemp maintains competitive accuracy across multiple video understanding benchmarks while significantly accelerating inference compared to existing thinking-with-frames methods.
Conclusion: The dual-model collaborative approach effectively balances efficiency and accuracy for long video understanding, mirroring human brain pathways and overcoming the redundancy bottleneck of current methods.
Abstract: Long video understanding is essential for human-like intelligence, enabling coherent perception and reasoning over extended temporal contexts. While the emerging thinking-with-frames paradigm, which alternates between global temporal reasoning and local frame examination, has advanced the reasoning capabilities of video multi-modal large language models (MLLMs), it suffers from a significant efficiency bottleneck due to the progressively growing and redundant multi-modal context. To address this, we propose SpecTemp, a reinforcement learning-based Speculative Temporal reasoning framework that decouples temporal perception from reasoning via a cooperative dual-model design. In SpecTemp, a lightweight draft MLLM rapidly explores and proposes salient frames from densely sampled temporal regions, while a powerful target MLLM focuses on temporal reasoning and verifies the draft’s proposals, iteratively refining its attention until convergence. This design mirrors the collaborative pathways of the human brain, balancing efficiency with accuracy. To support training, we construct the SpecTemp-80K dataset, featuring synchronized dual-level annotations for coarse evidence spans and fine-grained frame-level evidence. Experiments across multiple video understanding benchmarks demonstrate that SpecTemp not only maintains competitive accuracy but also significantly accelerates inference compared with existing thinking-with-frames methods.
[247] StreamGaze: Gaze-Guided Temporal Reasoning and Proactive Understanding in Streaming Videos
Daeun Lee, Subhojyoti Mukherjee, Branislav Kveton, Ryan A. Rossi, Viet Dac Lai, Seunghyun Yoon, Trung Bui, Franck Dernoncourt, Mohit Bansal
Main category: cs.CV
TL;DR: StreamGaze is the first benchmark for evaluating MLLMs’ ability to use human gaze signals for temporal and proactive reasoning in streaming videos, revealing significant performance gaps between current models and humans.
Details
Motivation: Existing streaming benchmarks don't measure whether MLLMs can interpret or leverage human gaze signals in streaming settings, which is crucial for realistic applications like AR glasses that need to anticipate user intentions.Method: Developed a gaze-video QA generation pipeline that aligns egocentric videos with raw gaze trajectories via fixation extraction, region-specific visual prompting, and scanpath construction to produce spatio-temporally grounded QA pairs reflecting human perceptual dynamics.
Result: Substantial performance gaps between state-of-the-art MLLMs and human performance across all StreamGaze tasks, revealing fundamental limitations in gaze-based temporal reasoning, intention modeling, and proactive prediction.
Conclusion: Current MLLMs struggle with gaze-guided streaming video understanding, and the benchmark provides insights into failure modes and capabilities needed for future models, with all data and code being publicly released for continued research.
Abstract: Streaming video understanding requires models not only to process temporally incoming frames, but also to anticipate user intention for realistic applications like AR glasses. While prior streaming benchmarks evaluate temporal reasoning, none measure whether MLLMs can interpret or leverage human gaze signals within a streaming setting. To fill this gap, we introduce StreamGaze, the first benchmark designed to evaluate how effectively MLLMs use gaze for temporal and proactive reasoning in streaming videos. StreamGaze introduces gaze-guided past, present, and proactive tasks that comprehensively evaluate streaming video understanding. These tasks assess whether models can use real-time gaze to follow shifting attention and infer user intentions from only past and currently observed frames. To build StreamGaze, we develop a gaze-video QA generation pipeline that aligns egocentric videos with raw gaze trajectories via fixation extraction, region-specific visual prompting, and scanpath construction. This pipeline produces spatio-temporally grounded QA pairs that closely reflect human perceptual dynamics. Across all StreamGaze tasks, we observe substantial performance gaps between state-of-the-art MLLMs and human performance, revealing fundamental limitations in gaze-based temporal reasoning, intention modeling, and proactive prediction. We further provide detailed analyses of gaze-prompting strategies, reasoning behaviors, and task-specific failure modes, offering deeper insight into why current MLLMs struggle and what capabilities future models must develop. All data and code will be publicly released to support continued research in gaze-guided streaming video understanding.
[248] IRPO: Boosting Image Restoration via Post-training GRPO
Haoxuan Xu. Yi Liu, Boyuan Jiang, Jinlong Peng, Donghao Luo, Xiaobin Hu, Shuicheng Yan, Haoang Li
Main category: cs.CV
TL;DR: IRPO is a novel GRPO-based post-training paradigm for low-level vision tasks that addresses limitations of existing image restoration methods through optimized data formulation and a multi-component reward system balancing objective accuracy and human perception.
Details
Motivation: Existing image restoration methods rely on pixel-level hard-fitting to ground-truth images, which leads to over-smoothing and poor generalization. While post-training paradigms have succeeded in high-level generation tasks, their potential for low-level vision remains unexplored.Method: IRPO introduces two key innovations: 1) A data formulation principle that selects underperforming samples from pre-training for optimal performance and efficiency, and 2) A three-component reward system including General Reward for structural fidelity, Expert Reward using Qwen-VL for perceptual alignment, and Restoration Reward for task-specific low-level quality.
Result: IRPO achieves state-of-the-art results on six in-domain and five out-of-domain low-level benchmarks, surpassing the AdaIR baseline by 0.83 dB on in-domain tasks and 3.43 dB on OOD settings across diverse degradation types.
Conclusion: IRPO successfully demonstrates the potential of post-training paradigms for low-level vision tasks, providing a systematic approach that balances objective accuracy with human perceptual preference while achieving superior performance and generalization.
Abstract: Recent advances in post-training paradigms have achieved remarkable success in high-level generation tasks, yet their potential for low-level vision remains rarely explored. Existing image restoration (IR) methods rely on pixel-level hard-fitting to ground-truth images, struggling with over-smoothing and poor generalization. To address these limitations, we propose IRPO, a low-level GRPO-based post-training paradigm that systematically explores both data formulation and reward modeling. We first explore a data formulation principle for low-level post-training paradigm, in which selecting underperforming samples from the pre-training stage yields optimal performance and improved efficiency. Furthermore, we model a reward-level criteria system that balances objective accuracy and human perceptual preference through three complementary components: a General Reward for structural fidelity, an Expert Reward leveraging Qwen-VL for perceptual alignment, and a Restoration Reward for task-specific low-level quality. Comprehensive experiments on six in-domain and five out-of-domain (OOD) low-level benchmarks demonstrate that IRPO achieves state-of-the-art results across diverse degradation types, surpassing the AdaIR baseline by 0.83 dB on in-domain tasks and 3.43 dB on OOD settings. Our code can be shown in https://github.com/HaoxuanXU1024/IRPO.
[249] PanFlow: Decoupled Motion Control for Panoramic Video Generation
Cheng Zhang, Hanwen Liang, Donny Y. Chen, Qianyi Wu, Konstantinos N. Plataniotis, Camilo Cruz Gambardella, Jianfei Cai
Main category: cs.CV
TL;DR: PanFlow is a novel panoramic video generation method that enables explicit motion control for large and complex motions by decoupling camera rotation from optical flow and using spherical noise warping for loop consistency.
Details
Motivation: Existing panoramic video generation methods lack explicit motion control and struggle with large/complex motions, limiting their practical applications in virtual reality and immersive media.Method: PanFlow exploits the spherical nature of panoramas to decouple camera rotation from input optical flow, introduces spherical noise warping for boundary consistency, and uses a curated large-scale motion-rich dataset with pose and flow annotations.
Result: PanFlow significantly outperforms prior methods in motion fidelity, visual quality, and temporal coherence, and demonstrates effectiveness in motion transfer and video editing applications.
Conclusion: PanFlow provides a robust solution for controllable panoramic video generation with large motions, advancing the field for VR and immersive media applications.
Abstract: Panoramic video generation has attracted growing attention due to its applications in virtual reality and immersive media. However, existing methods lack explicit motion control and struggle to generate scenes with large and complex motions. We propose PanFlow, a novel approach that exploits the spherical nature of panoramas to decouple the highly dynamic camera rotation from the input optical flow condition, enabling more precise control over large and dynamic motions. We further introduce a spherical noise warping strategy to promote loop consistency in motion across panorama boundaries. To support effective training, we curate a large-scale, motion-rich panoramic video dataset with frame-level pose and flow annotations. We also showcase the effectiveness of our method in various applications, including motion transfer and video editing. Extensive experiments demonstrate that PanFlow significantly outperforms prior methods in motion fidelity, visual quality, and temporal coherence. Our code, dataset, and models are available at https://github.com/chengzhag/PanFlow.
[250] AFRAgent : An Adaptive Feature Renormalization Based High Resolution Aware GUI agent
Neeraj Anand, Rishabh Jain, Sohan Patnaik, Balaji Krishnamurthy, Mausoom Sarkar
Main category: cs.CV
TL;DR: AFRAgent is a compact instruct-BLIP-based multimodal model for mobile UI automation that achieves state-of-the-art performance while being 4x smaller than competitors through adaptive feature renormalization.
Details
Motivation: Mobile UI automation has growing industrial demand, and while VLMs enable autonomous task execution, current models struggle with accurate widget identification and action determination due to limited spatial information in vision features. Additionally, top-performing models are large, requiring extensive training and causing inference delays.Method: AFRAgent uses an instruct-BLIP-based multimodal architecture with adaptive feature renormalization - a token-level affine transformation technique that enriches low-resolution image embeddings and fuses high-resolution details to enhance image embeddings in the LLM pipeline.
Result: AFRAgent achieves superior performance on Meta-GUI and AITW benchmarks for smartphone automation while being less than one-fourth the size of its nearest competitor, establishing a new state-of-the-art baseline.
Conclusion: AFRAgent demonstrates that efficient, compact models can achieve state-of-the-art performance in GUI automation through innovative feature enhancement techniques, addressing both accuracy and efficiency challenges in mobile UI automation.
Abstract: There is a growing demand for mobile user interface (UI) automation, driven by its broad applications across industries. With the advent of visual language models (VLMs), GUI automation has progressed from generating text-based instructions for humans to autonomously executing tasks, thus optimizing automation workflows. Recent approaches leverage VLMs for this problem due to their ability to 1) process on-screen content directly, 2) remain independent of device-specific APIs by utilizing human actions (e.g., clicks, typing), and 3) apply real-world contextual knowledge for task understanding. However, these models often have trouble accurately identifying widgets and determining actions due to limited spatial information in vision encoder features. Additionally, top-performing models are often large, requiring extensive training and resulting in inference delays. In this work, we introduce AFRAgent, an instruct-BLIP-based multimodal architecture that achieves superior performance in GUI automation while being less than one-fourth the size of its nearest competitor. To enhance image embeddings in the large language model (LLM) pipeline, we propose an adaptive feature renormalization-based (a token-level affine transformation) technique that effectively enriches low-resolution image embeddings and fuses high-resolution details. We evaluate AFRAgent on Meta-GUI and AITW benchmarks, establishing a new state-of-the-art baseline for smartphone automation.
[251] Smol-GS: Compact Representations for Abstract 3D Gaussian Splatting
Haishan Wang, Mohammad Hassan Vali, Arno Solin
Main category: cs.CV
TL;DR: Smol-GS is a novel compression method for 3D Gaussian Splatting that learns compact 3D representations integrating spatial and semantic information, achieving state-of-the-art compression while maintaining rendering quality.
Details
Motivation: To create highly efficient encodings for 3D Gaussian Splatting that can compress 3D scenes by orders of magnitude without sacrificing flexibility, enabling both high-quality rendering and potential use in downstream 3D understanding tasks.Method: Uses a recursive voxel hierarchy to capture splat coordinates, while splat-wise features store abstracted cues including color, opacity, transformation, and material properties, integrating both spatial and semantic information in 3D space.
Result: Achieves state-of-the-art compression on standard benchmarks while maintaining high rendering quality, compressing 3D scenes by orders of magnitude without loss of flexibility.
Conclusion: Smol-GS provides an effective approach for learning compact 3D representations that could serve as a foundation for downstream tasks like navigation, planning, and broader 3D scene understanding beyond just visual fidelity.
Abstract: We present Smol-GS, a novel method for learning compact representations for 3D Gaussian Splatting (3DGS). Our approach learns highly efficient encodings in 3D space that integrate both spatial and semantic information. The model captures the coordinates of the splats through a recursive voxel hierarchy, while splat-wise features store abstracted cues, including color, opacity, transformation, and material properties. This design allows the model to compress 3D scenes by orders of magnitude without loss of flexibility. Smol-GS achieves state-of-the-art compression on standard benchmarks while maintaining high rendering quality. Beyond visual fidelity, the discrete representations could potentially serve as a foundation for downstream tasks such as navigation, planning, and broader 3D scene understanding.
[252] TAP-CT: 3D Task-Agnostic Pretraining of Computed Tomography Foundation Models
Tim Veenboer, George Yiasemis, Eric Marcus, Vivien Van Veldhuizen, Cees G. M. Snoek, Jonas Teuwen, Kevin B. W. Groot Lipman
Main category: cs.CV
TL;DR: TAP-CT introduces a task-agnostic foundation model for CT imaging that requires minimal fine-tuning, using Vision Transformers adapted for 3D volumetric data with targeted architectural modifications.
Details
Motivation: Existing medical foundation models require extensive fine-tuning or resource-intensive decoders, and many encoders have task-biased pretraining objectives, creating a need for strong, task-agnostic models that need minimal fine-tuning beyond feature extraction.Method: Simple adaptation of Vision Transformers and DINOv2 for volumetric data with targeted modifications to patch embeddings, positional encodings, and volumetric augmentations, enabling scalable self-supervised pretraining directly on 3D CT volumes.
Result: Large-scale 3D pretraining on 105K CT volumes yields stable, robust frozen representations that generalize strongly across downstream tasks.
Conclusion: TAP-CT provides a powerful, low-resource baseline for medical imaging research, with all models, configurations, and benchmark code released to promote transparency and reproducibility.
Abstract: Existing foundation models (FMs) in the medical domain often require extensive fine-tuning or rely on training resource-intensive decoders, while many existing encoders are pretrained with objectives biased toward specific tasks. This illustrates a need for a strong, task-agnostic foundation model that requires minimal fine-tuning beyond feature extraction. In this work, we introduce a suite of task-agnostic pretraining of CT foundation models (TAP-CT): a simple yet effective adaptation of Vision Transformers (ViTs) and DINOv2 for volumetric data, enabling scalable self-supervised pretraining directly on 3D CT volumes. Our approach incorporates targeted modifications to patch embeddings, positional encodings, and volumetric augmentations, making the architecture depth-aware while preserving the simplicity of the underlying architectures. We show that large-scale 3D pretraining on an extensive in-house CT dataset (105K volumes) yields stable, robust frozen representations that generalize strongly across downstream tasks. To promote transparency and reproducibility, and to establish a powerful, low-resource baseline for future research in medical imaging, we will release all pretrained models, experimental configurations, and downstream benchmark code at https://huggingface.co/fomofo/tap-ct-b-3d.
[253] Neural Discrete Representation Learning for Sparse-View CBCT Reconstruction: From Algorithm Design to Prospective Multicenter Clinical Evaluation
Haoshen Wang, Lei Chen, Wei-Hua Zhang, Linxia Wu, Yong Luo, Zengmao Wang, Yuan Xiong, Chengcheng Zhu, Wenjuan Tang, Xueyi Zhang, Wei Zhou, Xuhua Duan, Lefei Zhang, Gao-Jun Teng, Bo Du, Huangxuan Zhao
Main category: cs.CV
TL;DR: DeepPriorCBCT is a three-stage deep learning framework that achieves diagnostic-grade CBCT reconstruction using only 1/6 of conventional radiation dose, validated through large multicenter retrospective data and a prospective clinical trial.
Details
Motivation: CBCT-guided puncture for thoracic tumors exposes patients to significant radiation, increasing secondary malignancy risk. Existing low-dose CBCT strategies lack validation with large-scale multicenter data and prospective clinical evaluation.Method: Three-stage deep learning framework for sparse-sampling CBCT reconstruction. Developed and validated using 4102 patients with 8675 CBCT scans from 12 centers. Prospective cross-over trial (NCT07035977) with 138 patients scheduled for percutaneous thoracic puncture.
Result: Reconstructed images indistinguishable from original scans per 11 physicians. Diagnostic performance and image quality comparable to standard reconstruction. Prospective trial: no significant differences in image quality or lesion assessment (all P>0.05). 25 interventionalists expressed no preference between model-based and full-sampling images (Kappa<0.2). Radiation reduced to ~1/6 of conventional approach.
Conclusion: DeepPriorCBCT enables high-quality CBCT reconstruction under sparse sampling while markedly decreasing intraoperative radiation risk, validated through both retrospective and prospective clinical studies.
Abstract: Cone beam computed tomography (CBCT)-guided puncture has become an established approach for diagnosing and treating early- to mid-stage thoracic tumours, yet the associated radiation exposure substantially elevates the risk of secondary malignancies. Although multiple low-dose CBCT strategies have been introduced, none have undergone validation using large-scale multicenter retrospective datasets, and prospective clinical evaluation remains lacking. Here, we propose DeepPriorCBCT - a three-stage deep learning framework that achieves diagnostic-grade reconstruction using only one-sixth of the conventional radiation dose. 4102 patients with 8675 CBCT scans from 12 centers were included to develop and validate DeepPriorCBCT. Additionally, a prospective cross-over trial (Registry number: NCT07035977) which recruited 138 patients scheduled for percutaneous thoracic puncture was conducted to assess the model’s clinical applicability. Assessment by 11 physicians confirmed that reconstructed images were indistinguishable from original scans. Moreover, diagnostic performance and overall image quality were comparable to those generated by standard reconstruction algorithms. In the prospective trial, five radiologists reported no significant differences in image quality or lesion assessment between DeepPriorCBCT and the clinical standard (all P>0.05). Likewise, 25 interventionalists expressed no preference between model-based and full-sampling images for surgical guidance (Kappa<0.2). Radiation exposure with DeepPriorCBCT was reduced to approximately one-sixth of that with the conventional approach, and collectively, the findings confirm that it enables high-quality CBCT reconstruction under sparse sampling conditions while markedly decreasing intraoperative radiation risk.
[254] Feed-Forward 3D Gaussian Splatting Compression with Long-Context Modeling
Zhening Liu, Rui Song, Yushi Huang, Yingdong Hu, Xinjie Zhang, Jiawei Shao, Zehong Lin, Jun Zhang
Main category: cs.CV
TL;DR: A novel feed-forward compression framework for 3D Gaussian Splatting that achieves 20× compression ratio by modeling long-range spatial dependencies through large-scale context structures and attention-based transform coding.
Details
Motivation: 3D Gaussian Splatting (3DGS) has emerged as a revolutionary 3D representation but suffers from substantial data size that limits widespread adoption. Existing feed-forward compression methods struggle with modeling long-range spatial dependencies due to limited receptive fields and inadequate context capacity in entropy models.Method: Proposes a feed-forward 3DGS compression framework with: 1) Large-scale context structure comprising thousands of Gaussians based on Morton serialization, 2) Fine-grained space-channel auto-regressive entropy model to leverage expansive context, and 3) Attention-based transform coding model to extract informative latent priors by aggregating features from wide-range neighboring Gaussians.
Result: Achieves 20× compression ratio for 3DGS in feed-forward inference and establishes state-of-the-art performance among generalizable codecs.
Conclusion: The proposed framework effectively models long-range correlations to enable highly compact and generalizable 3D representations, overcoming limitations of existing methods and making 3DGS more practical for widespread adoption.
Abstract: 3D Gaussian Splatting (3DGS) has emerged as a revolutionary 3D representation. However, its substantial data size poses a major barrier to widespread adoption. While feed-forward 3DGS compression offers a practical alternative to costly per-scene per-train compressors, existing methods struggle to model long-range spatial dependencies, due to the limited receptive field of transform coding networks and the inadequate context capacity in entropy models. In this work, we propose a novel feed-forward 3DGS compression framework that effectively models long-range correlations to enable highly compact and generalizable 3D representations. Central to our approach is a large-scale context structure that comprises thousands of Gaussians based on Morton serialization. We then design a fine-grained space-channel auto-regressive entropy model to fully leverage this expansive context. Furthermore, we develop an attention-based transform coding model to extract informative latent priors by aggregating features from a wide range of neighboring Gaussians. Our method yields a $20\times$ compression ratio for 3DGS in a feed-forward inference and achieves state-of-the-art performance among generalizable codecs.
[255] Quantum-Inspired Spectral Geometry for Neural Operator Equivalence and Structured Pruning
Haijian Shao, Wei Liu, Xing Deng
Main category: cs.CV
TL;DR: A quantum-inspired geometric framework for neural operators using normalized singular value spectra on Bloch hypersphere enables cross-modal/cross-architecture operator substitutability with provable functional closeness.
Details
Motivation: Address bottlenecks in multimodal intelligence on resource-constrained domestic hardware: multimodal feature heterogeneity, real-time requirements in dynamic scenarios, and hardware-specific operator redundancy.Method: Quantum-inspired geometric framework representing operators by normalized singular value spectrum on Bloch hypersphere; spectral-to-functional equivalence theorem; Quantum Metric-Driven Functional Redundancy Graphs (QM-FRG) and one-shot structured pruning.
Result: Proposed metric shows superiority over magnitude and random baselines in controlled simulation; extensive experimental validation on large-scale multimodal transformers and domestic heterogeneous hardware (Huawei Ascend, Cambricon MLU, Kunlunxin) is deferred to journal version.
Conclusion: Establishes first rigorous foundation for cross-modal and cross-architecture operator substitutability through quantum-inspired geometric representation and spectral-to-functional equivalence.
Abstract: The rapid growth of multimodal intelligence on resource-constrained and heterogeneous domestic hardware exposes critical bottlenecks: multimodal feature heterogeneity, real-time requirements in dynamic scenarios, and hardware-specific operator redundancy. This work introduces a quantum-inspired geometric framework for neural operators that represents each operator by its normalized singular value spectrum on the Bloch hypersphere. We prove a tight spectral-to-functional equivalence theorem showing that vanishing Fubini–Study/Wasserstein-2 distance implies provable functional closeness, establishing the first rigorous foundation for cross-modal and cross-architecture operator substitutability. Based on this metric, we propose Quantum Metric-Driven Functional Redundancy Graphs (QM-FRG) and one-shot structured pruning. Controlled simulation validates the superiority of the proposed metric over magnitude and random baselines. An extensive experimental validation on large-scale multimodal transformers and domestic heterogeneous hardware (Huawei Ascend, Cambricon MLU, Kunlunxin) hardware is deferred to an extended journal version currently in preparation.
[256] Look, Recite, Then Answer: Enhancing VLM Performance via Self-Generated Knowledge Hints
Xisheng Feng
Main category: cs.CV
TL;DR: A framework called “Look, Recite, Then Answer” improves Vision-Language Models for specialized domains by reducing hallucinations through self-generated knowledge hints and evidence alignment.
Details
Motivation: VLMs struggle in specialized domains like precision agriculture due to "Reasoning-Driven Hallucination" where linguistic priors override visual perception, and the "Modality Gap" where visual embeddings fail to activate fine-grained expert knowledge already in model parameters.Method: A parameter-efficient framework with three decoupled stages: (1) Look generates objective visual descriptions and candidate sets; (2) Recite uses a lightweight 1.7B router to transform visual cues into targeted queries that trigger candidate-specific parametric knowledge; (3) Answer performs parallel evidence alignment between descriptions and recited knowledge to select the most consistent label.
Result: Achieves state-of-the-art results on AgroBench, improving Weed Identification accuracy by 23.6% over Qwen-VL and surpassing GPT-4o without external search overhead.
Conclusion: The modular design mitigates hallucinations by transforming passive perception into active, controllable knowledge retrieval, enhancing VLM performance in specialized domains while keeping backbone models frozen.
Abstract: Vision-Language Models (VLMs) exhibit significant performance plateaus in specialized domains like precision agriculture, primarily due to “Reasoning-Driven Hallucination” where linguistic priors override visual perception. A key bottleneck is the “Modality Gap”: visual embeddings fail to reliably activate the fine-grained expert knowledge already encoded in model parameters. We propose “Look, Recite, Then Answer,” a parameter-efficient framework that enhances VLMs via self-generated knowledge hints while keeping backbone models frozen. The framework decouples inference into three stages: (1) Look generates objective visual descriptions and candidate sets; (2) Recite employs a lightweight 1.7B router to transform visual cues into targeted queries that trigger candidate-specific parametric knowledge; (3) Answer performs parallel evidence alignment between descriptions and recited knowledge to select the most consistent label. On AgroBench, our method achieves state-of-the-art results, improving Weed Identification accuracy by 23.6% over Qwen-VL and surpassing GPT-4o without external search overhead. This modular design mitigates hallucinations by transforming passive perception into active, controllable knowledge retrieval
[257] HanDyVQA: A Video QA Benchmark for Fine-Grained Hand-Object Interaction Dynamics
Masatoshi Tateno, Gido Kato, Hirokatsu Kataoka, Yoichi Sato, Takuma Yagi
Main category: cs.CV
TL;DR: HanDyVQA is a fine-grained video QA benchmark for hand-object interaction dynamics, covering manipulation and effects with 6 question types and 11.1K QA pairs, revealing significant gaps between current models (73% accuracy) and human performance (97%).
Details
Motivation: Existing HOI benchmarks focus either on manipulation or effects at coarse levels, lacking fine-grained spatio-temporal reasoning to capture the underlying dynamics in hand-object interactions.Method: Created HanDyVQA benchmark with 6 complementary question types (Action, Process, Objects, Location, State Change, Object Parts), totaling 11.1K multiple-choice QA pairs and 10.3K segmentation masks for object/part-level reasoning.
Result: Best-performing model (Gemini-2.5-Pro) achieved only 73% average accuracy, far from human performance (97%). Analysis revealed challenges in spatial relationships, motion, and part-level geometric understanding.
Conclusion: HanDyVQA highlights significant gaps in current models’ understanding of HOI dynamics. Integrating explicit HOI-related cues improves performance, offering insights for developing future models with deeper understanding of hand-object interaction dynamics.
Abstract: Hand-object interaction (HOI) inherently involves dynamics where human manipulations produce distinct spatio-temporal effects on objects. However, existing semantic HOI benchmarks focused either on manipulation or on the resulting effects at a coarse level, lacking fine-grained spatio-temporal reasoning to capture the underlying dynamics in HOI. We introduce HanDyVQA, a fine-grained video question-answering benchmark that comprehensively covers both the manipulation and effect aspects of HOI. HanDyVQA comprises six complementary question types (Action, Process, Objects, Location, State Change, and Object Parts), totalling 11.1K multiple-choice QA pairs. Collected QA pairs recognizing manipulation styles, hand/object motions, and part-level state changes. HanDyVQA also includes 10.3K segmentation masks for Objects and Object Parts questions, enabling the evaluation of object/part-level reasoning in video object segmentation. We evaluated recent video foundation models on our benchmark and found that even the best-performing model, Gemini-2.5-Pro, reached only 73% average accuracy, which is far from human performance (97%). Further analysis shows the remaining challenges in spatial relationship, motion, and part-level geometric understanding. We also found that integrating explicit HOI-related cues into visual features improves performance, offering insights for developing future models with a deeper understanding of HOI dynamics.
[258] Multilingual Training-Free Remote Sensing Image Captioning
Carlos Rebelo, Gil Rocha, João Daniel Silva, Bruno Martins
Main category: cs.CV
TL;DR: First training-free multilingual remote sensing image captioning using retrieval-augmented prompting with SigLIP2 encoder and language models, achieving competitive performance with supervised English-only systems across 10 languages.
Details
Motivation: Address limitations of current remote sensing captioning: reliance on large annotated datasets, English-only focus restricting global applicability, and need for more inclusive/scalable Earth observation systems.Method: Retrieval-augmented prompting approach: 1) Use domain-adapted SigLIP2 encoder to retrieve related captions/examples from datastore, 2) Two variants: image-blind (multilingual LLM with textual prompts) and image-aware (VLM processing prompt+image), 3) Graph-based re-ranking using PageRank on image-caption graph for coherence.
Result: Competitive with fully supervised English-only systems across 4 benchmark datasets and 10 languages. PageRank re-ranking yields up to 35% performance improvements. VLMs generate visually grounded but lexically diverse captions, while LLMs achieve stronger BLEU/CIDEr scores. Direct target language generation outperforms translation-based strategies.
Conclusion: First systematic evaluation of multilingual, training-free captioning for remote sensing imagery, advancing toward more inclusive and scalable multimodal Earth observation systems without requiring large annotated datasets.
Abstract: Remote sensing image captioning has advanced rapidly through encoder–decoder models, although the reliance on large annotated datasets and the focus on English restricts global applicability. To address these limitations, we propose the first training-free multilingual approach, based on retrieval-augmented prompting. For a given aerial image, we employ a domain-adapted SigLIP2 encoder to retrieve related captions and few-shot examples from a datastore, which are then provided to a language model. We explore two variants: an image-blind setup, where a multilingual Large Language Model (LLM) generates the caption from textual prompts alone, and an image-aware setup, where a Vision–Language Model (VLM) jointly processes the prompt and the input image. To improve the coherence of the retrieved content, we introduce a graph-based re-ranking strategy using PageRank on a graph of images and captions. Experiments on four benchmark datasets across ten languages demonstrate that our approach is competitive with fully supervised English-only systems and generalizes to other languages. Results also highlight the importance of re-ranking with PageRank, yielding up to 35% improvements in performance metrics. Additionally, it was observed that while VLMs tend to generate visually grounded but lexically diverse captions, LLMs can achieve stronger BLEU and CIDEr scores. Lastly, directly generating captions in the target language consistently outperforms other translation-based strategies. Overall, our work delivers one of the first systematic evaluations of multilingual, training-free captioning for remote sensing imagery, advancing toward more inclusive and scalable multimodal Earth observation systems.
[259] Accelerating Streaming Video Large Language Models via Hierarchical Token Compression
Yiyu Wang, Xuyang Liu, Xiyan Gui, Xinying Lin, Boxue Yang, Chenfei Liao, Tailai Chen, Linfeng Zhang
Main category: cs.CV
TL;DR: STC is a hierarchical token compression framework that accelerates streaming VideoLLMs by reducing redundant ViT encoding and compressing visual tokens before LLM processing, achieving significant latency reductions with minimal accuracy loss.
Details
Motivation: Streaming VideoLLMs face computational bottlenecks in real-time deployment due to redundant processing of temporally similar frames in ViT encoding and inflated token sequences during LLM pre-filling, leading to high latency and memory overhead.Method: STC introduces two token-level accelerators: STC-Cacher reduces ViT encoding overhead by caching and reusing features from similar frames, and STC-Pruner compresses visual token sequences before LLM processing by preserving only the most salient tokens based on spatial and temporal relevance.
Result: STC outperforms other compression methods across four baseline streaming VideoLLMs on five benchmarks, retaining up to 99% accuracy on ReKV while reducing ViT encoding latency by 24.5% and LLM pre-filling latency by 45.3%.
Conclusion: STC provides an effective plug-and-play framework for accelerating streaming VideoLLMs by addressing computational bottlenecks in both ViT encoding and LLM pre-filling stages, enabling more efficient real-time video understanding with minimal accuracy trade-offs.
Abstract: Streaming Video Large Language Models (VideoLLMs) have demonstrated impressive performance across various video understanding tasks, but they face significant challenges in real-time deployment due to the high computational cost of processing dense visual tokens from continuous video streams. In streaming video scenarios, the primary bottleneck lies in the Vision Transformer (ViT) encoding stage, where redundant processing of temporally similar frames leads to inefficiency. Additionally, inflated token sequences during LLM pre-filling further exacerbate latency and memory overhead. To address these challenges, we propose \textbf{S}treaming \textbf{T}oken \textbf{C}ompression (\textbf{STC}), a plug-and-play hierarchical framework that seamlessly integrates into existing streaming VideoLLMs, optimizing both ViT encoding and LLM pre-filling stages to accelerate processing. STC introduces two token-level accelerators: \textbf{STC-Cacher}, which reduces ViT encoding overhead by caching and reusing features from temporally similar frames, and \textbf{STC-Pruner}, which compresses the visual token sequence before it enters the LLM, preserving only the most salient tokens based on both spatial and temporal relevance. Extensive experiments on four baseline streaming VideoLLMs across five benchmarks demonstrate that STC outperforms other compression methods. Notably, STC retains up to \textbf{99%} of accuracy on the ReKV framework while reducing ViT encoding latency and LLM pre-filling latency by \textbf{24.5%} and \textbf{45.3%}.
[260] SwiftVLA: Unlocking Spatiotemporal Dynamics for Lightweight VLA Models at Minimal Overhead
Chaojun Ni, Cheng Chen, Xiaofeng Wang, Zheng Zhu, Wenzhao Zheng, Boyuan Wang, Tianrun Chen, Guosheng Zhao, Haoyun Li, Zhehao Dong, Qiang Zhang, Yun Ye, Yang Wang, Guan Huang, Wenjun Mei
Main category: cs.CV
TL;DR: SwiftVLA enhances compact Vision-Language-Action models with 4D understanding using a pretrained 4D visual geometry transformer, Fusion Tokens for unified representations, and a mask-and-reconstruct strategy, achieving performance comparable to much larger models while being 18× faster and 12× more memory-efficient.
Details
Motivation: Current Vision-Language-Action (VLA) models are too large for practical deployment, while lightweight alternatives compromise spatiotemporal reasoning. Existing methods that incorporate 3D inputs still rely on large VLMs and lack temporal understanding.Method: 1) Uses pretrained 4D visual geometry transformer with temporal cache to extract 4D features from 2D images; 2) Introduces Fusion Tokens (learnable tokens trained with future prediction) to unify 2D and 4D representations; 3) Employs mask-and-reconstruct strategy to train VLM on 4D representations, allowing 4D branch to be dropped at inference.
Result: SwiftVLA outperforms lightweight baselines and rivals VLAs up to 7× larger, achieving comparable performance on edge devices while being 18× faster and reducing memory footprint by 12×.
Conclusion: SwiftVLA demonstrates that compact VLA models can achieve strong 4D understanding through efficient architectural design, enabling practical deployment on edge devices without sacrificing spatiotemporal reasoning capabilities.
Abstract: Vision-Language-Action (VLA) models built on pretrained Vision-Language Models (VLMs) show strong potential but are limited in practicality due to their large parameter counts. To mitigate this issue, using a lightweight VLM has been explored, but it compromises spatiotemporal reasoning. Although some methods suggest that incorporating additional 3D inputs can help, they usually rely on large VLMs to fuse 3D and 2D inputs and still lack temporal understanding. Therefore, we propose SwiftVLA, an architecture that enhances a compact model with 4D understanding while preserving design efficiency. Specifically, our approach features a pretrained 4D visual geometry transformer with a temporal cache that extracts 4D features from 2D images. Then, to enhance the VLM’s ability to exploit both 2D images and 4D features, we introduce Fusion Tokens, a set of learnable tokens trained with a future prediction objective to generate unified representations for action generation. Finally, we introduce a mask-and-reconstruct strategy that masks 4D inputs to the VLM and trains the VLA to reconstruct them, enabling the VLM to learn effective 4D representations and allowing the 4D branch to be dropped at inference with minimal performance loss. Experiments in real and simulated environments show that SwiftVLA outperforms lightweight baselines and rivals VLAs up to 7 times larger, achieving comparable performance on edge devices while being 18 times faster and reducing memory footprint by 12 times.
[261] Hierarchical Semantic Alignment for Image Clustering
Xingyu Zhu, Beier Zhu, Yunfan Li, Junfeng Fang, Shuo Wang, Kesen Zhao, Hanwang Zhang
Main category: cs.CV
TL;DR: CAE is a training-free hierarchical semantic alignment method for image clustering that combines caption-level descriptions and noun-level concepts to address noun ambiguity and improve clustering performance.
Details
Motivation: Existing image clustering methods that use nouns as external semantic knowledge often overlook noun ambiguity, which can distort semantic representations and degrade clustering quality.Method: Proposes CAE with three steps: 1) Construct semantic space using relevant nouns from WordNet and descriptions from caption datasets, 2) Align image features with selected nouns and captions via optimal transport to get discriminative semantic space, 3) Combine enhanced semantic and image features for clustering.
Result: Extensive experiments on 8 datasets show effectiveness, notably surpassing state-of-the-art training-free approach with 4.2% accuracy improvement and 2.9% ARI improvement on ImageNet-1K.
Conclusion: CAE effectively addresses noun ambiguity in image clustering by combining hierarchical semantic alignment, achieving superior performance without requiring training.
Abstract: Image clustering is a classic problem in computer vision, which categorizes images into different groups. Recent studies utilize nouns as external semantic knowledge to improve clus- tering performance. However, these methods often overlook the inherent ambiguity of nouns, which can distort semantic representations and degrade clustering quality. To address this issue, we propose a hierarChical semAntic alignmEnt method for image clustering, dubbed CAE, which improves cluster- ing performance in a training-free manner. In our approach, we incorporate two complementary types of textual seman- tics: caption-level descriptions, which convey fine-grained attributes of image content, and noun-level concepts, which represent high-level object categories. We first select relevant nouns from WordNet and descriptions from caption datasets to construct a semantic space aligned with image features. Then, we align image features with selected nouns and captions via optimal transport to obtain a more discriminative semantic space. Finally, we combine the enhanced semantic and image features to perform clustering. Extensive experiments across 8 datasets demonstrate the effectiveness of our method, notably surpassing the state-of-the-art training-free approach with a 4.2% improvement in accuracy and a 2.9% improvement in adjusted rand index (ARI) on the ImageNet-1K dataset.
[262] TalkingPose: Efficient Face and Gesture Animation with Feedback-guided Diffusion Model
Alireza Javanmardi, Pragati Jaiswal, Tewodros Amberbir Habtegebrial, Christen Millerdurai, Shaoxiang Wang, Alain Pagani, Didier Stricker
Main category: cs.CV
TL;DR: TalkingPose is a diffusion-based framework for generating long-form, temporally coherent human upper-body animations from single images and driving poses, overcoming computational limitations of existing methods.
Details
Motivation: Existing diffusion models for character animation struggle with generating temporally coherent long-form content due to computational/memory constraints from training on short video segments, limiting their effectiveness for extended coherent generation.Method: TalkingPose uses a diffusion-based framework that leverages driving frames to capture facial/hand movements and transfers them to target actors via stable diffusion backbone. It introduces a feedback-driven mechanism for temporal coherence without additional computational costs or secondary training.
Result: The framework enables generation of animations with unlimited duration while maintaining temporal consistency, and introduces a comprehensive large-scale dataset as a new benchmark for human upper-body animation.
Conclusion: TalkingPose addresses key limitations in long-form character animation generation by providing a computationally efficient solution for temporally coherent extended animations, supported by a new benchmark dataset.
Abstract: Recent advancements in diffusion models have significantly improved the realism and generalizability of character-driven animation, enabling the synthesis of high-quality motion from just a single RGB image and a set of driving poses. Nevertheless, generating temporally coherent long-form content remains challenging. Existing approaches are constrained by computational and memory limitations, as they are typically trained on short video segments, thus performing effectively only over limited frame lengths and hindering their potential for extended coherent generation. To address these constraints, we propose TalkingPose, a novel diffusion-based framework specifically designed for producing long-form, temporally consistent human upper-body animations. TalkingPose leverages driving frames to precisely capture expressive facial and hand movements, transferring these seamlessly to a target actor through a stable diffusion backbone. To ensure continuous motion and enhance temporal coherence, we introduce a feedback-driven mechanism built upon image-based diffusion models. Notably, this mechanism does not incur additional computational costs or require secondary training stages, enabling the generation of animations with unlimited duration. Additionally, we introduce a comprehensive, large-scale dataset to serve as a new benchmark for human upper-body animation.
[263] Dual-Projection Fusion for Accurate Upright Panorama Generation in Robotic Vision
Yuhao Shan, Qianyi Yuan, Jingguo Liu, Shigang Li, Jianfeng Li, Tong Chen
Main category: cs.CV
TL;DR: Dual-stream network jointly estimates camera inclination angles and reconstructs upright panoramic images using CNN and ViT branches with adaptive fusion.
Details
Motivation: Panoramic cameras are crucial for robotic vision in sparse environments, but non-upright panoramas from unstable robot postures hinder downstream tasks. IMU-based methods suffer from drift and disturbances, so vision-based alternatives are needed.Method: Dual-stream angle-aware generation network with CNN branch for local geometric structures from equirectangular projections and ViT branch for global contextual cues from cubemap projections. Features are integrated through dual-projection adaptive fusion module with high-frequency enhancement, circular padding, and channel attention mechanisms.
Result: Outperforms existing approaches on SUN360 and M3D datasets in both inclination estimation and upright panorama generation. Ablation studies validate each module’s contribution and highlight task synergy.
Conclusion: Proposed method effectively addresses non-upright panorama correction for robotic vision, offering improved performance over traditional approaches through dual-stream architecture and specialized enhancements for 360° continuity.
Abstract: Panoramic cameras, capable of capturing a 360-degree field of view, are crucial in robotic vision, particularly in environments with sparse features. However, non-upright panoramas due to unstable robot postures hinder downstream tasks. Traditional IMU-based correction methods suffer from drift and external disturbances, while vision-based approaches offer a promising alternative. This study presents a dual-stream angle-aware generation network that jointly estimates camera inclination angles and reconstructs upright panoramic images. The network comprises a CNN branch that extracts local geometric structures from equirectangular projections and a ViT branch that captures global contextual cues from cubemap projections. These are integrated through a dual-projection adaptive fusion module that aligns spatial features across both domains. To further enhance performance, we introduce a high-frequency enhancement block, circular padding, and channel attention mechanisms to preserve 360° continuity and improve geometric sensitivity. Experiments on the SUN360 and M3D datasets demonstrate that our method outperforms existing approaches in both inclination estimation and upright panorama generation. Ablation studies further validate the contribution of each module and highlight the synergy between the two tasks. The code and related datasets can be found at: https://github.com/YuhaoShine/DualProjectionFusion.
[264] ForamDeepSlice: A High-Accuracy Deep Learning Framework for Foraminifera Species Classification from 2D Micro-CT Slices
Abdelghafour Halimi, Ali Alibrahim, Didier Barradas-Bautista, Ronell Sicat, Abdulkader M. Afifi
Main category: cs.CV
TL;DR: Deep learning pipeline achieves 95.64% accuracy for automated classification of 12 foraminifera species using 2D micro-CT slices, with interactive dashboard for practical deployment.
Details
Motivation: To establish AI-assisted micropaleontological identification benchmarks and bridge the gap between deep learning and applied geosciences through automated foraminifera classification.Method: Used 97 micro-CT scanned specimens across 27 species, selected 12 species with sufficient representation, employed specimen-level data splitting (109,617 slices total), evaluated 7 CNN architectures with transfer learning, and created ensemble model combining ConvNeXt-Large and EfficientNetV2-Small.
Result: Final ensemble model achieved 95.64% test accuracy, 99.6% top-3 accuracy, and 0.998 AUC across all species. Developed interactive dashboard with real-time classification and 3D slice matching using SSIM, NCC, and Dice coefficient metrics.
Conclusion: Establishes new benchmarks for AI-assisted micropaleontological identification and provides a fully reproducible framework for foraminifera classification research, bridging deep learning with applied geosciences.
Abstract: This study presents a comprehensive deep learning pipeline for the automated classification of 12 foraminifera species using 2D micro-CT slices derived from 3D scans. We curated a scientifically rigorous dataset comprising 97 micro-CT scanned specimens across 27 species, selecting 12 species with sufficient representation for robust machine learning. To ensure methodological integrity and prevent data leakage, we employed specimen-level data splitting, resulting in 109,617 high-quality 2D slices (44,103 for training, 14,046 for validation, and 51,468 for testing). We evaluated seven state-of-the-art 2D convolutional neural network (CNN) architectures using transfer learning. Our final ensemble model, combining ConvNeXt-Large and EfficientNetV2-Small, achieved a test accuracy of 95.64%, with a top-3 accuracy of 99.6% and an area under the ROC curve (AUC) of 0.998 across all species. To facilitate practical deployment, we developed an interactive advanced dashboard that supports real-time slice classification and 3D slice matching using advanced similarity metrics, including SSIM, NCC, and the Dice coefficient. This work establishes new benchmarks for AI-assisted micropaleontological identification and provides a fully reproducible framework for foraminifera classification research, bridging the gap between deep learning and applied geosciences.
[265] LAHNet: Local Attentive Hashing Network for Point Cloud Registration
Wentao Qu, Xiaoshui Huang, Liang Xiao
Main category: cs.CV
TL;DR: LAHNet introduces a Local Attentive Hashing Network for point cloud registration using local attention with locality bias, Group Transformer for long-range context, and Interaction Transformer for overlap region matching.
Details
Motivation: Most existing point cloud descriptors focus on local information, but broader receptive fields are essential for enhancing feature distinctiveness in point cloud registration.Method: Proposes LAHNet with: 1) Group Transformer using Locality-Sensitive Hashing to partition point clouds into non-overlapping windows for long-range context; 2) Cross-window strategy to expand receptive field; 3) Interaction Transformer with overlap matrix to match overlap regions between point cloud pairs.
Result: Extensive results demonstrate LAHNet learns robust and distinctive features, achieving significant registration results on real-world indoor and outdoor benchmarks.
Conclusion: LAHNet effectively enhances feature distinctiveness through local attention mechanisms and window-based strategies, improving point cloud registration performance across diverse real-world scenarios.
Abstract: Most existing learning-based point cloud descriptors for point cloud registration focus on perceiving local information of point clouds to generate distinctive features. However, a reasonable and broader receptive field is essential for enhancing feature distinctiveness. In this paper, we propose a Local Attentive Hashing Network for point cloud registration, called LAHNet, which introduces a local attention mechanism with the inductive bias of locality of convolution-like operators into point cloud descriptors. Specifically, a Group Transformer is designed to capture reasonable long-range context between points. This employs a linear neighborhood search strategy, Locality-Sensitive Hashing, enabling uniformly partitioning point clouds into non-overlapping windows. Meanwhile, an efficient cross-window strategy is adopted to further expand the reasonable feature receptive field. Furthermore, building on this effective windowing strategy, we propose an Interaction Transformer to enhance the feature interactions of the overlap regions within point cloud pairs. This computes an overlap matrix to match overlap regions between point cloud pairs by representing each window as a global signal. Extensive results demonstrate that LAHNet can learn robust and distinctive features, achieving significant registration results on real-world indoor and outdoor benchmarks.
[266] SceneProp: Combining Neural Network and Markov Random Field for Scene-Graph Grounding
Keita Otani, Tatsuya Harada
Main category: cs.CV
TL;DR: SceneProp reformulates scene-graph grounding as MAP inference in an MRF, using differentiable Belief Propagation to jointly satisfy all constraints in complex visual queries.
Details
Motivation: Standard phrase grounding methods fail with complex compositional queries containing multiple objects and relationships, lacking structural inductive bias. Existing scene-graph grounding methods paradoxically perform worse as query graphs grow, failing to leverage relational context that should make grounding easier.Method: SceneProp reformulates scene-graph grounding as Maximum a Posteriori (MAP) inference in a Markov Random Field (MRF). It performs global inference over the entire query graph to find optimal assignment of image regions to nodes, implemented via differentiable Belief Propagation in an end-to-end framework.
Result: Experiments on four benchmarks show SceneProp significantly outperforms prior work. Critically, its accuracy consistently improves with the size and complexity of query graphs, demonstrating for the first time that more relational context leads to better grounding.
Conclusion: SceneProp successfully addresses the structural deficit in grounding complex visual queries by treating scene-graph grounding as a joint inference problem, enabling models to leverage relational context rather than being hindered by it.
Abstract: Grounding complex, compositional visual queries with multiple objects and relationships is a fundamental challenge for vision-language models. While standard phrase grounding methods excel at localizing single objects, they lack the structural inductive bias to parse intricate relational descriptions, often failing as queries become more descriptive. To address this structural deficit, we focus on scene-graph grounding, a powerful but less-explored formulation where the query is an explicit graph of objects and their relationships. However, existing methods for this task also struggle, paradoxically showing decreased performance as the query graph grows – failing to leverage the very information that should make grounding easier. We introduce SceneProp, a novel method that resolves this issue by reformulating scene-graph grounding as a Maximum a Posteriori (MAP) inference problem in a Markov Random Field (MRF). By performing global inference over the entire query graph, SceneProp finds the optimal assignment of image regions to nodes that jointly satisfies all constraints. This is achieved within an end-to-end framework via a differentiable implementation of the Belief Propagation algorithm. Experiments on four benchmarks show that our dedicated focus on the scene-graph grounding formulation allows SceneProp to significantly outperform prior work. Critically, its accuracy consistently improves with the size and complexity of the query graph, demonstrating for the first time that more relational context can, and should, lead to better grounding. Codes are available at https://github.com/keitaotani/SceneProp.
[267] Binary-Gaussian: Compact and Progressive Representation for 3D Gaussian Segmentation
An Yang, Chenyu Liu, Jun Du, Jianqing Gao, Jia Pan, Jinshui Hu, Baocai Yin, Bing Yin, Cong Liu
Main category: cs.CV
TL;DR: Binary encoding scheme for 3D Gaussian Splatting segmentation reduces memory overhead and enables fine-grained panoptic segmentation through progressive training.
Details
Motivation: Existing 3D-GS segmentation methods suffer from high memory overhead due to high-dimensional category features, and struggle with fine-grained segmentation due to label space congestion and lack of multi-granularity control.Method: 1) Coarse-to-fine binary encoding scheme compresses per-Gaussian category features into single integers via binary-to-decimal mapping; 2) Progressive training strategy decomposes panoptic segmentation into independent sub-tasks; 3) Fine-tuning opacity during segmentation training to address photometric rendering vs. semantic segmentation incompatibility.
Result: Achieves state-of-the-art segmentation performance on multiple benchmarks while significantly reducing memory consumption and accelerating inference.
Conclusion: The proposed binary encoding and progressive training approach effectively addresses memory and fine-grained segmentation challenges in 3D-GS, making it a practical solution for semantic segmentation tasks.
Abstract: 3D Gaussian Splatting (3D-GS) has emerged as an efficient 3D representation and a promising foundation for semantic tasks like segmentation. However, existing 3D-GS-based segmentation methods typically rely on high-dimensional category features, which introduce substantial memory overhead. Moreover, fine-grained segmentation remains challenging due to label space congestion and the lack of stable multi-granularity control mechanisms. To address these limitations, we propose a coarse-to-fine binary encoding scheme for per-Gaussian category representation, which compresses each feature into a single integer via the binary-to-decimal mapping, drastically reducing memory usage. We further design a progressive training strategy that decomposes panoptic segmentation into a series of independent sub-tasks, reducing inter-class conflicts and thereby enhancing fine-grained segmentation capability. Additionally, we fine-tune opacity during segmentation training to address the incompatibility between photometric rendering and semantic segmentation, which often leads to foreground-background confusion. Extensive experiments on multiple benchmarks demonstrate that our method achieves state-of-the-art segmentation performance while significantly reducing memory consumption and accelerating inference.
[268] Adaptive Evidential Learning for Temporal-Semantic Robustness in Moment Retrieval
Haojian Huang, Kaijing Ma, Jin Chen, Haodong Chen, Zhou Wu, Xianghao Zang, Han Fang, Chao Ban, Hao Sun, Mulin Chen, Zhongjiang He
Main category: cs.CV
TL;DR: Proposes DEMR framework for moment retrieval using debiased evidential learning with Reflective Flipped Fusion and Geom-regularizer to improve uncertainty estimation and handle modality imbalance.
Details
Motivation: Traditional moment retrieval methods struggle with fine-grained information, deterministic reasoning, and aligning with complex/ambiguous moments. Existing approaches lack adaptability for complex video scenarios and suffer from modality imbalance issues in uncertainty estimation.Method: DEMR framework with: 1) Reflective Flipped Fusion (RFF) block for cross-modal alignment, 2) query reconstruction task to enhance text sensitivity, and 3) Geom-regularizer to refine uncertainty predictions for adaptive alignment with difficult moments.
Result: Extensive testing shows significant improvements in effectiveness, robustness, and interpretability on standard datasets and debiased datasets ActivityNet-CD and Charades-CD.
Conclusion: DEMR provides a promising solution for temporal-semantic robustness in moment retrieval by addressing uncertainty estimation biases and modality imbalance through debiased evidential learning.
Abstract: In the domain of moment retrieval, accurately identifying temporal segments within videos based on natural language queries remains challenging. Traditional methods often employ pre-trained models that struggle with fine-grained information and deterministic reasoning, leading to difficulties in aligning with complex or ambiguous moments. To overcome these limitations, we explore Deep Evidential Regression (DER) to construct a vanilla Evidential baseline. However, this approach encounters two major issues: the inability to effectively handle modality imbalance and the structural differences in DER’s heuristic uncertainty regularizer, which adversely affect uncertainty estimation. This misalignment results in high uncertainty being incorrectly associated with accurate samples rather than challenging ones. Our observations indicate that existing methods lack the adaptability required for complex video scenarios. In response, we propose Debiased Evidential Learning for Moment Retrieval (DEMR), a novel framework that incorporates a Reflective Flipped Fusion (RFF) block for cross-modal alignment and a query reconstruction task to enhance text sensitivity, thereby reducing bias in uncertainty estimation. Additionally, we introduce a Geom-regularizer to refine uncertainty predictions, enabling adaptive alignment with difficult moments and improving retrieval accuracy. Extensive testing on standard datasets and debiased datasets ActivityNet-CD and Charades-CD demonstrates significant enhancements in effectiveness, robustness, and interpretability, positioning our approach as a promising solution for temporal-semantic robustness in moment retrieval. The code is publicly available at https://github.com/KaijingOfficial/DEMR.
[269] Efficient and Scalable Monocular Human-Object Interaction Motion Reconstruction
Boran Wen, Ye Lu, Keyan Wan, Sirui Wang, Jiahong Zhou, Junxuan Liang, Xinpeng Liu, Bang Xiao, Dingbang Huang, Ruiyang Liu, Yong-Lu Li
Main category: cs.CV
TL;DR: 4DHOISolver is an optimization framework for reconstructing 4D human-object interactions from monocular videos using sparse human-in-the-loop contact annotations, enabling creation of Open4DHOI dataset and demonstrating motion imitation capabilities.
Details
Motivation: Monocular internet videos offer vast, diverse human-object interaction data but extracting accurate 4D interaction data remains an unsolved challenge. Current methods lack scalability and accuracy for real-world applications.Method: 4DHOISolver: an efficient optimization framework that constrains ill-posed 4D HOI reconstruction using sparse human-in-the-loop contact point annotations while maintaining spatio-temporal coherence and physical plausibility.
Result: Created Open4DHOI dataset with 144 object types and 103 actions. Demonstrated RL-based agents can imitate recovered motions. Benchmark shows existing 3D foundation models cannot automatically predict precise human-object contact correspondences.
Conclusion: Human-in-the-loop strategy is currently necessary for accurate 4D HOI reconstruction. The work provides a valuable dataset and framework while highlighting the open challenge of automated contact prediction for the community.
Abstract: Generalized robots must learn from diverse, large-scale human-object interactions (HOI) to operate robustly in the real world. Monocular internet videos offer a nearly limitless and readily available source of data, capturing an unparalleled diversity of human activities, objects, and environments. However, accurately and scalably extracting 4D interaction data from these in-the-wild videos remains a significant and unsolved challenge. Thus, in this work, we introduce 4DHOISolver, a novel and efficient optimization framework that constrains the ill-posed 4D HOI reconstruction problem by leveraging sparse, human-in-the-loop contact point annotations, while maintaining high spatio-temporal coherence and physical plausibility. Leveraging this framework, we introduce Open4DHOI, a new large-scale 4D HOI dataset featuring a diverse catalog of 144 object types and 103 actions. Furthermore, we demonstrate the effectiveness of our reconstructions by enabling an RL-based agent to imitate the recovered motions. However, a comprehensive benchmark of existing 3D foundation models indicates that automatically predicting precise human-object contact correspondences remains an unsolved problem, underscoring the immediate necessity of our human-in-the-loop strategy while posing an open challenge to the community. Data and code will be publicly available at https://wenboran2002.github.io/open4dhoi/
[270] MM-ACT: Learn from Multimodal Parallel Generation to Act
Haotian Liang, Xinyi Chen, Bin Wang, Mingkang Chen, Yitian Liu, Yuhao Zhang, Zanxin Chen, Tianshuo Yang, Yilun Chen, Jiangmiao Pang, Dong Liu, Xiaokang Yang, Yao Mu, Wenqi Shao, Ping Luo
Main category: cs.CV
TL;DR: MM-ACT is a unified Vision-Language-Action model that integrates text, image, and action in shared token space, using re-mask parallel decoding for text/image and one-step parallel decoding for actions, achieving state-of-the-art performance on robotic tasks.
Details
Motivation: Generalist robotic policies need both semantic understanding for task planning and predictive capabilities for environment interaction. Current approaches often separate these capabilities, limiting unified reasoning and action generation.Method: MM-ACT integrates text, image, and action in shared token space with Context-Shared Multimodal Learning. Uses re-mask parallel decoding for text/image generation and one-step parallel decoding for action generation to improve efficiency.
Result: Achieves 96.3% success rate on LIBERO simulation, 72.0% across three real Franka tasks, and 52.38% across eight bimanual tasks on RoboTwin2.0, with additional 9.25% gain from cross-modal learning.
Conclusion: MM-ACT demonstrates effective unified multimodal learning for robotics, showing strong in-domain and out-of-domain performance with cross-modal learning providing significant improvements. The shared token space approach enables better integration of semantic understanding and action generation.
Abstract: A generalist robotic policy needs both semantic understanding for task planning and the ability to interact with the environment through predictive capabilities. To tackle this, we present MM-ACT, a unified Vision-Language-Action (VLA) model that integrates text, image, and action in shared token space and performs generation across all three modalities. MM-ACT adopts a re-mask parallel decoding strategy for text and image generation, and employs a one-step parallel decoding strategy for action generation to improve efficiency. We introduce Context-Shared Multimodal Learning, a unified training paradigm that supervises generation in all three modalities from a shared context, enhancing action generation through cross-modal learning. Experiments were conducted on the LIBERO simulation and Franka real-robot setups as well as RoboTwin2.0 to assess in-domain and out-of-domain performances respectively. Our approach achieves a success rate of 96.3% on LIBERO, 72.0% across three tasks of real Franka, and 52.38% across eight bimanual tasks of RoboTwin2.0 with an additional gain of 9.25% from cross-modal learning. We release our codes, models and data at https://github.com/HHYHRHY/MM-ACT.
[271] PhotoFramer: Multi-modal Image Composition Instruction
Zhiyuan You, Ke Wang, He Zhang, Xin Cai, Jinjin Gu, Tianfan Xue, Chao Dong, Zhoutong Zhang
Main category: cs.CV
TL;DR: PhotoFramer is a multi-modal framework that provides composition guidance by generating both textual instructions and example images to help users improve poorly composed photos.
Details
Motivation: Many casual users struggle with photo composition, needing guidance to frame well-composed images. Existing tools lack the ability to provide both actionable textual advice and visual examples.Method: The framework uses a multi-modal model trained on a curated large-scale dataset. Composition guidance is organized hierarchically into shift, zoom-in, and view-change tasks. Training data is synthesized by degrading well-composed photos into poor ones using a learned degradation model applied to multi-view datasets and expert photos.
Result: Experiments show that textual instructions effectively steer image composition, and combining them with visual exemplars consistently outperforms exemplar-only baselines.
Conclusion: PhotoFramer represents a practical step toward composition assistants that make expert photographic knowledge accessible to everyday users, with released code, models, and datasets.
Abstract: Composition matters during the photo-taking process, yet many casual users struggle to frame well-composed images. To provide composition guidance, we introduce PhotoFramer, a multi-modal composition instruction framework. Given a poorly composed image, PhotoFramer first describes how to improve the composition in natural language and then generates a well-composed example image. To train such a model, we curate a large-scale dataset. Inspired by how humans take photos, we organize composition guidance into a hierarchy of sub-tasks: shift, zoom-in, and view-change tasks. Shift and zoom-in data are sampled from existing cropping datasets, while view-change data are obtained via a two-stage pipeline. First, we sample pairs with varying viewpoints from multi-view datasets, and train a degradation model to transform well-composed photos into poorly composed ones. Second, we apply this degradation model to expert-taken photos to synthesize poor images to form training pairs. Using this dataset, we finetune a model that jointly processes and generates both text and images, enabling actionable textual guidance with illustrative examples. Extensive experiments demonstrate that textual instructions effectively steer image composition, and coupling them with exemplars yields consistent improvements over exemplar-only baselines. PhotoFramer offers a practical step toward composition assistants that make expert photographic priors accessible to everyday users. Codes, model weights, and datasets have been released in https://zhiyuanyou.github.io/photoframer.
[272] S2AM3D: Scale-controllable Part Segmentation of 3D Point Cloud
Han Su, Tianyu Huang, Zichen Wan, Xiaohe Wu, Wangmeng Zuo
Main category: cs.CV
TL;DR: S2AM3D: A novel approach for part-level point cloud segmentation that combines 2D segmentation priors with 3D consistent supervision, featuring point-consistent part encoder, scale-aware prompt decoder, and a new large-scale dataset.
Details
Motivation: Existing part-level point cloud segmentation faces two major challenges: 1) Native 3D models lack generalization due to data scarcity, and 2) Introducing 2D pre-trained knowledge often leads to inconsistent segmentation results across different views.Method: Proposes S2AM3D with three key components: 1) Point-consistent part encoder that aggregates multi-view 2D features through native 3D contrastive learning for globally consistent point features, 2) Scale-aware prompt decoder enabling real-time adjustment of segmentation granularity via continuous scale signals, and 3) A large-scale, high-quality part-level point cloud dataset with over 100k samples.
Result: Extensive experiments demonstrate that S2AM3D achieves leading performance across multiple evaluation settings, exhibiting exceptional robustness and controllability when handling complex structures and parts with significant size variations.
Conclusion: S2AM3D effectively addresses the challenges of 3D part segmentation by combining 2D priors with 3D consistency, providing a robust and controllable solution that outperforms existing methods while handling complex structures and varying part sizes.
Abstract: Part-level point cloud segmentation has recently attracted significant attention in 3D computer vision. Nevertheless, existing research is constrained by two major challenges: native 3D models lack generalization due to data scarcity, while introducing 2D pre-trained knowledge often leads to inconsistent segmentation results across different views. To address these challenges, we propose S2AM3D, which incorporates 2D segmentation priors with 3D consistent supervision. We design a point-consistent part encoder that aggregates multi-view 2D features through native 3D contrastive learning, producing globally consistent point features. A scale-aware prompt decoder is then proposed to enable real-time adjustment of segmentation granularity via continuous scale signals. Simultaneously, we introduce a large-scale, high-quality part-level point cloud dataset with more than 100k samples, providing ample supervision signals for model training. Extensive experiments demonstrate that S2AM3D achieves leading performance across multiple evaluation settings, exhibiting exceptional robustness and controllability when handling complex structures and parts with significant size variations.
[273] Provenance-Driven Reliable Semantic Medical Image Vector Reconstruction via Lightweight Blockchain-Verified Latent Fingerprints
Mohsin Rasheed, Abdullah Al-Mamun
Main category: cs.CV
TL;DR: Semantic-aware medical image reconstruction with blockchain provenance for trustworthy AI-assisted diagnosis.
Details
Motivation: Real-world medical images often have corruption, noise, and potential tampering, challenging AI reliability. Conventional reconstruction methods may produce visually plausible outputs but compromise anatomical fidelity, impacting clinical outcomes.Method: Semantic-aware medical image reconstruction framework integrating high-level latent embeddings with hybrid U-Net architecture to preserve clinically relevant structures. Includes lightweight blockchain-based provenance layer using scale-free graph design for verifiable recording of reconstruction events.
Result: Extensive evaluation across multiple datasets and corruption types demonstrates improved structural consistency, restoration accuracy, and provenance integrity compared to existing approaches.
Conclusion: The solution advances dependable AI for medical imaging by uniting semantic-guided reconstruction with secure traceability, enhancing both diagnostic confidence and regulatory compliance in healthcare environments.
Abstract: Medical imaging is essential for clinical diagnosis, yet real-world data frequently suffers from corruption, noise, and potential tampering, challenging the reliability of AI-assisted interpretation. Conventional reconstruction techniques prioritize pixel-level recovery and may produce visually plausible outputs while compromising anatomical fidelity, an issue that can directly impact clinical outcomes. We propose a semantic-aware medical image reconstruction framework that integrates high-level latent embeddings with a hybrid U-Net architecture to preserve clinically relevant structures during restoration. To ensure trust and accountability, we incorporate a lightweight blockchain-based provenance layer using scale-free graph design, enabling verifiable recording of each reconstruction event without imposing significant overhead. Extensive evaluation across multiple datasets and corruption types demonstrates improved structural consistency, restoration accuracy, and provenance integrity compared with existing approaches. By uniting semantic-guided reconstruction with secure traceability, our solution advances dependable AI for medical imaging, enhancing both diagnostic confidence and regulatory compliance in healthcare environments.
[274] LISA-3D: Lifting Language-Image Segmentation to 3D via Multi-View Consistency
Zhongbin Guo, Jiahe Liu, Wenyu Gao, Yushan Li, Chengzhi Li, Ping Jian
Main category: cs.CV
TL;DR: LISA-3D is a two-stage framework that lifts 2D language-image segmentation to 3D reconstruction using geometry-aware LoRA layers and frozen SAM-3D, achieving improved language-to-3D accuracy with minimal parameter adaptation.
Details
Motivation: Text-driven 3D reconstruction requires mask generators that understand open-vocabulary instructions while maintaining cross-view consistency, which existing methods struggle to achieve simultaneously.Method: Two-stage framework: 1) Retrofit LISA with geometry-aware LoRA layers for cross-view consistent masks, 2) Use frozen SAM-3D reconstructor with RGBA prompts (RGB + masks) to output Gaussian splats or textured meshes without retraining.
Result: Improves language-to-3D accuracy by up to +15.6 points over single-view baselines on ScanRefer and Nr3D datasets, adapting only 11.6M parameters while supporting zero-shot deployment on unseen categories.
Conclusion: LISA-3D provides a practical, modular, and data-efficient recipe for language-guided 3D content creation by effectively combining 2D language-image segmentation with 3D reconstruction without requiring 3D-text supervision.
Abstract: Text-driven 3D reconstruction demands a mask generator that simultaneously understands open-vocabulary instructions and remains consistent across viewpoints. We present LISA-3D, a two-stage framework that lifts language-image segmentation into 3D by retrofitting the instruction-following model LISA with geometry-aware Low-Rank Adaptation (LoRA) layers and reusing a frozen SAM-3D reconstructor. During training we exploit off-the-shelf RGB-D sequences and their camera poses to build a differentiable reprojection loss that enforces cross-view agreement without requiring any additional 3D-text supervision. The resulting masks are concatenated with RGB images to form RGBA prompts for SAM-3D, which outputs Gaussian splats or textured meshes without retraining. Across ScanRefer and Nr3D, LISA-3D improves language-to-3D accuracy by up to +15.6 points over single-view baselines while adapting only 11.6M parameters. The system is modular, data-efficient, and supports zero-shot deployment on unseen categories, providing a practical recipe for language-guided 3D content creation. Our code will be available at https://github.com/binisalegend/LISA-3D.
[275] Lotus-2: Advancing Geometric Dense Prediction with Powerful Image Generative Model
Jing He, Haodong Li, Mingzhi Sheng, Ying-Cong Chen
Main category: cs.CV
TL;DR: Lotus-2: A two-stage deterministic framework that adapts diffusion models for stable, accurate geometric dense prediction (depth/normal estimation) using only 59K training samples.
Details
Motivation: Monocular geometric inference is ill-posed due to appearance ambiguity. Existing discriminative models are limited by data scale/quality, while diffusion models have powerful world priors but their stochastic generative formulation is suboptimal for deterministic geometric prediction.Method: Two-stage framework: 1) Core predictor with single-step deterministic formulation, clean-data objective, and lightweight local continuity module for globally coherent structures. 2) Detail sharpener with constrained multi-step rectified-flow refinement within the manifold to enhance fine-grained geometry.
Result: Using only 59K training samples (<1% of existing datasets), achieves SOTA in monocular depth estimation and competitive surface normal prediction, demonstrating diffusion models can serve as deterministic world priors.
Conclusion: Diffusion models can be adapted as deterministic world priors for high-quality geometric reasoning, bridging traditional discriminative and generative paradigms with minimal training data.
Abstract: Recovering pixel-wise geometric properties from a single image is fundamentally ill-posed due to appearance ambiguity and non-injective mappings between 2D observations and 3D structures. While discriminative regression models achieve strong performance through large-scale supervision, their success is bounded by the scale, quality and diversity of available data and limited physical reasoning. Recent diffusion models exhibit powerful world priors that encode geometry and semantics learned from massive image-text data, yet directly reusing their stochastic generative formulation is suboptimal for deterministic geometric inference: the former is optimized for diverse and high-fidelity image generation, whereas the latter requires stable and accurate predictions. In this work, we propose Lotus-2, a two-stage deterministic framework for stable, accurate and fine-grained geometric dense prediction, aiming to provide an optimal adaption protocol to fully exploit the pre-trained generative priors. Specifically, in the first stage, the core predictor employs a single-step deterministic formulation with a clean-data objective and a lightweight local continuity module (LCM) to generate globally coherent structures without grid artifacts. In the second stage, the detail sharpener performs a constrained multi-step rectified-flow refinement within the manifold defined by the core predictor, enhancing fine-grained geometry through noise-free deterministic flow matching. Using only 59K training samples, less than 1% of existing large-scale datasets, Lotus-2 establishes new state-of-the-art results in monocular depth estimation and highly competitive surface normal prediction. These results demonstrate that diffusion models can serve as deterministic world priors, enabling high-quality geometric reasoning beyond traditional discriminative and generative paradigms.
[276] TRoVe: Discovering Error-Inducing Static Feature Biases in Temporal Vision-Language Models
Maya Varma, Jean-Benoit Delbrouck, Sophie Ostmeier, Akshay Chaudhari, Curtis Langlotz
Main category: cs.CV
TL;DR: TRoVe is an automated method to discover error-inducing static feature biases in temporal vision-language models, improving bias detection by 28.6% over baselines and helping improve model performance.
Details
Motivation: Vision-language models may rely on static feature biases (like background or object features) rather than dynamic visual changes when making predictions on temporal understanding tasks, leading to systematic errors that need identification before real-world deployment.Method: TRoVe extracts candidate static features from annotated validation datasets and scores each feature based on: (1) its effect on classification errors, and (2) the extent to which the VLM relies on the feature for predictions.
Result: TRoVe achieves 28.6% improvement over the closest baseline in identifying error-inducing static feature biases. When applied to 7 off-the-shelf VLMs and 2 temporal understanding tasks, it surfaces previously-unknown biases and shows this knowledge can improve test-time performance.
Conclusion: TRoVe provides an effective automated approach for discovering static feature biases in temporal VLMs, enabling better model understanding and potential performance improvements through bias awareness.
Abstract: Vision-language models (VLMs) have made great strides in addressing temporal understanding tasks, which involve characterizing visual changes across a sequence of images. However, recent works have suggested that when making predictions, VLMs may rely on static feature biases, such as background or object features, rather than dynamic visual changes. Static feature biases are a type of shortcut and can contribute to systematic prediction errors on downstream tasks; as a result, identifying and characterizing error-inducing static feature biases is critical prior to real-world model deployment. In this work, we introduce TRoVe, an automated approach for discovering error-inducing static feature biases learned by temporal VLMs. Given a trained VLM and an annotated validation dataset associated with a downstream classification task, TRoVe extracts candidate static features from the dataset and scores each feature by (i) the effect of the feature on classification errors as well as (ii) the extent to which the VLM relies on the feature when making predictions. In order to quantitatively evaluate TRoVe, we introduce an evaluation framework consisting of 101 trained temporal VLMs paired with ground-truth annotations for learned static feature biases. We use this framework to demonstrate that TRoVe can accurately identify error-inducing static feature biases in VLMs, achieving a 28.6% improvement over the closest baseline. Finally, we apply TRoVe to 7 off-the-shelf VLMs and 2 temporal understanding tasks, surfacing previously-unknown static feature biases and demonstrating that knowledge of learned biases can aid in improving model performance at test time. Our code is available at https://github.com/Stanford-AIMI/TRoVe.
[277] Parameter Reduction Improves Vision Transformers: A Comparative Study of Sharing and Width Reduction
Anantha Padmanaban Krishna Kumar
Main category: cs.CV
TL;DR: The paper shows that reducing MLP capacity in ViT-B/16 through parameter sharing or width reduction can improve performance and training stability, suggesting the model operates in an overparameterized regime.
Details
Motivation: While scaling laws suggest bigger models perform better, the authors question whether Vision Transformers are always monotonically improving with scale. They investigate if ViT-B/16 on ImageNet-1K might be overparameterized, particularly in MLP blocks, and explore whether parameter reduction strategies could maintain or even improve performance.Method: Two simple parameter-reduction strategies applied to MLP blocks of ViT-B/16: 1) GroupedMLP - shares MLP weights between adjacent transformer blocks (32.7% parameter reduction), and 2) ShallowMLP - halves the MLP hidden dimension (32.7% parameter reduction). Both maintain baseline computational cost or improve inference throughput.
Result: GroupedMLP achieves 81.47% top-1 accuracy (vs baseline 81.05%) with same computational cost. ShallowMLP reaches 81.25% accuracy with 38% increase in inference throughput. Both show substantially improved training stability, reducing peak-to-final accuracy degradation from 0.47% to 0.03-0.06%.
Conclusion: ViT-B/16 on ImageNet-1K operates in an overparameterized regime where MLP capacity can be reduced without harming performance. Architectural constraints like parameter sharing and reduced width act as useful inductive biases, highlighting the importance of parameter allocation in Vision Transformer design.
Abstract: Although scaling laws and many empirical results suggest that increasing the size of Vision Transformers often improves performance, model accuracy and training behavior are not always monotonically increasing with scale. Focusing on ViT-B/16 trained on ImageNet-1K, we study two simple parameter-reduction strategies applied to the MLP blocks, each removing 32.7% of the baseline parameters. Our \emph{GroupedMLP} variant shares MLP weights between adjacent transformer blocks and achieves 81.47% top-1 accuracy while maintaining the baseline computational cost. Our \emph{ShallowMLP} variant halves the MLP hidden dimension and reaches 81.25% top-1 accuracy with a 38% increase in inference throughput. Both models outperform the 86.6M-parameter baseline (81.05%) and exhibit substantially improved training stability, reducing peak-to-final accuracy degradation from 0.47% to the range 0.03% to 0.06%. These results suggest that, for ViT-B/16 on ImageNet-1K with a standard training recipe, the model operates in an overparameterized regime in which MLP capacity can be reduced without harming performance and can even slightly improve it. More broadly, our findings suggest that architectural constraints such as parameter sharing and reduced width may act as useful inductive biases, and highlight the importance of how parameters are allocated when designing Vision Transformers. All code is available at: https://github.com/AnanthaPadmanaban-KrishnaKumar/parameter-efficient-vit-mlps.
[278] Accelerating Inference of Masked Image Generators via Reinforcement Learning
Pranav Subbaraman, Shufan Li, Siyan Zhao, Aditya Grover
Main category: cs.CV
TL;DR: Speed-RL uses reinforcement learning to accelerate masked generative models by 3x while maintaining image quality, unlike traditional distillation methods.
Details
Motivation: Masked Generative Models (MGMs) produce high-fidelity images but require many sampling steps, resulting in slow inference speed. Current distillation methods treat acceleration as distribution matching, but the authors propose a more direct reinforcement learning approach.Method: Proposes Speed-RL, a reinforcement learning paradigm that combines quality and speed rewards to finetune pretrained MGMs. Instead of matching teacher-student distributions, it directly optimizes for generating high-quality images in fewer steps using RL.
Result: The method accelerates base models by 3x while maintaining comparable image quality, as demonstrated through extensive experiments.
Conclusion: Speed-RL offers an effective reinforcement learning alternative to conventional distillation for accelerating masked generative models, achieving significant speedup without sacrificing image quality.
Abstract: Masked Generative Models (MGM)s demonstrate strong capabilities in generating high-fidelity images. However, they need many sampling steps to create high-quality generations, resulting in slow inference speed. In this work, we propose Speed-RL, a novel paradigm for accelerating a pretrained MGMs to generate high-quality images in fewer steps. Unlike conventional distillation methods which formulate the acceleration problem as a distribution matching problem, where a few-step student model is trained to match the distribution generated by a many-step teacher model, we consider this problem as a reinforcement learning problem. Since the goal of acceleration is to generate high quality images in fewer steps, we can combine a quality reward with a speed reward and finetune the base model using reinforcement learning with the combined reward as the optimization target. Through extensive experiments, we show that the proposed method was able to accelerate the base model by a factor of 3x while maintaining comparable image quality.
[279] CycliST: A Video Language Model Benchmark for Reasoning on Cyclical State Transitions
Simon Kohaut, Daniel Ochs, Shun Zhang, Benedict Flade, Julian Eggert, Kristian Kersting, Devendra Singh Dhami
Main category: cs.CV
TL;DR: CycliST is a new benchmark dataset for evaluating Video Language Models’ ability to reason about cyclical state transitions in synthetic videos with periodic patterns, revealing current VLMs’ limitations in temporal understanding and quantitative reasoning.
Details
Motivation: Current Video Language Models lack proper understanding of cyclical dynamics and temporal patterns in visual scenes, creating a need for targeted evaluation of their spatio-temporal cognition abilities.Method: Created synthetic video sequences with periodic patterns in object motion and visual attributes, using a tiered evaluation system with progressive difficulty through variations in cyclic objects, scene clutter, and lighting conditions.
Result: Current VLMs (both open-source and proprietary) struggle with cyclical dynamics, lack temporal understanding, cannot extract quantitative insights, and show no consistent performance correlation with model size or architecture.
Conclusion: CycliST provides a comprehensive evaluation framework that highlights significant technical gaps in current VLMs and paves the way for improved visual reasoning models capable of understanding periodic patterns.
Abstract: We present CycliST, a novel benchmark dataset designed to evaluate Video Language Models (VLM) on their ability for textual reasoning over cyclical state transitions. CycliST captures fundamental aspects of real-world processes by generating synthetic, richly structured video sequences featuring periodic patterns in object motion and visual attributes. CycliST employs a tiered evaluation system that progressively increases difficulty through variations in the number of cyclic objects, scene clutter, and lighting conditions, challenging state-of-the-art models on their spatio-temporal cognition. We conduct extensive experiments with current state-of-the-art VLMs, both open-source and proprietary, and reveal their limitations in generalizing to cyclical dynamics such as linear and orbital motion, as well as time-dependent changes in visual attributes like color and scale. Our results demonstrate that present-day VLMs struggle to reliably detect and exploit cyclic patterns, lack a notion of temporal understanding, and are unable to extract quantitative insights from scenes, such as the number of objects in motion, highlighting a significant technical gap that needs to be addressed. More specifically, we find no single model consistently leads in performance: neither size nor architecture correlates strongly with outcomes, and no model succeeds equally well across all tasks. By providing a targeted challenge and a comprehensive evaluation framework, CycliST paves the way for visual reasoning models that surpass the state-of-the-art in understanding periodic patterns.
[280] Learning Eigenstructures of Unstructured Data Manifolds
Roy Velich, Arkadi Piven, David Bensaïd, Daniel Cremers, Thomas Dagès, Ron Kimmel
Main category: cs.CV
TL;DR: A novel learning framework that directly learns spectral bases from unstructured data without traditional operator construction or eigendecomposition, recovering spectral bases, eigenvalues, and metric sampling density in a unified manner.
Details
Motivation: Traditional spectral analysis in geometry processing requires manual operator selection, discretization, and expensive eigendecomposition, which becomes challenging for unstructured data and high-dimensional manifolds. There's a need for a more flexible, data-driven approach that can handle arbitrary datasets without assumptions about manifold structure.Method: The method trains a network to decompose an implicit approximation operator by minimizing reconstruction error in the learned basis over a distribution of probe functions. Grounded in optimal-approximation theory, it learns spectral basis, eigenvalues, and metric sampling density simultaneously without explicit operator construction.
Result: The approach successfully recovers meaningful spectral bases that resemble those of the Laplacian operator on both 3D surface point clouds and high-dimensional image manifolds, without explicit operator construction. It scales to arbitrary datasets of any dimension.
Conclusion: This learning-based framework provides a principled, data-driven alternative to conventional spectral analysis pipelines, eliminating the need for operator selection, discretization, and eigendecomposition. It opens new possibilities for geometry processing on unstructured data, especially in high-dimensional spaces.
Abstract: We introduce a novel framework that directly learns a spectral basis for shape and manifold analysis from unstructured data, eliminating the need for traditional operator selection, discretization, and eigensolvers. Grounded in optimal-approximation theory, we train a network to decompose an implicit approximation operator by minimizing the reconstruction error in the learned basis over a chosen distribution of probe functions. For suitable distributions, they can be seen as an approximation of the Laplacian operator and its eigendecomposition, which are fundamental in geometry processing. Furthermore, our method recovers in a unified manner not only the spectral basis, but also the implicit metric’s sampling density and the eigenvalues of the underlying operator. Notably, our unsupervised method makes no assumption on the data manifold, such as meshing or manifold dimensionality, allowing it to scale to arbitrary datasets of any dimension. On point clouds lying on surfaces in 3D and high-dimensional image manifolds, our approach yields meaningful spectral bases, that can resemble those of the Laplacian, without explicit construction of an operator. By replacing the traditional operator selection, construction, and eigendecomposition with a learning-based approach, our framework offers a principled, data-driven alternative to conventional pipelines. This opens new possibilities in geometry processing for unstructured data, particularly in high-dimensional spaces.
[281] Structural Prognostic Event Modeling for Multimodal Cancer Survival Analysis
Yilan Zhang, Li Nanbo, Changchun Yang, Jürgen Schmidhuber, Xin Gao
Main category: cs.CV
TL;DR: SlotSPE: A slot-based framework that compresses multimodal histology-gene data into compact, distinctive slots to model sparse structural prognostic events for improved cancer survival prediction.
Details
Motivation: Current multimodal approaches struggle to efficiently model intra- and inter-modal interactions due to high-dimensional inputs. Critical prognostic events are sparse, patient-specific, and unannotated, making them difficult to uncover despite being key determinants of patient outcomes.Method: SlotSPE uses slot attention to compress multimodal inputs into compact, modality-specific sets of mutually distinctive slots, inspired by factorial coding. These slots represent prognostic events and enable efficient modeling of complex interactions while allowing incorporation of biological priors.
Result: Outperforms existing methods on 8 out of 10 cancer benchmarks with 2.9% overall improvement. Remains robust under missing genomic data and provides improved interpretability through structured event decomposition.
Conclusion: SlotSPE effectively models sparse structural prognostic events in multimodal cancer data, achieving superior survival prediction performance while offering enhanced interpretability and robustness to missing data.
Abstract: The integration of histology images and gene profiles has shown great promise for improving survival prediction in cancer. However, current approaches often struggle to model intra- and inter-modal interactions efficiently and effectively due to the high dimensionality and complexity of the inputs. A major challenge is capturing critical prognostic events that, though few, underlie the complexity of the observed inputs and largely determine patient outcomes. These events, manifested as high-level structural signals such as spatial histologic patterns or pathway co-activations, are typically sparse, patient-specific, and unannotated, making them inherently difficult to uncover. To address this, we propose SlotSPE, a slot-based framework for structural prognostic event modeling. Specifically, inspired by the principle of factorial coding, we compress each patient’s multimodal inputs into compact, modality-specific sets of mutually distinctive slots using slot attention. By leveraging these slot representations as encodings for prognostic events, our framework enables both efficient and effective modeling of complex intra- and inter-modal interactions, while also facilitating seamless incorporation of biological priors that enhance prognostic relevance. Extensive experiments on ten cancer benchmarks show that SlotSPE outperforms existing methods in 8 out of 10 cohorts, achieving an overall improvement of 2.9%. It remains robust under missing genomic data and delivers markedly improved interpretability through structured event decomposition.
[282] OmniFD: A Unified Model for Versatile Face Forgery Detection
Haotian Liu, Haoyu Chen, Chenhui Pan, You Hu, Guoying Zhao, Xiaobai Li
Main category: cs.CV
TL;DR: OmniFD is a unified framework that jointly handles four face forgery detection tasks (image/video classification, spatial/temporal localization) in a single model, achieving better performance with fewer parameters and training time.
Details
Motivation: Current face forgery detection approaches use task-specific models with independent architectures, causing computational redundancy and ignoring correlations across related tasks.Method: Three components: (1) shared Swin Transformer encoder for unified 4D spatiotemporal representations, (2) cross-task interaction module with learnable queries for attention-based inter-task dependencies, (3) lightweight decoding heads for task-specific predictions.
Result: Superior performance across benchmarks with high efficiency: 4.63% video classification improvement with image data, 63% parameter reduction, 50% training time reduction compared to task-specific models.
Conclusion: OmniFD establishes a practical, generalizable solution for comprehensive face forgery detection by unifying multiple tasks in one framework, leveraging multi-task learning for generalized representations and knowledge transfer.
Abstract: Face forgery detection encompasses multiple critical tasks, including identifying forged images and videos and localizing manipulated regions and temporal segments. Current approaches typically employ task-specific models with independent architectures, leading to computational redundancy and ignoring potential correlations across related tasks. We introduce OmniFD, a unified framework that jointly addresses four core face forgery detection tasks within a single model, i.e., image and video classification, spatial localization, and temporal localization. Our architecture consists of three principal components: (1) a shared Swin Transformer encoder that extracts unified 4D spatiotemporal representations from both images and video inputs, (2) a cross-task interaction module with learnable queries that dynamically captures inter-task dependencies through attention-based reasoning, and (3) lightweight decoding heads that transform refined representations into corresponding predictions for all FFD tasks. Extensive experiments demonstrate OmniFD’s advantage over task-specific models. Its unified design leverages multi-task learning to capture generalized representations across tasks, especially enabling fine-grained knowledge transfer that facilitates other tasks. For example, video classification accuracy improves by 4.63% when image data are incorporated. Furthermore, by unifying images, videos and the four tasks within one framework, OmniFD achieves superior performance across diverse benchmarks with high efficiency and scalability, e.g., reducing 63% model parameters and 50% training time. It establishes a practical and generalizable solution for comprehensive face forgery detection in real-world applications. The source code is made available at https://github.com/haotianll/OmniFD.
[283] Weakly Supervised Continuous Micro-Expression Intensity Estimation Using Temporal Deep Neural Network
Riyadh Mohammed Almushrafy
Main category: cs.CV
TL;DR: A unified framework for continuous micro-expression intensity estimation using only weak temporal labels (onset, apex, offset) without frame-level annotations.
Details
Motivation: Most prior work focuses on classifying discrete micro-expression categories, but few address continuous intensity evolution over time. Progress is limited by lack of frame-level intensity labels, making fully supervised regression impractical.Method: Uses triangular prior to convert sparse temporal landmarks into dense pseudo-intensity trajectories. A lightweight temporal regression model combines ResNet18 encoder with bidirectional GRU to predict frame-wise intensity directly from image sequences. Requires no frame-level annotation and uses consistent preprocessing pipeline.
Result: On SAMM: Spearman correlation 0.9014, Kendall correlation 0.7999. On CASME II: up to 0.9116 and 0.8168 respectively when trained without apex-ranking term. Outperforms frame-wise baseline. Strong temporal agreement with pseudo-intensity trajectories.
Conclusion: First unified approach for continuous micro-expression intensity estimation using only sparse temporal annotations. Temporal modeling and structured pseudo labels are central to capturing rise-apex-fall dynamics of micro-facial movements.
Abstract: Micro-facial expressions are brief and involuntary facial movements that reflect genuine emotional states. While most prior work focuses on classifying discrete micro-expression categories, far fewer studies address the continuous evolution of intensity over time. Progress in this direction is limited by the lack of frame-level intensity labels, which makes fully supervised regression impractical. We propose a unified framework for continuous micro-expression intensity estimation using only weak temporal labels (onset, apex, offset). A simple triangular prior converts sparse temporal landmarks into dense pseudo-intensity trajectories, and a lightweight temporal regression model that combines a ResNet18 encoder with a bidirectional GRU predicts frame-wise intensity directly from image sequences. The method requires no frame-level annotation effort and is applied consistently across datasets through a single preprocessing and temporal alignment pipeline. Experiments on SAMM and CASME II show strong temporal agreement with the pseudo-intensity trajectories. On SAMM, the model reaches a Spearman correlation of 0.9014 and a Kendall correlation of 0.7999, outperforming a frame-wise baseline. On CASME II, it achieves up to 0.9116 and 0.8168, respectively, when trained without the apex-ranking term. Ablation studies confirm that temporal modeling and structured pseudo labels are central to capturing the rise-apex-fall dynamics of micro-facial movements. To our knowledge, this is the first unified approach for continuous micro-expression intensity estimation using only sparse temporal annotations.
[284] SocialFusion: Addressing Social Degradation in Pre-trained Vision-Language Models
Hamza Tahboub, Weiyan Shi, Gang Hua, Huaizu Jiang
Main category: cs.CV
TL;DR: SocialFusion: A unified framework that connects frozen visual encoder with language model to overcome “social degradation” in VLMs, enabling positive transfer across multiple social perception tasks.
Details
Motivation: Current vision-language models (VLMs) surprisingly struggle with multiple social perception tasks simultaneously, showing negative transfer due to "social degradation" - where general pre-training impairs visual encoder's ability to represent nuanced social information.Method: Propose SocialFusion framework that learns minimal connection between frozen visual encoder and language model. Investigate social degradation through linear representation probing (decodability) and gradient conflict analysis (compatibility).
Result: SocialFusion exhibits positive transfer across all five social tasks, leverages synergies between tasks to enhance overall performance, and achieves comparable performance to task-specific state-of-the-art models on various benchmarks.
Conclusion: Current VLM pre-training strategies may be detrimental to acquiring general social competence, highlighting need for more socially-aware training paradigms. SocialFusion demonstrates effective approach to unified social perception learning.
Abstract: Understanding social interactions from visual cues is a fundamental challenge for a socially competent AI. While powerful pre-trained vision-language models (VLMs) have shown remarkable general capabilities, they surprisingly struggle to unify and learn multiple social perception tasks simultaneously, often exhibiting negative transfer. We identify that this negative transfer stems from a critical issue we term “social degradation,” whereby the general visual-linguistic pre-training process of VLMs impairs the visual encoder’s ability to represent nuanced social information. We investigate this behavior further under two lenses: decodability through linear representation probing and compatibility through gradient conflict analysis, revealing that both play a role in the degradation, especially the former, which is significantly compromised in the VLM pre-training process. To address these issues, we propose SocialFusion, a unified framework that learns a minimal connection between a frozen visual encoder and a language model. Compared with existing VLMs, it exhibits positive transfer across all five social tasks, leveraging synergies between them to enhance overall performance and achieves comparable performance to task-specific state-of-the-art models on various benchmarks. Our findings suggest that current VLM pre-training strategies may be detrimental to acquiring general social competence and highlight the need for more socially-aware training paradigms.
[285] DPAC: Distribution-Preserving Adversarial Control for Diffusion Sampling
Han-Jin Lee, Han-Ju Lee, Jin-Seong Kim, Seok-Hwan Choi
Main category: cs.CV
TL;DR: DPAC: A diffusion guidance method that projects adversarial gradients onto tangent spaces to preserve sample quality while achieving target classification.
Details
Motivation: Adversarial guidance for diffusion sampling often achieves target classification but degrades sample quality due to accumulated deviations between controlled and nominal trajectories.Method: Formalize degradation as path-space KL divergence, derive optimal control via Girsanov’s theorem, and propose DPAC which projects adversarial gradients onto tangent spaces orthogonal to the score function.
Result: DPAC achieves lower FID and path-KL at matched attack success rates, with theoretical guarantees of O(Δt²) quality gap in discrete solvers.
Conclusion: Distribution-preserving adversarial control (DPAC) provides principled connection between adversarial control energy and perceptual fidelity, enabling high-quality targeted sampling.
Abstract: Adversarially guided diffusion sampling often achieves the target class, but sample quality degrades as deviations between the adversarially controlled and nominal trajectories accumulate. We formalize this degradation as a path-space Kullback-Leibler divergence(path-KL) between controlled and nominal (uncontrolled) diffusion processes, thereby showing via Girsanov’s theorem that it exactly equals the control energy. Building on this stochastic optimal control (SOC) view, we theoretically establish that minimizing this path-KL simultaneously tightens upper bounds on both the 2-Wasserstein distance and Fréchet Inception Distance (FID), revealing a principled connection between adversarial control energy and perceptual fidelity. From a variational perspective, we derive a first-order optimality condition for the control: among all directions that yield the same classification gain, the component tangent to iso-(log-)density surfaces (i.e., orthogonal to the score) minimizes path-KL, whereas the normal component directly increases distributional drift. This leads to DPAC (Distribution-Preserving Adversarial Control), a diffusion guidance rule that projects adversarial gradients onto the tangent space defined by the generative score geometry. We further show that in discrete solvers, the tangent projection cancels the O(Δt) leading error term in the Wasserstein distance, achieving an O(Δt^2) quality gap; moreover, it remains second-order robust to score or metric approximation. Empirical studies on ImageNet-100 validate the theoretical predictions, confirming that DPAC achieves lower FID and estimated path-KL at matched attack success rates.
[286] Real-Time On-the-Go Annotation Framework Using YOLO for Automated Dataset Generation
Mohamed Abdallah Salem, Ahmed Harb Rabia
Main category: cs.CV
TL;DR: Real-time annotation system using YOLO models on edge devices for agricultural applications, with comprehensive evaluation of YOLOv5, YOLOv8, and YOLOv12 architectures showing pretrained single-class configurations perform best.
Details
Motivation: Traditional dataset annotation for object detection models like YOLO is labor-intensive and time-consuming, especially in agriculture where rapid decision-making is critical. There's a need for more efficient annotation methods that can reduce manual labeling efforts post-data collection.Method: Proposed a real-time annotation approach using YOLO models deployed on edge devices for immediate labeling during image capture. Conducted extensive comparative analysis of three YOLO architectures (YOLOv5, YOLOv8, YOLOv12) with different configurations: single-class vs multi-class annotation and pretrained vs scratch-based training.
Result: Pretrained and single-class configurations showed significant advantages in model convergence, performance, and robustness. The real-time annotation framework was validated as feasible and effective, capable of drastically reducing dataset preparation time while maintaining high annotation quality.
Conclusion: The proposed real-time annotation system using YOLO models on edge devices is a practical solution for agricultural applications, with pretrained single-class configurations offering the best performance for efficient and accurate dataset annotation.
Abstract: Efficient and accurate annotation of datasets remains a significant challenge for deploying object detection models such as You Only Look Once (YOLO) in real-world applications, particularly in agriculture where rapid decision-making is critical. Traditional annotation techniques are labor-intensive, requiring extensive manual labeling post data collection. This paper presents a novel real-time annotation approach leveraging YOLO models deployed on edge devices, enabling immediate labeling during image capture. To comprehensively evaluate the efficiency and accuracy of our proposed system, we conducted an extensive comparative analysis using three prominent YOLO architectures (YOLOv5, YOLOv8, YOLOv12) under various configurations: single-class versus multi-class annotation and pretrained versus scratch-based training. Our analysis includes detailed statistical tests and learning dynamics, demonstrating significant advantages of pretrained and single-class configurations in terms of model convergence, performance, and robustness. Results strongly validate the feasibility and effectiveness of our real-time annotation framework, highlighting its capability to drastically reduce dataset preparation time while maintaining high annotation quality.
[287] VSRD++: Autolabeling for 3D Object Detection via Instance-Aware Volumetric Silhouette Rendering
Zihua Liu, Hiroki Sakuma, Masatoshi Okutomi
Main category: cs.CV
TL;DR: VSRD++ is a weakly supervised monocular 3D object detection framework that eliminates the need for 3D annotations by using neural-field-based volumetric rendering with only 2D supervision.
Details
Motivation: Existing monocular 3D object detection methods heavily depend on supervised learning with extensive 3D annotations, which are expensive and labor-intensive to acquire from LiDAR point clouds. There's a need for approaches that can work with weaker supervision.Method: Two-stage pipeline: 1) Multi-view 3D autolabeling using signed distance fields (SDFs) with instance-aware volumetric silhouette rendering, decomposing SDFs into cuboid SDFs and residual distance fields, modeling dynamic objects with velocity attributes and confidence scores, and using 3D attribute initialization; 2) Training monocular 3D detectors using the generated pseudo-labels.
Result: Extensive experiments on KITTI-360 dataset show VSRD++ significantly outperforms existing weakly supervised approaches for monocular 3D object detection on both static and dynamic scenes.
Conclusion: VSRD++ successfully demonstrates that weakly supervised monocular 3D object detection is feasible by leveraging neural-field-based volumetric rendering with only 2D supervision, eliminating the need for expensive 3D annotations while achieving strong performance.
Abstract: Monocular 3D object detection is a fundamental yet challenging task in 3D scene understanding. Existing approaches heavily depend on supervised learning with extensive 3D annotations, which are often acquired from LiDAR point clouds through labor-intensive labeling processes. To tackle this problem, we propose VSRD++, a novel weakly supervised framework for monocular 3D object detection that eliminates the reliance on 3D annotations and leverages neural-field-based volumetric rendering with weak 2D supervision. VSRD++ consists of a two-stage pipeline: multi-view 3D autolabeling and subsequent monocular 3D detector training. In the multi-view autolabeling stage, object surfaces are represented as signed distance fields (SDFs) and rendered as instance masks via the proposed instance-aware volumetric silhouette rendering. To optimize 3D bounding boxes, we decompose each instance’s SDF into a cuboid SDF and a residual distance field (RDF) that captures deviations from the cuboid. To address the geometry inconsistency commonly observed in volume rendering methods applied to dynamic objects, we model the dynamic objects by including velocity into bounding box attributes as well as assigning confidence to each pseudo-label. Moreover, we also employ a 3D attribute initialization module to initialize the dynamic bounding box parameters. In the monocular 3D object detection phase, the optimized 3D bounding boxes serve as pseudo labels for training monocular 3D object detectors. Extensive experiments on the KITTI-360 dataset demonstrate that VSRD++ significantly outperforms existing weakly supervised approaches for monocular 3D object detection on both static and dynamic scenes. Code is available at https://github.com/Magicboomliu/VSRD_plus_plus
[288] TabletopGen: Instance-Level Interactive 3D Tabletop Scene Generation from Text or Single Image
Ziqian Wang, Yonghao He, Licheng Yang, Wei Zou, Hongxuan Ma, Liu Liu, Wei Sui, Yuxin Guo, Hu Su
Main category: cs.CV
TL;DR: TabletopGen: A training-free framework for generating diverse, physically interactive 3D tabletop scenes from reference images, with novel pose/scale alignment for accurate 3D reconstruction.
Details
Motivation: Current text/image-driven 3D scene generation methods focus on large-scale scenes and struggle with high-density layouts and complex spatial relations of tabletop scenes, which are essential for embodied AI, robotic manipulation policy learning, and data synthesis.Method: TabletopGen accepts reference images (can be synthesized by text-to-image models), performs instance segmentation/completion to get per-instance images, reconstructs each instance into 3D models with canonical coordinate alignment, then uses a novel two-stage pose/scale alignment approach (Differentiable Rotation Optimizer + Top-view Spatial Alignment) before assembling into collision-free, simulation-ready scenes.
Result: Extensive experiments and user studies show TabletopGen achieves state-of-the-art performance, surpassing existing methods in visual fidelity, layout accuracy, and physical plausibility, capable of generating realistic tabletop scenes with rich stylistic and spatial diversity.
Conclusion: TabletopGen provides an effective training-free framework for generating high-fidelity, physically interactive 3D tabletop scenes, addressing limitations of current methods and enabling better support for embodied AI applications like robotic manipulation.
Abstract: Generating high-fidelity, physically interactive 3D simulated tabletop scenes is essential for embodied AI–especially for robotic manipulation policy learning and data synthesis. However, current text- or image-driven 3D scene generation methods mainly focus on large-scale scenes, struggling to capture the high-density layouts and complex spatial relations that characterize tabletop scenes. To address these challenges, we propose TabletopGen, a training-free, fully automatic framework that generates diverse, instance-level interactive 3D tabletop scenes. TabletopGen accepts a reference image as input, which can be synthesized by a text-to-image model to enhance scene diversity. We then perform instance segmentation and completion on the reference to obtain per-instance images. Each instance is reconstructed into a 3D model followed by canonical coordinate alignment. The aligned 3D models then undergo pose and scale estimation before being assembled into a collision-free, simulation-ready tabletop scene. A key component of our framework is a novel pose and scale alignment approach that decouples the complex spatial reasoning into two stages: a Differentiable Rotation Optimizer for precise rotation recovery and a Top-view Spatial Alignment mechanism for robust translation and scale estimation, enabling accurate 3D reconstruction from 2D reference. Extensive experiments and user studies show that TabletopGen achieves state-of-the-art performance, markedly surpassing existing methods in visual fidelity, layout accuracy, and physical plausibility, capable of generating realistic tabletop scenes with rich stylistic and spatial diversity. Our code will be publicly available.
[289] Closing the Approximation Gap of Partial AUC Optimization: A Tale of Two Formulations
Yangbangyan Jiang, Qianqian Xu, Huiyang Shao, Zhiyong Yang, Shilong Bao, Xiaochun Cao, Qingming Huang
Main category: cs.CV
TL;DR: The paper presents two instance-wise minimax reformulations for partial AUC (PAUC) optimization that close the approximation gap, offering either asymptotically vanishing gap or unbiasedness with more variables.
Details
Motivation: PAUC is crucial for real-world scenarios with class imbalance and decision constraints, but existing methods suffer from uncontrollable approximation errors or limited scalability when optimizing approximate PAUC objectives.Method: Two instance-wise minimax reformulations: (1) with asymptotically vanishing gap, (2) unbiased with more variables. Key ideas: establish equivalent instance-wise problem to lower time complexity, simplify sample selection via threshold learning, apply different smoothing techniques.
Result: Algorithms achieve linear per-iteration computational complexity w.r.t. sample size and convergence rate of O(ε^{-1/3}) for typical PAUCs. Provides tight generalization bound showing impact of TPR/FPR constraints α/β with order of Õ(α^{-1}n_+^{-1} + β^{-1}n_-^{-1}). Extensive experiments validate method strength.
Conclusion: The proposed methods close the approximation gap in PAUC optimization with efficient solvers, linear complexity, and proven convergence rates, while providing theoretical generalization bounds that explicitly quantify constraint impacts.
Abstract: As a variant of the Area Under the ROC Curve (AUC), the partial AUC (PAUC) focuses on a specific range of false positive rate (FPR) and/or true positive rate (TPR) in the ROC curve. It is a pivotal evaluation metric in real-world scenarios with both class imbalance and decision constraints. However, selecting instances within these constrained intervals during its calculation is NP-hard, and thus typically requires approximation techniques for practical resolution. Despite the progress made in PAUC optimization over the last few years, most existing methods still suffer from uncontrollable approximation errors or a limited scalability when optimizing the approximate PAUC objectives. In this paper, we close the approximation gap of PAUC optimization by presenting two simple instance-wise minimax reformulations: one with an asymptotically vanishing gap, the other with the unbiasedness at the cost of more variables. Our key idea is to first establish an equivalent instance-wise problem to lower the time complexity, simplify the complicated sample selection procedure by threshold learning, and then apply different smoothing techniques. Equipped with an efficient solver, the resulting algorithms enjoy a linear per-iteration computational complexity w.r.t. the sample size and a convergence rate of $O(ε^{-1/3})$ for typical one-way and two-way PAUCs. Moreover, we provide a tight generalization bound of our minimax reformulations. The result explicitly demonstrates the impact of the TPR/FPR constraints $α$/$β$ on the generalization and exhibits a sharp order of $\tilde{O}(α^{-1}\n_+^{-1} + β^{-1}\n_-^{-1})$. Finally, extensive experiments on several benchmark datasets validate the strength of our proposed methods.
[290] M4-BLIP: Advancing Multi-Modal Media Manipulation Detection through Face-Enhanced Local Analysis
Hang Wu, Ke Sun, Jiayi Ji, Xiaoshuai Sun, Rongrong Ji
Main category: cs.CV
TL;DR: M4-BLIP framework for multi-modal media manipulation detection that focuses on local facial features using BLIP-2 for feature extraction, integrates local facial information as prior knowledge, and enhances interpretability with LLM integration.
Details
Motivation: Current multi-modal media manipulation detection methods overlook localized information, despite manipulations frequently occurring in specific areas like facial regions, creating a significant societal threat to information reliability and integrity.Method: Proposes M4-BLIP framework using BLIP-2 model for local feature extraction, incorporates local facial information as prior knowledge, includes alignment and fusion module to integrate local and global features, and integrates with Large Language Models for interpretability.
Result: Extensive quantitative and visualization experiments validate the framework’s effectiveness against state-of-the-art competitors, showing enhanced detection accuracy.
Conclusion: The M4-BLIP framework successfully addresses the limitation of overlooking localized information in media manipulation detection by focusing on facial regions and improving interpretability through LLM integration.
Abstract: In the contemporary digital landscape, multi-modal media manipulation has emerged as a significant societal threat, impacting the reliability and integrity of information dissemination. Current detection methodologies in this domain often overlook the crucial aspect of localized information, despite the fact that manipulations frequently occur in specific areas, particularly in facial regions. In response to this critical observation, we propose the M4-BLIP framework. This innovative framework utilizes the BLIP-2 model, renowned for its ability to extract local features, as the cornerstone for feature extraction. Complementing this, we incorporate local facial information as prior knowledge. A specially designed alignment and fusion module within M4-BLIP meticulously integrates these local and global features, creating a harmonious blend that enhances detection accuracy. Furthermore, our approach seamlessly integrates with Large Language Models (LLM), significantly improving the interpretability of the detection outcomes. Extensive quantitative and visualization experiments validate the effectiveness of our framework against the state-of-the-art competitors.
[291] S$^2$-MLLM: Boosting Spatial Reasoning Capability of MLLMs for 3D Visual Grounding with Structural Guidance
Beining Xu, Siting Zhu, Zhao Jin, Junxian Li, Hesheng Wang
Main category: cs.CV
TL;DR: S²-MLLM: An efficient framework that enhances spatial reasoning in Multi-modal Large Language Models for 3D Visual Grounding through implicit spatial reasoning, avoiding inefficient point cloud reconstruction.
Details
Motivation: Existing MLLMs struggle with 3D spatial understanding from limited 2D perspectives, and current methods using viewpoint-dependent rendering of point clouds are inefficient and limit spatial reasoning capabilities.Method: Proposes S²-MLLM with spatial guidance strategy leveraging feed-forward 3D reconstruction’s structure awareness. Includes structure-enhanced module (SE) with intra-view/inter-view attention and multi-level position encoding to associate visual representations with spatial positions.
Result: Achieves significant performance improvements over existing methods across ScanRefer, Nr3D, and Sr3D datasets, unifying superior performance, generalization, and efficiency.
Conclusion: S²-MLLM effectively addresses inefficiency and limited spatial reasoning in existing 3DVG methods by enabling implicit spatial understanding without point cloud reconstruction, demonstrating strong performance across multiple benchmarks.
Abstract: 3D Visual Grounding (3DVG) focuses on locating objects in 3D scenes based on natural language descriptions, serving as a fundamental task for embodied AI and robotics. Recent advances in Multi-modal Large Language Models (MLLMs) have motivated research into extending them to 3DVG. However, MLLMs primarily process 2D visual inputs and struggle with understanding 3D spatial structure of scenes solely from these limited perspectives. Existing methods mainly utilize viewpoint-dependent rendering of reconstructed point clouds to provide explicit structural guidance for MLLMs in 3DVG tasks, leading to inefficiency and limited spatial reasoning. To address this issue, we propose S$^2$-MLLM, an efficient framework that enhances spatial reasoning in MLLMs through implicit spatial reasoning. We introduce a spatial guidance strategy that leverages the structure awareness of feed-forward 3D reconstruction. By acquiring 3D structural understanding during training, our model can implicitly reason about 3D scenes without relying on inefficient point cloud reconstruction. Moreover, we propose a structure-enhanced module (SE), which first employs intra-view and inter-view attention mechanisms to capture dependencies within views and correspondences across views. The module further integrates multi-level position encoding to associate visual representations with spatial positions and viewpoint information, enabling more accurate structural understanding. Extensive experiments demonstrate that S$^2$-MLLM unifies superior performance, generalization, and efficiency, achieving significant performance over existing methods across the ScanRefer, Nr3D, and Sr3D datasets. Code will be available upon acceptance.
[292] PSR: Scaling Multi-Subject Personalized Image Generation with Pairwise Subject-Consistency Rewards
Shulei Wang, Longhui Wei, Xin He, Jianbo Ouyang, Hui Lu, Zhou Zhao, Qi Tian
Main category: cs.CV
TL;DR: PSR introduces a scalable pipeline for multi-subject personalized image generation using single-subject models, with reinforcement learning rewards for subject consistency and text control.
Details
Motivation: Existing personalized generation models degrade when extended to multiple subjects, struggling with subject consistency and prompt adherence due to lack of high-quality multi-subject datasets and refined training strategies.Method: 1) Scalable multi-subject data generation pipeline using single-subject models; 2) Pairwise Subject-Consistency Rewards and general-purpose rewards; 3) Reinforcement learning stage for refinement; 4) New benchmark with 7 subsets across 3 dimensions.
Result: Extensive experiments demonstrate effectiveness in advancing multi-subject personalized image generation, with improved subject consistency and text controllability.
Conclusion: The proposed approach successfully addresses multi-subject personalization challenges through scalable data generation and refined reinforcement learning with specialized rewards.
Abstract: Personalized generation models for a single subject have demonstrated remarkable effectiveness, highlighting their significant potential. However, when extended to multiple subjects, existing models often exhibit degraded performance, particularly in maintaining subject consistency and adhering to textual prompts. We attribute these limitations to the absence of high-quality multi-subject datasets and refined post-training strategies. To address these challenges, we propose a scalable multi-subject data generation pipeline that leverages powerful single-subject generation models to construct diverse and high-quality multi-subject training data. Through this dataset, we first enable single-subject personalization models to acquire knowledge of synthesizing multi-image and multi-subject scenarios. Furthermore, to enhance both subject consistency and text controllability, we design a set of Pairwise Subject-Consistency Rewards and general-purpose rewards, which are incorporated into a refined reinforcement learning stage. To comprehensively evaluate multi-subject personalization, we introduce a new benchmark that assesses model performance using seven subsets across three dimensions. Extensive experiments demonstrate the effectiveness of our approach in advancing multi-subject personalized image generation. Github Link: https://github.com/wang-shulei/PSR
[293] TRivia: Self-supervised Fine-tuning of Vision-Language Models for Table Recognition
Junyuan Zhang, Bin Wang, Qintong Zhang, Fan Wu, Zichen Wen, Jialin Lu, Junjie Shan, Ziqi Zhao, Shuya Yang, Ziling Wang, Ziyang Miao, Huaping Zhong, Yuhang Zang, Xiaoyi Dong, Ka-Ho Chow, Conghui He
Main category: cs.CV
TL;DR: TRivia: A self-supervised fine-tuning method for table recognition that learns from unlabeled table images using question-answering rewards, achieving SOTA performance with TRivia-3B model.
Details
Motivation: Supervised table recognition requires costly labeled data, creating a gap between proprietary models (with large labeled datasets) and open-source models (with limited resources). Privacy regulations often restrict use of proprietary models, making open-source alternatives necessary but underperforming.Method: Built on Group Relative Policy Optimization, TRivia automatically identifies effective unlabeled samples and uses a question-answering-based reward mechanism without human annotations. An attention-guided module generates diverse questions for each table image, and the model’s ability to interpret recognition results and answer correctly provides feedback for optimization.
Result: TRivia-3B, an open-sourced 3B parameter model, surpasses existing systems (including Gemini 2.5 Pro and MinerU2.5) on three popular benchmarks, demonstrating state-of-the-art performance.
Conclusion: TRivia enables pretrained VLMs to autonomously learn table recognition from unlabeled data, bridging the performance gap between proprietary and open-source models while eliminating the need for costly labeled datasets.
Abstract: Table recognition (TR) aims to transform table images into semi-structured representations such as HTML or Markdown. As a core component of document parsing, TR has long relied on supervised learning, with recent efforts dominated by fine-tuning vision-language models (VLMs) using labeled data. While VLMs have brought TR to the next level, pushing performance further demands large-scale labeled data that is costly to obtain. Consequently, although proprietary models have continuously pushed the performance boundary, open-source models, often trained with limited resources and, in practice, the only viable option for many due to privacy regulations, still lag far behind. To bridge this gap, we introduce TRivia, a self-supervised fine-tuning method that enables pretrained VLMs to learn TR directly from unlabeled table images in the wild. Built upon Group Relative Policy Optimization, TRivia automatically identifies unlabeled samples that most effectively facilitate learning and eliminates the need for human annotations through a question-answering-based reward mechanism. An attention-guided module generates diverse questions for each table image, and the ability to interpret the recognition results and answer them correctly provides feedback to optimize the TR model. This closed-loop process allows the TR model to autonomously learn to recognize, structure, and reason over tables without labeled data. Leveraging this pipeline, we present TRivia-3B, an open-sourced, compact, and state-of-the-art TR model that surpasses existing systems (e.g., Gemini 2.5 Pro, MinerU2.5) on three popular benchmarks. Model and code are released at: https://github.com/opendatalab/TRivia
[294] ViscNet: Vision-Based In-line Viscometry for Fluid Mixing Process
Jongwon Sohn, Juhyeon Moon, Hyunjoon Jung, Jaewook Nam
Main category: cs.CV
TL;DR: Vision-based viscometer uses optical distortion of background patterns through fluid surface to measure viscosity non-invasively, achieving 0.113 MAE in log units and 81% classification accuracy.
Details
Motivation: Conventional viscometers are invasive and require controlled lab environments that don't match real process conditions, limiting process monitoring and autonomous lab operations.Method: Computer vision system that infers viscosity by analyzing how a fixed background pattern becomes optically distorted as light refracts through the mixing-driven, continuously deforming free surface of the fluid.
Result: Achieves mean absolute error of 0.113 in log m² s⁻¹ units for regression and up to 81% accuracy in viscosity-class prediction. Multi-pattern strategy improves robustness, and uncertainty quantification provides confidence estimates.
Conclusion: The stand-off viscometer offers a practical, automation-ready alternative to existing viscometry methods, enabling non-invasive viscosity measurement in real process conditions.
Abstract: Viscosity measurement is essential for process monitoring and autonomous laboratory operation, yet conventional viscometers remain invasive and require controlled laboratory environments that differ substantially from real process conditions. We present a computer-vision-based viscometer that infers viscosity by exploiting how a fixed background pattern becomes optically distorted as light refracts through the mixing-driven, continuously deforming free surface. Under diverse lighting conditions, the system achieves a mean absolute error of 0.113 in log m2 s^-1 units for regression and reaches up to 81% accuracy in viscosity-class prediction. Although performance declines for classes with closely clustered viscosity values, a multi-pattern strategy improves robustness by providing enriched visual cues. To ensure sensor reliability, we incorporate uncertainty quantification, enabling viscosity predictions with confidence estimates. This stand-off viscometer offers a practical, automation-ready alternative to existing viscometry methods.
[295] nnMobileNet++: Towards Efficient Hybrid Networks for Retinal Image Analysis
Xin Li, Wenhui Zhu, Xuanzhao Dong, Hao Wang, Yujian Xiong, Oana Dumitrascu, Yalin Wang
Main category: cs.CV
TL;DR: nnMobileNet++ is a hybrid CNN-transformer architecture for retinal image analysis that combines dynamic snake convolution, transformer blocks, and retinal pretraining to achieve state-of-the-art performance with low computational cost.
Details
Motivation: Traditional CNN architectures like nnMobileNet struggle to capture long-range dependencies and model irregular lesions and elongated vascular patterns in retinal images, which are critical for clinical diagnosis. There's a need for lightweight yet effective models that can handle these complex patterns.Method: Proposes nnMobileNet++, a hybrid architecture with three key components: 1) dynamic snake convolution for boundary-aware feature extraction, 2) stage-specific transformer blocks after the second down-sampling stage for global context modeling, and 3) retinal image pretraining to improve generalization.
Result: Achieves state-of-the-art or highly competitive accuracy on multiple public retinal datasets for classification tasks while maintaining low computational cost, as demonstrated through experiments and ablation studies.
Conclusion: nnMobileNet++ represents an effective lightweight framework for retinal image analysis that successfully bridges convolutional and transformer representations, addressing the limitations of purely convolutional architectures while maintaining computational efficiency.
Abstract: Retinal imaging is a critical, non-invasive modality for the early detection and monitoring of ocular and systemic diseases. Deep learning, particularly convolutional neural networks (CNNs), has significant progress in automated retinal analysis, supporting tasks such as fundus image classification, lesion detection, and vessel segmentation. As a representative lightweight network, nnMobileNet has demonstrated strong performance across multiple retinal benchmarks while remaining computationally efficient. However, purely convolutional architectures inherently struggle to capture long-range dependencies and model the irregular lesions and elongated vascular patterns that characterize on retinal images, despite the critical importance of vascular features for reliable clinical diagnosis. To further advance this line of work and extend the original vision of nnMobileNet, we propose nnMobileNet++, a hybrid architecture that progressively bridges convolutional and transformer representations. The framework integrates three key components: (i) dynamic snake convolution for boundary-aware feature extraction, (ii) stage-specific transformer blocks introduced after the second down-sampling stage for global context modeling, and (iii) retinal image pretraining to improve generalization. Experiments on multiple public retinal datasets for classification, together with ablation studies, demonstrate that nnMobileNet++ achieves state-of-the-art or highly competitive accuracy while maintaining low computational cost, underscoring its potential as a lightweight yet effective framework for retinal image analysis.
[296] Supervised Contrastive Machine Unlearning of Background Bias in Sonar Image Classification with Fine-Grained Explainable AI
Kamal Basha S, Athira Nambiar
Main category: cs.CV
TL;DR: Proposes a novel framework combining Targeted Contrastive Unlearning (TCU) and Unlearn to Explain Sonar Framework (UESF) to reduce seafloor bias in sonar image analysis and improve model generalization and interpretability.
Details
Motivation: Existing AI models for sonar image analysis over-rely on seafloor features, leading to poor generalization despite high accuracy on training data. This bias problem needs to be addressed for better real-world performance.Method: Two key modules: 1) Targeted Contrastive Unlearning (TCU) extends triplet loss to reduce seafloor-induced background bias, 2) Unlearn to Explain Sonar Framework (UESF) provides visual insights into what the model has forgotten and adapts LIME explainer for better attribution evaluation.
Result: Extensive experiments on real and synthetic sonar datasets show significant improvements in unlearning effectiveness, model robustness, and interpretability compared to existing approaches.
Conclusion: The proposed framework successfully addresses seafloor bias in sonar image analysis, improving generalization while maintaining interpretability through novel unlearning and explanation techniques.
Abstract: Acoustic sonar image analysis plays a critical role in object detection and classification, with applications in both civilian and defense domains. Despite the availability of real and synthetic datasets, existing AI models that achieve high accuracy often over-rely on seafloor features, leading to poor generalization. To mitigate this issue, we propose a novel framework that integrates two key modules: (i) a Targeted Contrastive Unlearning (TCU) module, which extends the traditional triplet loss to reduce seafloor-induced background bias and improve generalization, and (ii) the Unlearn to Explain Sonar Framework (UESF), which provides visual insights into what the model has deliberately forgotten while adapting the LIME explainer to generate more faithful and localized attributions for unlearning evaluation. Extensive experiments across both real and synthetic sonar datasets validate our approach, demonstrating significant improvements in unlearning effectiveness, model robustness, and interpretability.
[297] Diffusion Model in Latent Space for Medical Image Segmentation Task
Huynh Trinh Ngoc, Toan Nguyen Hai, Ba Luong Son, Long Tran Quoc
Main category: cs.CV
TL;DR: MedSegLatDiff is a diffusion-based framework combining VAE with latent diffusion for efficient medical image segmentation that generates multiple plausible masks and confidence maps while achieving state-of-the-art performance.
Details
Motivation: Traditional medical image segmentation methods produce only single masks, failing to capture inherent uncertainty in medical images. Recent generative models can create multiple plausible masks but are computationally heavy, limiting clinical deployment.Method: Combines variational autoencoder (VAE) with latent diffusion model. VAE compresses input into low-dimensional latent space to reduce noise and accelerate training. Diffusion process operates in this compact representation. Uses weighted cross-entropy loss in VAE mask reconstruction to better preserve tiny structures like small nodules.
Result: Achieves state-of-the-art or highly competitive Dice and IoU scores on ISIC-2018 (skin lesions), CVC-Clinic (polyps), and LIDC-IDRI (lung nodules). Simultaneously generates diverse segmentation hypotheses and confidence maps for enhanced interpretability.
Conclusion: MedSegLatDiff provides efficient uncertainty-aware medical image segmentation with enhanced interpretability and reliability compared to deterministic baselines, making it particularly suitable for clinical deployment.
Abstract: Medical image segmentation is crucial for clinical diagnosis and treatment planning. Traditional methods typically produce a single segmentation mask, failing to capture inherent uncertainty. Recent generative models enable the creation of multiple plausible masks per image, mimicking the collaborative interpretation of several clinicians. However, these approaches remain computationally heavy. We propose MedSegLatDiff, a diffusion based framework that combines a variational autoencoder (VAE) with a latent diffusion model for efficient medical image segmentation. The VAE compresses the input into a low dimensional latent space, reducing noise and accelerating training, while the diffusion process operates directly in this compact representation. We further replace the conventional MSE loss with weighted cross entropy in the VAE mask reconstruction path to better preserve tiny structures such as small nodules. MedSegLatDiff is evaluated on ISIC-2018 (skin lesions), CVC-Clinic (polyps), and LIDC-IDRI (lung nodules). It achieves state of the art or highly competitive Dice and IoU scores while simultaneously generating diverse segmentation hypotheses and confidence maps. This provides enhanced interpretability and reliability compared to deterministic baselines, making the model particularly suitable for clinical deployment.
[298] EGG-Fusion: Efficient 3D Reconstruction with Geometry-aware Gaussian Surfel on the Fly
Xiaokun Pan, Zhenzhe Li, Zhichao Ye, Hongjia Zhai, Guofeng Zhang
Main category: cs.CV
TL;DR: EGG-Fusion: A real-time differentiable rendering SLAM system using geometry-aware Gaussian surfels with information filter fusion for robust, high-precision 3D reconstruction at 24 FPS.
Details
Motivation: Current differentiable rendering methods for 3D reconstruction face dual challenges: real-time computation limitations and sensitivity to sensor noise, leading to degraded geometric fidelity and limited practicality for real-world applications.Method: Proposes EGG-Fusion with robust sparse-to-dense camera tracking and geometry-aware Gaussian surfel mapping module. Uses information filter-based fusion method that explicitly accounts for sensor noise. Differentiable Gaussian surfel mapping models multi-view consistent surfaces while enabling efficient parameter optimization.
Result: Achieves surface reconstruction error of 0.6cm on standardized benchmarks (Replica and ScanNet++), representing over 20% improvement in accuracy compared to state-of-the-art GS-based methods. Maintains real-time processing at 24 FPS.
Conclusion: EGG-Fusion establishes itself as one of the most accurate differentiable-rendering-based real-time reconstruction systems, successfully addressing both real-time computation and sensor noise challenges while achieving high-precision surface reconstruction.
Abstract: Real-time 3D reconstruction is a fundamental task in computer graphics. Recently, differentiable-rendering-based SLAM system has demonstrated significant potential, enabling photorealistic scene rendering through learnable scene representations such as Neural Radiance Fields (NeRF) and 3D Gaussian Splatting (3DGS). Current differentiable rendering methods face dual challenges in real-time computation and sensor noise sensitivity, leading to degraded geometric fidelity in scene reconstruction and limited practicality. To address these challenges, we propose a novel real-time system EGG-Fusion, featuring robust sparse-to-dense camera tracking and a geometry-aware Gaussian surfel mapping module, introducing an information filter-based fusion method that explicitly accounts for sensor noise to achieve high-precision surface reconstruction. The proposed differentiable Gaussian surfel mapping effectively models multi-view consistent surfaces while enabling efficient parameter optimization. Extensive experimental results demonstrate that the proposed system achieves a surface reconstruction error of 0.6\textit{cm} on standardized benchmark datasets including Replica and ScanNet++, representing over 20% improvement in accuracy compared to state-of-the-art (SOTA) GS-based methods. Notably, the system maintains real-time processing capabilities at 24 FPS, establishing it as one of the most accurate differentiable-rendering-based real-time reconstruction systems. Project Page: https://zju3dv.github.io/eggfusion/
[299] FOD-S2R: A FOD Dataset for Sim2Real Transfer Learning based Object Detection
Ashish Vashist, Qiranul Saadiyean, Suresh Sundaram, Chandra Sekhar Seelamantula
Main category: cs.CV
TL;DR: FOD-S2R: First dataset combining real and synthetic images of Foreign Object Debris in aircraft fuel tanks to study synthetic data’s effectiveness for detection in confined environments.
Details
Motivation: Foreign Object Debris in aircraft fuel tanks poses critical safety risks, but there's a lack of dedicated datasets for these complex, enclosed environments. Existing datasets focus on external or open-air settings, not confined fuel tank interiors.Method: Created FOD-S2R dataset with 3,114 real HD images captured in a controlled fuel tank replica and 3,137 synthetic images generated using Unreal Engine. Dataset includes varied FOVs, object distances, lighting conditions, colors, and object sizes. Benchmarked state-of-the-art object detection models.
Result: Synthetic data improves detection accuracy and generalization to real-world conditions. Experiments demonstrate synthetic data enhances model performance and narrows the Sim2Real gap.
Conclusion: FOD-S2R provides the first systematic evaluation of synthetic data for FOD detection in confined fuel tanks, offering a valuable foundation for developing automated aviation maintenance systems.
Abstract: Foreign Object Debris (FOD) within aircraft fuel tanks presents critical safety hazards including fuel contamination, system malfunctions, and increased maintenance costs. Despite the severity of these risks, there is a notable lack of dedicated datasets for the complex, enclosed environments found inside fuel tanks. To bridge this gap, we present a novel dataset, FOD-S2R, composed of real and synthetic images of the FOD within a simulated aircraft fuel tank. Unlike existing datasets that focus on external or open-air environments, our dataset is the first to systematically evaluate the effectiveness of synthetic data in enhancing the real-world FOD detection performance in confined, closed structures. The real-world subset consists of 3,114 high-resolution HD images captured in a controlled fuel tank replica, while the synthetic subset includes 3,137 images generated using Unreal Engine. The dataset is composed of various Field of views (FOV), object distances, lighting conditions, color, and object size. Prior research has demonstrated that synthetic data can reduce reliance on extensive real-world annotations and improve the generalizability of vision models. Thus, we benchmark several state-of-the-art object detection models and demonstrate that introducing synthetic data improves the detection accuracy and generalization to real-world conditions. These experiments demonstrate the effectiveness of synthetic data in enhancing the model performance and narrowing the Sim2Real gap, providing a valuable foundation for developing automated FOD detection systems for aviation maintenance.
[300] TBT-Former: Learning Temporal Boundary Distributions for Action Localization
Thisara Rathnayaka, Uthayasanker Thayasivam
Main category: cs.CV
TL;DR: TBT-Former improves temporal action localization by enhancing ActionFormer with a scaled Transformer backbone, cross-scale FPN, and boundary distribution regression head to better handle fuzzy boundaries and multi-scale context.
Details
Motivation: Existing single-stage anchor-free models like ActionFormer struggle with precise localization of actions with ambiguous temporal boundaries and effective fusion of multi-scale contextual information.Method: TBT-Former introduces three key components: 1) scaled Transformer backbone with more attention heads and expanded MLP dimensions, 2) cross-scale FPN with top-down pathway and lateral connections, 3) novel boundary distribution regression head inspired by Generalized Focal Loss that treats boundary regression as probability distribution learning.
Result: Achieves state-of-the-art performance on THUMOS14 and EPIC-Kitchens 100 datasets, while remaining competitive on ActivityNet-1.3.
Conclusion: TBT-Former advances Transformer-based architectures for temporal action localization by better addressing boundary uncertainty and multi-scale feature fusion, establishing new benchmarks on competitive datasets.
Abstract: Temporal Action Localization (TAL) remains a fundamental challenge in video understanding, aiming to identify the start time, end time, and category of all action instances within untrimmed videos. While recent single-stage, anchor-free models like ActionFormer have set a high standard by leveraging Transformers for temporal reasoning, they often struggle with two persistent issues: the precise localization of actions with ambiguous or “fuzzy” temporal boundaries and the effective fusion of multi-scale contextual information. In this paper, we introduce the Temporal Boundary Transformer (TBT-Former), a new architecture that directly addresses these limitations. TBT-Former enhances the strong ActionFormer baseline with three core contributions: (1) a higher-capacity scaled Transformer backbone with an increased number of attention heads and an expanded Multi-Layer Perceptron (MLP) dimension for more powerful temporal feature extraction; (2) a cross-scale feature pyramid network (FPN) that integrates a top-down pathway with lateral connections, enabling richer fusion of high-level semantics and low-level temporal details; and (3) a novel boundary distribution regression head. Inspired by the principles of Generalized Focal Loss (GFL), this new head recasts the challenging task of boundary regression as a more flexible probability distribution learning problem, allowing the model to explicitly represent and reason about boundary uncertainty. Within the paradigm of Transformer-based architectures, TBT-Former advances the formidable benchmark set by its predecessors, establishing a new level of performance on the highly competitive THUMOS14 and EPIC-Kitchens 100 datasets, while remaining competitive on the large-scale ActivityNet-1.3. Our code is available at https://github.com/aaivu/In21-S7-CS4681-AML-Research-Projects/tree/main/projects/210536K-Multi-Modal-Learning_Video-Understanding
[301] DCText: Scheduled Attention Masking for Visual Text Generation via Divide-and-Conquer Strategy
Jaewoo Song, Jooyoung Choi, Kanghyun Baek, Sangyub Lee, Daemin Park, Sungroh Yoon
Main category: cs.CV
TL;DR: DCText is a training-free method that improves text rendering in text-to-image models using divide-and-conquer strategy with attention masks and localized noise initialization.
Details
Motivation: Current text-to-image models struggle with rendering long or multiple texts due to diluted global attention, which reduces text accuracy in generated images.Method: Uses divide-and-conquer strategy: 1) decomposes prompts by extracting and dividing target text, 2) assigns each segment to designated regions, 3) applies Text-Focus and Context-Expansion attention masks sequentially during denoising, 4) uses Localized Noise Initialization for better accuracy and alignment.
Result: Achieves best text accuracy on single- and multi-sentence benchmarks without compromising image quality, while also delivering the lowest generation latency.
Conclusion: DCText effectively solves the long/multiple text rendering problem in text-to-image generation through a training-free, computationally efficient approach that maintains both text accuracy and image quality.
Abstract: Despite recent text-to-image models achieving highfidelity text rendering, they still struggle with long or multiple texts due to diluted global attention. We propose DCText, a training-free visual text generation method that adopts a divide-and-conquer strategy, leveraging the reliable short-text generation of Multi-Modal Diffusion Transformers. Our method first decomposes a prompt by extracting and dividing the target text, then assigns each to a designated region. To accurately render each segment within their regions while preserving overall image coherence, we introduce two attention masks - Text-Focus and Context-Expansion - applied sequentially during denoising. Additionally, Localized Noise Initialization further improves text accuracy and region alignment without increasing computational cost. Extensive experiments on single- and multisentence benchmarks show that DCText achieves the best text accuracy without compromising image quality while also delivering the lowest generation latency.
[302] Rethinking Intracranial Aneurysm Vessel Segmentation: A Perspective from Computational Fluid Dynamics Applications
Feiyang Xiao, Yichi Zhang, Xigui Li, Yuanye Zhou, Chen Jiang, Xin Guo, Limei Han, Yuxin Li, Fengping Zhu, Yuan Cheng
Main category: cs.CV
TL;DR: First comprehensive multi-center dataset (IAVS) for intracranial aneurysm vessel segmentation with 641 3D MRA images and 587 annotations, including hemodynamic analysis outcomes and CFD applicability evaluation system.
Details
Motivation: Current segmentation methods focus on image-based metrics but neglect practical effectiveness in subsequent computational fluid dynamics (CFD) applications for hemodynamic analysis of intracranial aneurysms.Method: Created IAVS dataset with image-mask pairs and hemodynamic outcomes, established two-stage evaluation benchmarks (aneurysm localization + fine-grained segmentation), developed two-stage framework as baseline, and created standardized CFD applicability evaluation system for automated conversion to CFD models.
Result: First comprehensive multi-center dataset addressing limitations of existing datasets, providing tools for clinically relevant technique development with focus on topological integrity and CFD applicability.
Conclusion: IAVS dataset bridges gap between image segmentation and practical CFD applications, enabling development of segmentation methods that are clinically relevant for hemodynamic analysis of intracranial aneurysms.
Abstract: The precise segmentation of intracranial aneurysms and their parent vessels (IA-Vessel) is a critical step for hemodynamic analyses, which mainly depends on computational fluid dynamics (CFD). However, current segmentation methods predominantly focus on image-based evaluation metrics, often neglecting their practical effectiveness in subsequent CFD applications. To address this deficiency, we present the Intracranial Aneurysm Vessel Segmentation (IAVS) dataset, the first comprehensive, multi-center collection comprising 641 3D MRA images with 587 annotations of aneurysms and IA-Vessels. In addition to image-mask pairs, IAVS dataset includes detailed hemodynamic analysis outcomes, addressing the limitations of existing datasets that neglect topological integrity and CFD applicability. To facilitate the development and evaluation of clinically relevant techniques, we construct two evaluation benchmarks including global localization of aneurysms (Stage I) and fine-grained segmentation of IA-Vessel (Stage II) and develop a simple and effective two-stage framework, which can be used as a out-of-the-box method and strong baseline. For comprehensive evaluation of applicability of segmentation results, we establish a standardized CFD applicability evaluation system that enables the automated and consistent conversion of segmentation masks into CFD models, offering an applicability-focused assessment of segmentation outcomes. The dataset, code, and model will be public available at https://github.com/AbsoluteResonance/IAVS.
[303] Gaussian Swaying: Surface-Based Framework for Aerodynamic Simulation with 3D Gaussians
Hongru Yan, Xiang Zhang, Zeyuan Chen, Fangyin Wei, Zhuowen Tu
Main category: cs.CV
TL;DR: Gaussian Swaying: A surface-based aerodynamic simulation framework using 3D Gaussians that unifies simulation and rendering for realistic wind-driven motion.
Details
Motivation: Current methods for simulating aerodynamic effects (like swaying branches, rippling flags, rocking boats) have limitations: mesh-based approaches require costly meshing, while particle-based methods rely on discrete positional data. There's a need for an efficient, fine-grained approach to achieve realistic wind-driven motion in vision and graphics applications.Method: Proposes Gaussian Swaying, a surface-based framework using 3D Gaussians to model surfaces continuously. Uses Gaussian patches as a unified representation that supports both force computation for dynamics and provides normals for lightweight shading, eliminating the need for separate simulation and rendering representations.
Result: Comprehensive experiments on synthetic and real-world datasets across multiple metrics demonstrate state-of-the-art performance and efficiency. The framework offers scalable realistic aerodynamic scene simulation with fine-grained interaction.
Conclusion: Gaussian Swaying provides an efficient, unified approach for aerodynamic simulation that bridges the gap between simulation and rendering, enabling realistic wind-driven motion in computer vision and graphics applications with superior performance and scalability.
Abstract: Branches swaying in the breeze, flags rippling in the wind, and boats rocking on the water all show how aerodynamics shape natural motion – an effect crucial for realism in vision and graphics. In this paper, we present Gaussian Swaying, a surface-based framework for aerodynamic simulation using 3D Gaussians. Unlike mesh-based methods that require costly meshing, or particle-based approaches that rely on discrete positional data, Gaussian Swaying models surfaces continuously with 3D Gaussians, enabling efficient and fine-grained aerodynamic interaction. Our framework unifies simulation and rendering on the same representation: Gaussian patches, which support force computation for dynamics while simultaneously providing normals for lightweight shading. Comprehensive experiments on both synthetic and real-world datasets across multiple metrics demonstrate that Gaussian Swaying achieves state-of-the-art performance and efficiency, offering a scalable approach for realistic aerodynamic scene simulation.
[304] Lost in Distortion: Uncovering the Domain Gap Between Computer Vision and Brain Imaging - A Study on Pretraining for Age Prediction
Yanteng Zhang, Songheng Li, Zeyu Shen, Qizhen Lan, Lipei Zhang, Yang Liu, Vince Calhoun
Main category: cs.CV
TL;DR: This paper investigates whether noisy or low-quality neuroimaging scans can contribute meaningfully to pretraining foundation models, or if they hinder learning, by systematically exploring data quality levels in pretraining and their impact on downstream brain age prediction tasks.
Details
Motivation: Large-scale brain imaging datasets have high heterogeneity in quality, ranging from well-structured scans to severely distorted/incomplete volumes. Unlike natural image datasets in computer vision, this raises a fundamental question about whether noise/low-quality scans can meaningfully contribute to pretraining or hinder model learning.Method: The study systematically explores the role of data quality level in pretraining by performing pretraining on datasets with different quality levels, followed by fine-tuning for brain age prediction on external cohorts.
Result: Results show significant performance differences across quality levels, revealing both opportunities and limitations in using heterogeneous quality data for pretraining.
Conclusion: There is a gap between computer vision practices and clinical neuroimaging standards, emphasizing the necessity of domain-aware curation to ensure trusted and generalizable domain-specific foundation models.
Abstract: Large-scale brain imaging datasets provide unprecedented opportunities for developing domain foundation models through pretraining. However, unlike natural image datasets in computer vision, these neuroimaging data often exhibit high heterogeneity in quality, ranging from well-structured scans to severely distorted or incomplete brain volumes. This raises a fundamental question: can noise or low-quality scans contribute meaningfully to pretraining, or do they instead hinder model learning? In this study, we systematically explore the role of data quality level in pretraining and its impact on downstream tasks. Specifically, we perform pretraining on datasets with different quality levels and perform fine-tuning for brain age prediction on external cohorts. Our results show significant performance differences across quality levels, revealing both opportunities and limitations. We further discuss the gap between computer vision practices and clinical neuroimaging standards, emphasizing the necessity of domain-aware curation to ensure trusted and generalizable domain-specific foundation models.
[305] IVCR-200K: A Large-Scale Multi-turn Dialogue Benchmark for Interactive Video Corpus Retrieval
Ning Han, Yawen Zeng, Shaohua Long, Chengqing Li, Sijie Yang, Dun Tan, Jianfeng Dong, Jingjing Chen
Main category: cs.CV
TL;DR: Proposes Interactive Video Corpus Retrieval (IVCR) task and IVCR-200K dataset for multi-turn conversational video search, addressing limitations of one-way retrieval systems.
Details
Motivation: Current video retrieval systems lack meaningful interaction with users, failing to meet personalization needs of over 80% of users who require dynamic, conversational search experiences.Method: Introduces IVCR-200K dataset (bilingual, multi-turn, conversational) and proposes MLLM-based framework supporting multiple interaction modes with explainable solutions.
Result: Extensive experiments demonstrate effectiveness of both the dataset and the proposed framework for interactive video retrieval.
Conclusion: IVCR task represents a more realistic video search paradigm that enables conversational interactions, with the dataset and framework providing strong foundations for future research.
Abstract: In recent years, significant developments have been made in both video retrieval and video moment retrieval tasks, which respectively retrieve complete videos or moments for a given text query. These advancements have greatly improved user satisfaction during the search process. However, previous work has failed to establish meaningful “interaction” between the retrieval system and the user, and its one-way retrieval paradigm can no longer fully meet the personalization and dynamic needs of at least 80.8% of users. In this paper, we introduce the Interactive Video Corpus Retrieval (IVCR) task, a more realistic setting that enables multi-turn, conversational, and realistic interactions between the user and the retrieval system. To facilitate research on this challenging task, we introduce IVCR-200K, a high-quality, bilingual, multi-turn, conversational, and abstract semantic dataset that supports video retrieval and even moment retrieval. Furthermore, we propose a comprehensive framework based on multi-modal large language models (MLLMs) to help users interact in several modes with more explainable solutions. The extensive experiments demonstrate the effectiveness of our dataset and framework.
[306] Rice-VL: Evaluating Vision-Language Models for Cultural Understanding Across ASEAN Countries
Tushar Pranav, Eshan Pandey, Austria Lyka Diane Bala, Aman Chadha, Indriyati Atmosukarto, Donny Soh Cheng Lock
Main category: cs.CV
TL;DR: RICE-VL is a benchmark for evaluating Vision-Language Models’ cultural understanding across 11 ASEAN countries, revealing Western-centric biases and performance gaps in culturally diverse regions.
Details
Motivation: VLMs exhibit Western-centric biases that limit their effectiveness in culturally diverse regions like Southeast Asia, creating a need for culturally-aware evaluation benchmarks.Method: Created RICE-VL benchmark with 28,000+ human-curated VQA samples across 3 formats and 1,000 image-bounding box pairs for Visual Grounding, annotated by culturally informed experts. Proposed SEA-LAVE metric extension for assessing textual accuracy, cultural alignment, and country identification.
Result: Evaluation of 6 open- and closed-source VLMs revealed significant performance gaps in low-resource countries and abstract cultural domains, exposing limitations in VLMs’ cultural comprehension.
Conclusion: RICE-VL highlights the need for inclusive model development to better serve diverse global populations and addresses cultural biases in VLMs.
Abstract: Vision-Language Models (VLMs) excel in multimodal tasks but often exhibit Western-centric biases, limiting their effectiveness in culturally diverse regions like Southeast Asia (SEA). To address this, we introduce RICE-VL, a novel benchmark evaluating VLM cultural understanding across 11 ASEAN countries. RICE-VL includes over 28,000 human-curated Visual Question Answering (VQA) samples – covering True or False, Fill-in-the-Blank, and open-ended formats – and 1,000 image-bounding box pairs for Visual Grounding, annotated by culturally informed experts across 14 sub-ground categories. We propose SEA-LAVE, an extension of the LAVE metric, assessing textual accuracy, cultural alignment, and country identification. Evaluations of six open- and closed-source VLMs reveal significant performance gaps in low-resource countries and abstract cultural domains. The Visual Grounding task tests models’ ability to localize culturally significant elements in complex scenes, probing spatial and contextual accuracy. RICE-VL exposes limitations in VLMs’ cultural comprehension and highlights the need for inclusive model development to better serve diverse global populations.
[307] TokenPure: Watermark Removal through Tokenized Appearance and Structural Guidance
Pei Yang, Yepeng Liu, Kelly Peng, Yuan Gao, Yiren Song
Main category: cs.CV
TL;DR: TokenPure is a Diffusion Transformer framework for effective watermark removal that maintains content consistency by using token-based conditional reconstruction.
Details
Motivation: Digital watermarking is crucial for ownership proof of digital content, but robust watermark removal is needed while preserving content quality and consistency.Method: TokenPure decomposes watermarked images into visual tokens (texture) and structural tokens (geometry), then uses these tokens to condition a diffusion process for synthesizing watermark-free images while bypassing initial watermark-carrying noise.
Result: TokenPure achieves state-of-the-art watermark removal and reconstruction fidelity, substantially outperforming existing baselines in both perceptual quality and consistency.
Conclusion: The framework successfully solves the trade-off between thorough watermark destruction and content consistency through token-based conditional reconstruction, offering an effective solution for watermark removal in digital content.
Abstract: In the digital economy era, digital watermarking serves as a critical basis for ownership proof of massive replicable content, including AI-generated and other virtual assets. Designing robust watermarks capable of withstanding various attacks and processing operations is even more paramount. We introduce TokenPure, a novel Diffusion Transformer-based framework designed for effective and consistent watermark removal. TokenPure solves the trade-off between thorough watermark destruction and content consistency by leveraging token-based conditional reconstruction. It reframes the task as conditional generation, entirely bypassing the initial watermark-carrying noise. We achieve this by decomposing the watermarked image into two complementary token sets: visual tokens for texture and structural tokens for geometry. These tokens jointly condition the diffusion process, enabling the framework to synthesize watermark-free images with fine-grained consistency and structural integrity. Comprehensive experiments show that TokenPure achieves state-of-the-art watermark removal and reconstruction fidelity, substantially outperforming existing baselines in both perceptual quality and consistency.
[308] Optimizing Stroke Risk Prediction: A Machine Learning Pipeline Combining ROS-Balanced Ensembles and XAI
A S M Ahsanul Sarkar Akib, Raduana Khawla, Abdul Hasib
Main category: cs.CV
TL;DR: The paper presents an interpretable ML framework using ensemble modeling and XAI for stroke risk prediction, achieving 99.09% accuracy with Random Forest + ExtraTrees + XGBoost ensemble and identifying age, hypertension, and glucose levels as key predictors.
Details
Motivation: Stroke is a major global health concern causing death and permanent impairment. Early risk assessment is crucial for prompt intervention and effective prevention strategies, but current approaches need improvement in both accuracy and interpretability for clinical application.Method: Developed an interpretable ML framework using ensemble modeling and XAI techniques. Included comprehensive data preprocessing with feature engineering, addressed class imbalance using Random Over-Sampling (ROS), evaluated 10 ML models with 5-fold cross-validation, and created an optimized ensemble model combining Random Forest, ExtraTrees, and XGBoost. Used LIME-based interpretability analysis to identify key clinical variables.
Result: The optimized ensemble model achieved exceptional 99.09% accuracy on the Stroke Prediction Dataset (SPD). LIME-based analysis identified three critical clinical variables: age, hypertension, and glucose levels. The framework demonstrated high accuracy while maintaining interpretability for clinical use.
Conclusion: The study demonstrates that combining ensemble learning with explainable AI can provide highly accurate and interpretable stroke risk assessment. The framework enables early prediction, data-driven prevention, and personalized clinical decisions, potentially transforming stroke prediction and cardiovascular risk management.
Abstract: Stroke is a major cause of death and permanent impairment, making it a major worldwide health concern. For prompt intervention and successful preventative tactics, early risk assessment is essential. To address this challenge, we used ensemble modeling and explainable AI (XAI) techniques to create an interpretable machine learning framework for stroke risk prediction. A thorough evaluation of 10 different machine learning models using 5-fold cross-validation across several datasets was part of our all-inclusive strategy, which also included feature engineering and data pretreatment (using Random Over-Sampling (ROS) to solve class imbalance). Our optimized ensemble model (Random Forest + ExtraTrees + XGBoost) performed exceptionally well, obtaining a strong 99.09% accuracy on the Stroke Prediction Dataset (SPD). We improved the model’s transparency and clinical applicability by identifying three important clinical variables using LIME-based interpretability analysis: age, hypertension, and glucose levels. Through early prediction, this study highlights how combining ensemble learning with explainable AI (XAI) can deliver highly accurate and interpretable stroke risk assessment. By enabling data-driven prevention and personalized clinical decisions, our framework has the potential to transform stroke prediction and cardiovascular risk management.
[309] AlignVid: Training-Free Attention Scaling for Semantic Fidelity in Text-Guided Image-to-Video Generation
Yexin Liu, Wen-Jie Shu, Zile Huang, Haoze Zheng, Yueze Wang, Manyuan Zhang, Ser-Nam Lim, Harry Yang
Main category: cs.CV
TL;DR: AlignVid is a training-free framework that improves semantic adherence in text-guided image-to-video generation by using attention scaling modulation and guidance scheduling to address semantic negligence issues when prompts require substantial image transformations.
Details
Motivation: Existing TI2V methods struggle with semantic negligence - failing to properly adhere to fine-grained prompt semantics when prompts require substantial transformations of input images (object addition, deletion, or modification). A pilot study showed that Gaussian blur improves semantic adherence by creating clearer foreground-background separation in attention maps.Method: AlignVid has two components: 1) Attention Scaling Modulation (ASM) that directly reweights attention via lightweight Q or K scaling, and 2) Guidance Scheduling (GS) that applies ASM selectively across transformer blocks and denoising steps to reduce visual quality degradation. The framework is training-free and uses minimal intervention.
Result: The method improves semantic fidelity in TI2V generation. The authors also introduce OmitI2V, a new evaluation benchmark with 367 human-annotated samples spanning addition, deletion, and modification scenarios to measure semantic negligence.
Conclusion: AlignVid effectively addresses semantic negligence in TI2V generation through attention-based modulation, improving prompt adherence while limiting aesthetic degradation, and provides a new benchmark for evaluating these capabilities.
Abstract: Text-guided image-to-video (TI2V) generation has recently achieved remarkable progress, particularly in maintaining subject consistency and temporal coherence. However, existing methods still struggle to adhere to fine-grained prompt semantics, especially when prompts entail substantial transformations of the input image (e.g., object addition, deletion, or modification), a shortcoming we term semantic negligence. In a pilot study, we find that applying a Gaussian blur to the input image improves semantic adherence. Analyzing attention maps, we observe clearer foreground-background separation. From an energy perspective, this corresponds to a lower-entropy cross-attention distribution. Motivated by this, we introduce AlignVid, a training-free framework with two components: (i) Attention Scaling Modulation (ASM), which directly reweights attention via lightweight Q or K scaling, and (ii) Guidance Scheduling (GS), which applies ASM selectively across transformer blocks and denoising steps to reduce visual quality degradation. This minimal intervention improves prompt adherence while limiting aesthetic degradation. In addition, we introduce OmitI2V to evaluate semantic negligence in TI2V generation, comprising 367 human-annotated samples that span addition, deletion, and modification scenarios. Extensive experiments demonstrate that AlignVid can enhance semantic fidelity.
[310] Diffusion Fuzzy System: Fuzzy Rule Guided Latent Multi-Path Diffusion Modeling
Hailong Yang, Te Zhang, Kup-sze Choi, Zhaohong Deng
Main category: cs.CV
TL;DR: DFS is a fuzzy rule-guided multi-path diffusion model that improves feature learning and coordination while reducing computational costs for image generation.
Details
Motivation: Diffusion models struggle with image collections having significant feature differences, often failing to capture complex features and producing conflicting results. Existing multi-path approaches have inefficient coordination and high computational costs.Method: DFS uses multiple diffusion paths each dedicated to specific image features, rule-chain-based reasoning for dynamic coordination, and fuzzy membership-based latent-space compression to reduce computational costs.
Result: DFS achieves more stable training and faster convergence than existing single-path and multi-path diffusion models on LSUN Bedroom, LSUN Church, and MS COCO datasets. It surpasses baselines in image quality, text-image alignment, and accuracy compared to target references.
Conclusion: DFS effectively addresses the limitations of multi-path diffusion models by providing better feature specialization, efficient coordination through fuzzy rules, and reduced computational overhead, resulting in superior image generation performance.
Abstract: Diffusion models have emerged as a leading technique for generating images due to their ability to create high-resolution and realistic images. Despite their strong performance, diffusion models still struggle in managing image collections with significant feature differences. They often fail to capture complex features and produce conflicting results. Research has attempted to address this issue by learning different regions of an image through multiple diffusion paths and then combining them. However, this approach leads to inefficient coordination among multiple paths and high computational costs. To tackle these issues, this paper presents a Diffusion Fuzzy System (DFS), a latent-space multi-path diffusion model guided by fuzzy rules. DFS offers several advantages. First, unlike traditional multi-path diffusion methods, DFS uses multiple diffusion paths, each dedicated to learning a specific class of image features. By assigning each path to a different feature type, DFS overcomes the limitations of multi-path models in capturing heterogeneous image features. Second, DFS employs rule-chain-based reasoning to dynamically steer the diffusion process and enable efficient coordination among multiple paths. Finally, DFS introduces a fuzzy membership-based latent-space compression mechanism to reduce the computational costs of multi-path diffusion effectively. We tested our method on three public datasets: LSUN Bedroom, LSUN Church, and MS COCO. The results show that DFS achieves more stable training and faster convergence than existing single-path and multi-path diffusion models. Additionally, DFS surpasses baseline models in both image quality and alignment between text and images, and also shows improved accuracy when comparing generated images to target references.
[311] EvalTalker: Learning to Evaluate Real-Portrait-Driven Multi-Subject Talking Humans
Yingjie Zhou, Xilei Zhu, Siyu Ren, Ziyi Zhao, Ziwen Wang, Farong Wen, Yu Zhou, Jiezhang Cao, Xiongkuo Min, Fengjiao Chen, Xiaoyu Li, Xuezhi Cao, Guangtao Zhai, Xiaohong Liu
Main category: cs.CV
TL;DR: The paper introduces THQA-MT, the first large-scale dataset for Multi-Talker-generated Talking Human quality assessment, and proposes EvalTalker, a novel quality assessment framework that achieves superior correlation with subjective scores.
Details
Motivation: Current Multi-Talker systems (animating multiple subjects simultaneously) suffer from quality degradation due to technical limitations, leading to suboptimal user experiences. There's a need for better quality assessment methods to advance Multi-Talker generation.Method: 1) Construct THQA-MT dataset with 5,492 Multi-Talker-generated THs from 15 representative Multi-Talkers using 400 real portraits. 2) Conduct subjective experiments to analyze perceptual discrepancies and identify 12 distortion types. 3) Develop EvalTalker framework that perceives global quality, human characteristics, identity consistency, and integrates Qwen-Sync for multimodal synchrony perception.
Result: EvalTalker achieves superior correlation with subjective scores, demonstrating its effectiveness as a quality assessment framework for Multi-Talker-generated content.
Conclusion: The proposed THQA-MT dataset and EvalTalker framework provide a robust foundation for future research on high-quality Multi-Talker generation and evaluation, addressing current quality assessment limitations in multi-subject talking human animation.
Abstract: Speech-driven Talking Human (TH) generation, commonly known as “Talker,” currently faces limitations in multi-subject driving capabilities. Extending this paradigm to “Multi-Talker,” capable of animating multiple subjects simultaneously, introduces richer interactivity and stronger immersion in audiovisual communication. However, current Multi-Talkers still exhibit noticeable quality degradation caused by technical limitations, resulting in suboptimal user experiences. To address this challenge, we construct THQA-MT, the first large-scale Multi-Talker-generated Talking Human Quality Assessment dataset, consisting of 5,492 Multi-Talker-generated THs (MTHs) from 15 representative Multi-Talkers using 400 real portraits collected online. Through subjective experiments, we analyze perceptual discrepancies among different Multi-Talkers and identify 12 common types of distortion. Furthermore, we introduce EvalTalker, a novel TH quality assessment framework. This framework possesses the ability to perceive global quality, human characteristics, and identity consistency, while integrating Qwen-Sync to perceive multimodal synchrony. Experimental results demonstrate that EvalTalker achieves superior correlation with subjective scores, providing a robust foundation for future research on high-quality Multi-Talker generation and evaluation.
[312] Deep Unsupervised Anomaly Detection in Brain Imaging: Large-Scale Benchmarking and Bias Analysis
Alexander Frotscher, Christian F. Baumgartner, Thomas Wolfers
Main category: cs.CV
TL;DR: Large-scale multi-center benchmark of deep unsupervised anomaly detection in brain MRI shows reconstruction-based methods (especially diffusion-inspired) perform best for lesion segmentation, while feature-based methods are more robust to distribution shifts. Current limitations are algorithmic rather than data-related.
Details
Motivation: Fragmented evaluations, heterogeneous datasets, and inconsistent metrics have hindered progress in deep unsupervised anomaly detection for brain MRI toward clinical translation. There's a need for standardized benchmarking.Method: Large-scale multi-center benchmark using 2,976 T1 and 2,972 T2-weighted healthy scans for training, 92 for validation, and 2,221 T1w/1,262 T2w for testing across diverse clinical cohorts. Systematic evaluation of different algorithms’ robustness to scanners, lesion types/sizes, and demographics.
Result: Dice-based segmentation performance varied widely (0.03-0.65). Reconstruction-based methods (especially diffusion-inspired) achieved strongest lesion segmentation, while feature-based methods showed greater robustness to distribution shifts. Systematic biases observed: small/low-contrast lesions missed more often, false positives varied with age/sex. More training data yields only modest gains.
Conclusion: Current unsupervised anomaly detection frameworks are limited algorithmically rather than by data availability. Benchmark establishes foundation for future research and highlights priorities: image native pretraining, principled deviation measures, fairness-aware modeling, and robust domain adaptation for clinical translation.
Abstract: Deep unsupervised anomaly detection in brain magnetic resonance imaging offers a promising route to identify pathological deviations without requiring lesion-specific annotations. Yet, fragmented evaluations, heterogeneous datasets, and inconsistent metrics have hindered progress toward clinical translation. Here, we present a large-scale, multi-center benchmark of deep unsupervised anomaly detection for brain imaging. The training cohort comprised 2,976 T1 and 2,972 T2-weighted scans from healthy individuals across six scanners, with ages ranging from 6 to 89 years. Validation used 92 scans to tune hyperparameters and estimate unbiased thresholds. Testing encompassed 2,221 T1w and 1,262 T2w scans spanning healthy datasets and diverse clinical cohorts. Across all algorithms, the Dice-based segmentation performance varied between 0.03 and 0.65, indicating substantial variability. To assess robustness, we systematically evaluated the impact of different scanners, lesion types and sizes, as well as demographics (age, sex). Reconstruction-based methods, particularly diffusion-inspired approaches, achieved the strongest lesion segmentation performance, while feature-based methods showed greater robustness under distributional shifts. However, systematic biases, such as scanner-related effects, were observed for the majority of algorithms, including that small and low-contrast lesions were missed more often, and that false positives varied with age and sex. Increasing healthy training data yields only modest gains, underscoring that current unsupervised anomaly detection frameworks are limited algorithmically rather than by data availability. Our benchmark establishes a transparent foundation for future research and highlights priorities for clinical translation, including image native pretraining, principled deviation measures, fairness-aware modeling, and robust domain adaptation.
[313] InternVideo-Next: Towards General Video Foundation Models without Video-Text Supervision
Chenting Wang, Yuhan Zhu, Yicheng Xu, Jiange Yang, Ziang Yan, Yali Wang, Yi Wang, Limin Wang
Main category: cs.CV
TL;DR: InternVideo-Next introduces a two-stage pretraining framework with Encoder-Predictor-Decoder architecture to bridge the gap between pixel-level masked video modeling and semantic text supervision, achieving SOTA video representation learning.
Details
Motivation: Current video-text pretraining relies on noisy synthetic captions lacking implicit world knowledge, while masked video modeling struggles with architectural issues: pixel-level reconstruction conflicts with semantics and latent prediction encourages shortcut learning.Method: Proposes Encoder-Predictor-Decoder framework where predictor acts as latent world model. Two-stage pretraining: Stage 1 uses conditional diffusion decoder with image-level semantic priors to bridge pixel fidelity with semantic abstraction; Stage 2 predicts frozen Stage 1 targets to learn world knowledge while mitigating shortcuts.
Result: Achieves state-of-the-art results across benchmarks when trained on public, unlabeled videos, providing scalable path toward general video representation learning.
Conclusion: InternVideo-Next successfully addresses architectural limitations of traditional masked video modeling by disentangling encoder-decoder design and introducing semantic-aware pretraining, enabling effective learning of implicit world knowledge from unlabeled videos.
Abstract: Large-scale video-text pretraining achieves strong performance but depends on noisy, synthetic captions with limited semantic coverage, often overlooking implicit world knowledge such as object motion, 3D geometry, and physical cues. In contrast, masked video modeling (MVM) directly exploits spatiotemporal structures but trails text-supervised methods on general tasks. We find this gap arises from overlooked architectural issues: pixel-level reconstruction struggles with convergence and its low-level requirement often conflicts with semantics, while latent prediction often encourages shortcut learning. To address these, we disentangle the traditional encoder-decoder design into an Encoder-Predictor-Decoder (EPD) framework, where the predictor acts as a latent world model, and propose InternVideo-Next, a two-stage pretraining scheme that builds a semantically consistent yet detail-preserving latent space for this world model. First, conventional linear decoder in pixel MVM enforces the predictor output latent to be linearly projected to, thus separable in pixel space, causing the conflict with semantic abstraction. Our Stage 1 proposes a conditional diffusion decoder and injects reliable image-level semantic priors to enhance semantics and convergence, thus bridging pixel-level fidelity with high-level semantic abstraction. Stage 2 further learns world knowledge by predicting frozen Stage 1 targets within this space, mitigating shortcut learning. Trained on public, unlabeled videos, InternVideo-Next achieves state-of-the-art results across benchmarks and provides a scalable path toward general video representation learning.
[314] Handwritten Text Recognition for Low Resource Languages
Sayantan Dey, Alireza Alaei, Partha Pratim Roy
Main category: cs.CV
TL;DR: BharatOCR: A segmentation-free paragraph-level handwritten text recognition system for Hindi and Urdu using ViT-Transformer Decoder-LM architecture with implicit line segmentation.
Details
Motivation: Paragraph-level handwritten text recognition remains challenging for low-resource languages like Hindi and Urdu that lack comprehensive linguistic resources, requiring robust OCR systems for these scripts.Method: Proposes a ViT-Transformer Decoder-LM architecture: Vision Transformer (ViT) extracts visual features, Transformer decoder generates text sequences, and pre-trained language model (LM) refines output. Uses DeiT for masked image modeling and RoBERTa for masked language modeling. Implements implicit line segmentation to process paragraphs line by line without explicit segmentation.
Result: Achieved benchmark results on multiple datasets: 96.24% character recognition rate on NUST-UHWR, 92.05% on PUCIT-OUHL, 94.80% on Parimal-Urdu, and 80.64% on Hindi dataset. Outperformed several state-of-the-art Urdu text recognition methods.
Conclusion: BharatOCR demonstrates effective paragraph-level handwritten text recognition for low-resource languages through a segmentation-free approach combining vision transformers, transformer decoders, and language models, achieving state-of-the-art performance on Hindi and Urdu datasets.
Abstract: Despite considerable progress in handwritten text recognition, paragraph-level handwritten text recognition, especially in low-resource languages, such as Hindi, Urdu and similar scripts, remains a challenging problem. These languages, often lacking comprehensive linguistic resources, require special attention to develop robust systems for accurate optical character recognition (OCR). This paper introduces BharatOCR, a novel segmentation-free paragraph-level handwritten Hindi and Urdu text recognition. We propose a ViT-Transformer Decoder-LM architecture for handwritten text recognition, where a Vision Transformer (ViT) extracts visual features, a Transformer decoder generates text sequences, and a pre-trained language model (LM) refines the output to improve accuracy, fluency, and coherence. Our model utilizes a Data-efficient Image Transformer (DeiT) model proposed for masked image modeling in this research work. In addition, we adopt a RoBERTa architecture optimized for masked language modeling (MLM) to enhance the linguistic comprehension and generative capabilities of the proposed model. The transformer decoder generates text sequences from visual embeddings. This model is designed to iteratively process a paragraph image line by line, called implicit line segmentation. The proposed model was evaluated using our custom dataset (‘Parimal Urdu’) and (‘Parimal Hindi’), introduced in this research work, as well as two public datasets. The proposed model achieved benchmark results in the NUST-UHWR, PUCIT-OUHL, and Parimal-Urdu datasets, achieving character recognition rates of 96.24%, 92.05%, and 94.80%, respectively. The model also provided benchmark results using the Hindi dataset achieving a character recognition rate of 80.64%. The results obtained from our proposed model indicated that it outperformed several state-of-the-art Urdu text recognition methods.
[315] MasHeNe: A Benchmark for Head and Neck CT Mass Segmentation using Window-Enhanced Mamba with Frequency-Domain Integration
Thao Thi Phuong Dao, Tan-Cong Nguyen, Nguyen Chi Thanh, Truong Hoang Viet, Trong-Le Do, Mai-Khiem Tran, Minh-Khoi Pham, Trung-Nghia Le, Minh-Triet Tran, Thanh Dinh Le
Main category: cs.CV
TL;DR: Introduces MasHeNe dataset for head & neck mass segmentation (beyond just malignant tumors) and proposes WEMF model with tri-window enhancement and multi-frequency attention in U-Mamba architecture.
Details
Motivation: Public datasets for head & neck masses focus mainly on malignant lesions, overlooking other space-occupying conditions like cysts. Need for comprehensive dataset and benchmark for broader head & neck mass segmentation.Method: 1) Created MasHeNe dataset: 3,779 contrast-enhanced CT slices with pixel-level annotations of tumors and cysts. 2) Proposed WEMF model: applies tri-window enhancement to input, uses multi-frequency attention to fuse skip connections within U-shaped Mamba backbone.
Result: WEMF achieves best performance on MasHeNe: Dice 70.45%, IoU 66.89%, NSD 72.33%, HD95 5.12 mm. Dataset provides benchmark for head-and-neck mass segmentation beyond malignancy-only datasets.
Conclusion: MasHeNe addresses gap in head & neck mass segmentation datasets. WEMF shows strong performance but error patterns indicate task remains challenging, requiring further research. Dataset and code publicly available.
Abstract: Head and neck masses are space-occupying lesions that can compress the airway and esophagus and may affect nerves and blood vessels. Available public datasets primarily focus on malignant lesions and often overlook other space-occupying conditions in this region. To address this gap, we introduce MasHeNe, an initial dataset of 3,779 contrast-enhanced CT slices that includes both tumors and cysts with pixel-level annotations. We also establish a benchmark using standard segmentation baselines and report common metrics to enable fair comparison. In addition, we propose the Windowing-Enhanced Mamba with Frequency integration (WEMF) model. WEMF applies tri-window enhancement to enrich the input appearance before feature extraction. It further uses multi-frequency attention to fuse information across skip connections within a U-shaped Mamba backbone. On MasHeNe, WEMF attains the best performance among evaluated methods, with a Dice of 70.45%, IoU of 66.89%, NSD of 72.33%, and HD95 of 5.12 mm. This model indicates stable and strong results on this challenging task. MasHeNe provides a benchmark for head-and-neck mass segmentation beyond malignancy-only datasets. The observed error patterns also suggest that this task remains challenging and requires further research. Our dataset and code are available at https://github.com/drthaodao3101/MasHeNe.git.
[316] OpenBox: Annotate Any Bounding Boxes in 3D
In-Jae Lee, Mungyeom Kim, Kwonyoung Ryu, Pierre Musacchio, Jaesik Park
Main category: cs.CV
TL;DR: OpenBox: A two-stage automatic annotation pipeline using 2D vision foundation models for unsupervised 3D object detection, eliminating need for self-training while improving annotation quality.
Details
Motivation: Current unsupervised 3D object detection methods have limitations: uniform bounding box annotations ignore object physical states, require multiple self-training iterations (computationally expensive), and produce suboptimal quality annotations. Need to reduce annotation costs and recognize unseen objects for autonomous driving safety and scalability.Method: Two-stage pipeline: 1) Cross-modal instance alignment - associate 2D image instance cues from vision foundation model with 3D point clouds; 2) Instance categorization by rigidity/motion state, then generate adaptive bounding boxes using class-specific size statistics. No self-training required.
Result: Experiments on Waymo Open Dataset, Lyft Level 5 Perception dataset, and nuScenes dataset show improved accuracy and efficiency over baselines. Produces high-quality 3D bounding box annotations without self-training.
Conclusion: OpenBox addresses key challenges in unsupervised 3D object detection by leveraging 2D vision foundation models for automatic annotation, eliminating computational overhead of self-training while improving annotation quality through adaptive, state-aware bounding boxes.
Abstract: Unsupervised and open-vocabulary 3D object detection has recently gained attention, particularly in autonomous driving, where reducing annotation costs and recognizing unseen objects are critical for both safety and scalability. However, most existing approaches uniformly annotate 3D bounding boxes, ignore objects’ physical states, and require multiple self-training iterations for annotation refinement, resulting in suboptimal quality and substantial computational overhead. To address these challenges, we propose OpenBox, a two-stage automatic annotation pipeline that leverages a 2D vision foundation model. In the first stage, OpenBox associates instance-level cues from 2D images processed by a vision foundation model with the corresponding 3D point clouds via cross-modal instance alignment. In the second stage, it categorizes instances by rigidity and motion state, then generates adaptive bounding boxes with class-specific size statistics. As a result, OpenBox produces high-quality 3D bounding box annotations without requiring self-training. Experiments on the Waymo Open Dataset, the Lyft Level 5 Perception dataset, and the nuScenes dataset demonstrate improved accuracy and efficiency over baselines.
[317] BlinkBud: Detecting Hazards from Behind via Sampled Monocular 3D Detection on a Single Earbud
Yunzhe Li, Jiajun Yan, Yuzhou Wei, Kechen Liu, Yize Zhao, Chong Zhang, Hongzi Zhu, Li Lu, Shan Chang, Minyi Guo
Main category: cs.CV
TL;DR: BlinkBud: A system using a single earbud and paired phone to detect hazardous objects approaching from behind pedestrians/cyclists, achieving low power consumption and high detection accuracy.
Details
Motivation: Failing to be aware of speeding vehicles approaching from behind poses a huge threat to road safety for pedestrians and cyclists. There's a need for an effective warning system that doesn't require constant visual attention.Method: Uses a single earbud and paired phone to detect approaching hazards. Combines Kalman filter-based trajectory estimation with reinforcement learning for optimal image sampling. Corrects for user head movements using estimated pitch and yaw angles to align coordinate systems and improve depth estimation accuracy.
Result: Lightweight system with ultra-low power consumption (29.8 mW on earbud, 702.6 mW on smartphone). Achieves high detection accuracy with low average false positive ratio (4.90%) and false negative ratio (1.47%).
Conclusion: BlinkBud provides an effective, low-power solution for detecting approaching hazards from behind, significantly improving road safety for vulnerable road users without requiring constant visual attention.
Abstract: Failing to be aware of speeding vehicles approaching from behind poses a huge threat to the road safety of pedestrians and cyclists. In this paper, we propose BlinkBud, which utilizes a single earbud and a paired phone to online detect hazardous objects approaching from behind of a user. The core idea is to accurately track visually identified objects utilizing a small number of sampled camera images taken from the earbud. To minimize the power consumption of the earbud and the phone while guaranteeing the best tracking accuracy, a novel 3D object tracking algorithm is devised, integrating both a Kalman filter based trajectory estimation scheme and an optimal image sampling strategy based on reinforcement learning. Moreover, the impact of constant user head movements on the tracking accuracy is significantly eliminated by leveraging the estimated pitch and yaw angles to correct the object depth estimation and align the camera coordinate system to the user’s body coordinate system, respectively. We implement a prototype BlinkBud system and conduct extensive real-world experiments. Results show that BlinkBud is lightweight with ultra-low mean power consumptions of 29.8 mW and 702.6 mW on the earbud and smartphone, respectively, and can accurately detect hazards with a low average false positive ratio (FPR) and false negative ratio (FNR) of 4.90% and 1.47%, respectively.
[318] SRAM: Shape-Realism Alignment Metric for No Reference 3D Shape Evaluation
Sheng Liu, Tianyu Luan, Phani Nuney, Xuelu Feng, Junsong Yuan
Main category: cs.CV
TL;DR: Proposes Shape-Realism Alignment Metric (SRAM) using LLMs to evaluate 3D shape realism without ground truth, with new RealismGrading dataset and better correlation to human perception than existing methods.
Details
Motivation: Traditional 3D shape evaluation requires ground truth, but realism assessment often doesn't need reference shapes. Need for metric that aligns with human perception of realism without ground truth dependency.Method: 1) Encode 3D meshes into language token space, 2) Use LLM as bridge between shape info and realism evaluation, 3) Design dedicated realism decoder to align LLM output with human perception, 4) Create RealismGrading dataset with human-annotated scores from 16 algorithms across dozen objects.
Result: Metric correlates well with human perceptions, outperforms existing methods, shows good generalizability validated through k-fold cross-validation across different objects.
Conclusion: Proposed SRAM effectively evaluates 3D shape realism without ground truth, leveraging LLMs and human-annotated dataset for better alignment with human perception than traditional methods.
Abstract: 3D generation and reconstruction techniques have been widely used in computer games, film, and other content creation areas. As the application grows, there is a growing demand for 3D shapes that look truly realistic. Traditional evaluation methods rely on a ground truth to measure mesh fidelity. However, in many practical cases, a shape’s realism does not depend on having a ground truth reference. In this work, we propose a Shape-Realism Alignment Metric that leverages a large language model (LLM) as a bridge between mesh shape information and realism evaluation. To achieve this, we adopt a mesh encoding approach that converts 3D shapes into the language token space. A dedicated realism decoder is designed to align the language model’s output with human perception of realism. Additionally, we introduce a new dataset, RealismGrading, which provides human-annotated realism scores without the need for ground truth shapes. Our dataset includes shapes generated by 16 different algorithms on over a dozen objects, making it more representative of practical 3D shape distributions. We validate our metric’s performance and generalizability through k-fold cross-validation across different objects. Experimental results show that our metric correlates well with human perceptions and outperforms existing methods, and has good generalizability.
[319] RoleMotion: A Large-Scale Dataset towards Robust Scene-Specific Role-Playing Motion Synthesis with Fine-grained Descriptions
Junran Peng, Yiheng Huang, Silei Shen, Zeji Wei, Jingwei Yang, Baojie Wang, Yonghao He, Chuanchen Luo, Man Zhang, Xucheng Yin, Wei Sui
Main category: cs.CV
TL;DR: RoleMotion is a new large-scale human motion dataset focused on role-playing and functional motions for specific scenes, addressing limitations of existing datasets.
Details
Motivation: Existing text-to-motion datasets have several limitations: they are decentralized amalgamations of non-functional, isolated motions; have inconsistent quality; and lack fine-grained textual annotations. There's a need for a dataset specifically designed for role-playing and functional motions that work together in various social scenes.Method: The authors meticulously designed and collected RoleMotion with focus on scenes and roles, featuring 25 classic scenes, 110 functional roles, over 500 behaviors, and 10,296 high-quality human motion sequences of body and hands. They annotated these with 27,831 fine-grained text descriptions and built a stronger evaluator than existing counterparts.
Result: The dataset contains comprehensive motion data with fine-grained annotations. The authors proved the reliability of their evaluator and evaluated various text-to-motion methods on their dataset. They also explored the interplay of body and hand motion generation.
Conclusion: RoleMotion demonstrates high-quality and functionality for text-driven whole-body generation, providing a valuable resource for motion generation research with its scene-specific, role-playing focused approach and fine-grained annotations.
Abstract: In this paper, we introduce RoleMotion, a large-scale human motion dataset that encompasses a wealth of role-playing and functional motion data tailored to fit various specific scenes. Existing text datasets are mainly constructed decentrally as amalgamation of assorted subsets that their data are nonfunctional and isolated to work together to cover social activities in various scenes. Also, the quality of motion data is inconsistent, and textual annotation lacks fine-grained details in these datasets. In contrast, RoleMotion is meticulously designed and collected with a particular focus on scenes and roles. The dataset features 25 classic scenes, 110 functional roles, over 500 behaviors, and 10296 high-quality human motion sequences of body and hands, annotated with 27831 fine-grained text descriptions. We build an evaluator stronger than existing counterparts, prove its reliability, and evaluate various text-to-motion methods on our dataset. Finally, we explore the interplay of motion generation of body and hands. Experimental results demonstrate the high-quality and functionality of our dataset on text-driven whole-body generation.
[320] Textured Geometry Evaluation: Perceptual 3D Textured Shape Metric via 3D Latent-Geometry Network
Tianyu Luan, Xuelu Feng, Zixin Zhu, Phani Nuney, Sheng Liu, Xuan Gong, David Doermann, Chunming Qiao, Junsong Yuan
Main category: cs.CV
TL;DR: TGE is a new 3D mesh fidelity evaluation method that jointly uses geometry and color information without rendering, outperforming existing methods on real-world distortions.
Details
Motivation: Existing 3D evaluation metrics don't align well with human perception: Chamfer Distance fails to capture human evaluation, rendering-based methods have incomplete coverage and viewpoint sensitivity, and most are trained on synthetic distortions creating a domain gap.Method: Textured Geometry Evaluation (TGE) directly analyzes 3D meshes with texture without rendering, jointly using geometry and color information to calculate fidelity compared to a reference colored shape.
Result: TGE outperforms rendering-based and geometry-only methods on a human-annotated dataset with real-world distortions.
Conclusion: TGE provides a more human-aligned 3D fidelity evaluation method that works directly on textured meshes, addressing limitations of existing approaches and performing better on real-world distortions.
Abstract: Textured high-fidelity 3D models are crucial for games, AR/VR, and film, but human-aligned evaluation methods still fall behind despite recent advances in 3D reconstruction and generation. Existing metrics, such as Chamfer Distance, often fail to align with how humans evaluate the fidelity of 3D shapes. Recent learning-based metrics attempt to improve this by relying on rendered images and 2D image quality metrics. However, these approaches face limitations due to incomplete structural coverage and sensitivity to viewpoint choices. Moreover, most methods are trained on synthetic distortions, which differ significantly from real-world distortions, resulting in a domain gap. To address these challenges, we propose a new fidelity evaluation method that is based directly on 3D meshes with texture, without relying on rendering. Our method, named Textured Geometry Evaluation TGE, jointly uses the geometry and color information to calculate the fidelity of the input textured mesh with comparison to a reference colored shape. To train and evaluate our metric, we design a human-annotated dataset with real-world distortions. Experiments show that TGE outperforms rendering-based and geometry-only methods on real-world distortion dataset.
[321] Reversible Inversion for Training-Free Exemplar-guided Image Editing
Yuke Li, Lianli Gao, Ji Zhang, Pengpeng Zeng, Lichuan Xiang, Hongkai Wen, Heng Tao Shen, Jingkuan Song
Main category: cs.CV
TL;DR: ReInversion: A training-free two-stage denoising method for exemplar-guided image editing that improves quality and efficiency over standard inversion techniques.
Details
Motivation: Existing exemplar-guided image editing methods require expensive large-scale pre-training or suffer from poor quality and inefficiency with standard inversion techniques. There's a need for effective, efficient training-free alternatives.Method: Reversible Inversion (ReInversion) - a two-stage denoising process: first conditioned on source image, then on reference image. Plus Mask-Guided Selective Denoising (MSD) to constrain edits to target regions while preserving background structure.
Result: Achieves state-of-the-art exemplar-guided image editing performance with the lowest computational overhead, demonstrated through both qualitative and quantitative comparisons.
Conclusion: ReInversion provides an effective and efficient training-free solution for exemplar-guided image editing, overcoming limitations of both pre-training approaches and standard inversion techniques.
Abstract: Exemplar-guided Image Editing (EIE) aims to modify a source image according to a visual reference. Existing approaches often require large-scale pre-training to learn relationships between the source and reference images, incurring high computational costs. As a training-free alternative, inversion techniques can be used to map the source image into a latent space for manipulation. However, our empirical study reveals that standard inversion is sub-optimal for EIE, leading to poor quality and inefficiency. To tackle this challenge, we introduce \textbf{Reversible Inversion ({ReInversion})} for effective and efficient EIE. Specifically, ReInversion operates as a two-stage denoising process, which is first conditioned on the source image and subsequently on the reference. Besides, we introduce a Mask-Guided Selective Denoising (MSD) strategy to constrain edits to target regions, preserving the structural consistency of the background. Both qualitative and quantitative comparisons demonstrate that our ReInversion method achieves state-of-the-art EIE performance with the lowest computational overhead.
[322] Envision: Benchmarking Unified Understanding & Generation for Causal World Process Insights
Juanxi Tian, Siyuan Li, Conghui He, Lijun Wu, Cheng Tan
Main category: cs.CV
TL;DR: The paper introduces Envision, a benchmark for evaluating causal event progression in chained text-to-multi-image generation, revealing that current models struggle with spatiotemporal consistency despite aesthetic quality.
Details
Motivation: Current multimodal models focus on static single-image generation, leading to overfitting to pattern matching and semantic fusion while failing to model dynamic processes over time. There's a need to evaluate models' ability to generate coherent multi-image sequences grounded in world knowledge and causality.Method: Proposed Envision benchmark with 1,000 four-stage prompts across six scientific/humanities domains, structured by spatiotemporal causality. Introduced Envision-Score metric integrating multi-dimensional consistency, physicality, and aesthetics. Evaluated 15 models (10 specialized T2I, 5 unified multimodal).
Result: Specialized T2I models excel in aesthetics but lack world knowledge. Unified multimodal models outperform specialized ones in causal narrative coherence but still lag behind closed-source models and struggle with spatiotemporal consistency. Focus on single images impedes multi-frame reasoning.
Conclusion: Current emphasis on static single-image generation promotes pattern matching over dynamic world modeling, limiting world knowledge internalization. Future work should prioritize spatiotemporal consistency and causal reasoning in multi-frame generation.
Abstract: Current multimodal models aim to transcend the limitations of single-modality representations by unifying understanding and generation, often using text-to-image (T2I) tasks to calibrate semantic consistency. However, their reliance on static, single-image generation in training and evaluation leads to overfitting to static pattern matching and semantic fusion, while fundamentally hindering their ability to model dynamic processes that unfold over time. To address these constraints, we propose Envision-a causal event progression benchmark for chained text-to-multi-image generation. Grounded in world knowledge and structured by spatiotemporal causality, it reorganizes existing evaluation dimensions and includes 1,000 four-stage prompts spanning six scientific and humanities domains. To transition evaluation from single images to sequential frames and assess whether models truly internalize world knowledge while adhering to causal-temporal constraints, we introduce Envision-Score, a holistic metric integrating multi-dimensional consistency, physicality, and aesthetics. Comprehensive evaluation of 15 models (10 specialized T2I models, 5 unified models) uncovers: specialized T2I models demonstrate proficiency in aesthetic rendering yet lack intrinsic world knowledge. Unified multimodal models bridge this gap, consistently outperforming specialized counterparts in causal narrative coherence. However, even these unified architectures remain subordinate to closed-source models and struggle to overcome the core challenge of spatiotemporal consistency. This demonstrates that a focus on causally-isolated single images impedes multi-frame reasoning and generation, promoting static pattern matching over dynamic world modeling-ultimately limiting world knowledge internalization, generation.
[323] PointNet4D: A Lightweight 4D Point Cloud Video Backbone for Online and Offline Perception in Robotic Applications
Yunze Liu, Zifan Wang, Peiran Wu, Jiayang Ao
Main category: cs.CV
TL;DR: PointNet4D: A lightweight 4D backbone for processing streaming point cloud video in real-time using Hybrid Mamba-Transformer temporal fusion and 4DMAP pretraining, achieving strong performance across 9 tasks on 7 datasets with applications in robotics.
Details
Motivation: Current 4D backbone networks rely on computationally intensive spatiotemporal convolutions and Transformers, which are poorly suited for real-time applications in robotics and interactive systems that need to process streaming point cloud video under resource constraints.Method: Proposes PointNet4D with Hybrid Mamba-Transformer temporal fusion block combining efficient state-space modeling (Mamba) with bidirectional modeling power of Transformers. Introduces 4DMAP frame-wise masked auto-regressive pretraining strategy to capture motion cues across frames.
Result: Extensive evaluations across 9 tasks on 7 datasets show consistent improvements across diverse domains. Built two robotic application systems (4D Diffusion Policy and 4D Imitation Learning) achieving substantial gains on RoboTwin and HandoverSim benchmarks.
Conclusion: PointNet4D provides an efficient 4D backbone optimized for both online and offline settings, enabling real-time processing of streaming point cloud video with strong performance across various tasks and practical applications in robotics.
Abstract: Understanding dynamic 4D environments-3D space evolving over time-is critical for robotic and interactive systems. These applications demand systems that can process streaming point cloud video in real-time, often under resource constraints, while also benefiting from past and present observations when available. However, current 4D backbone networks rely heavily on spatiotemporal convolutions and Transformers, which are often computationally intensive and poorly suited to real-time applications. We propose PointNet4D, a lightweight 4D backbone optimized for both online and offline settings. At its core is a Hybrid Mamba-Transformer temporal fusion block, which integrates the efficient state-space modeling of Mamba and the bidirectional modeling power of Transformers. This enables PointNet4D to handle variable-length online sequences efficiently across different deployment scenarios. To enhance temporal understanding, we introduce 4DMAP, a frame-wise masked auto-regressive pretraining strategy that captures motion cues across frames. Our extensive evaluations across 9 tasks on 7 datasets, demonstrating consistent improvements across diverse domains. We further demonstrate PointNet4D’s utility by building two robotic application systems: 4D Diffusion Policy and 4D Imitation Learning, achieving substantial gains on the RoboTwin and HandoverSim benchmarks.
[324] FRAMER: Frequency-Aligned Self-Distillation with Adaptive Modulation Leveraging Diffusion Priors for Real-World Image Super-Resolution
Seungho Choi, Jeahun Sung, Jihyong Oh
Main category: cs.CV
TL;DR: FRAMER is a plug-and-play training scheme for Real-ISR that addresses diffusion models’ low-frequency bias by using final-layer features to teach intermediate layers through frequency-aware contrastive losses and adaptive modulation.
Details
Motivation: Diffusion models for Real-ISR under-reconstruct high-frequency details due to low-frequency bias and a "low-first, high-later" hierarchy in their depth-wise processing, despite surpassing GANs in perceptual quality.Method: Uses final-layer feature maps as teachers for intermediate layers, decomposing features into LF/HF bands via FFT masks. Applies Intra Contrastive Loss for LF structure stabilization and Inter Contrastive Loss for HF detail sharpening with random-layer/in-batch negatives. Includes Frequency-based Adaptive Weight and Frequency-based Alignment Modulation for per-layer signal reweighting and similarity-based distillation gating.
Result: Consistently improves PSNR/SSIM and perceptual metrics (LPIPS, NIQE, MANIQA, MUSIQ) across U-Net and DiT backbones including Stable Diffusion 2 and 3. Ablations validate the final-layer teacher and random-layer negatives.
Conclusion: FRAMER effectively addresses diffusion models’ frequency biases in Real-ISR through frequency-aware feature distillation and contrastive learning, providing plug-and-play improvements without modifying backbone architectures or inference procedures.
Abstract: Real-image super-resolution (Real-ISR) seeks to recover HR images from LR inputs with mixed, unknown degradations. While diffusion models surpass GANs in perceptual quality, they under-reconstruct high-frequency (HF) details due to a low-frequency (LF) bias and a depth-wise “low-first, high-later” hierarchy. We introduce FRAMER, a plug-and-play training scheme that exploits diffusion priors without changing the backbone or inference. At each denoising step, the final-layer feature map teaches all intermediate layers. Teacher and student feature maps are decomposed into LF/HF bands via FFT masks to align supervision with the model’s internal frequency hierarchy. For LF, an Intra Contrastive Loss (IntraCL) stabilizes globally shared structure. For HF, an Inter Contrastive Loss (InterCL) sharpens instance-specific details using random-layer and in-batch negatives. Two adaptive modulators, Frequency-based Adaptive Weight (FAW) and Frequency-based Alignment Modulation (FAM), reweight per-layer LF/HF signals and gate distillation by current similarity. Across U-Net and DiT backbones (e.g., Stable Diffusion 2, 3), FRAMER consistently improves PSNR/SSIM and perceptual metrics (LPIPS, NIQE, MANIQA, MUSIQ). Ablations validate the final-layer teacher and random-layer negatives.
[325] MDiff4STR: Mask Diffusion Model for Scene Text Recognition
Yongkun Du, Miaomiao Zhao, Songlin Fan, Zhineng Chen, Caiyan Jia, Yu-Gang Jiang
Main category: cs.CV
TL;DR: MDiff4STR introduces Mask Diffusion Models to Scene Text Recognition, outperforming auto-regressive models in accuracy while maintaining fast inference with only three denoising steps.
Details
Motivation: Mask Diffusion Models (MDMs) offer flexible efficiency-accuracy balance for vision-language tasks, but vanilla MDMs lag behind auto-regressive models in accuracy for STR. The paper aims to bridge this gap by adapting MDMs specifically for text recognition.Method: Proposes MDiff4STR with two key strategies: 1) Six noising strategies to align training/inference behavior, addressing the noising gap; 2) Token-replacement noise mechanism that provides non-mask noise to encourage revision of overconfident incorrect predictions.
Result: MDiff4STR consistently outperforms popular STR models across diverse benchmarks (irregular, artistic, occluded, Chinese text), surpassing state-of-the-art ARMs in accuracy while maintaining fast inference with only three denoising steps.
Conclusion: MDiff4STR successfully adapts Mask Diffusion Models to Scene Text Recognition, achieving superior accuracy over auto-regressive models while maintaining efficiency, demonstrating the potential of diffusion models for text recognition tasks.
Abstract: Mask Diffusion Models (MDMs) have recently emerged as a promising alternative to auto-regressive models (ARMs) for vision-language tasks, owing to their flexible balance of efficiency and accuracy. In this paper, for the first time, we introduce MDMs into the Scene Text Recognition (STR) task. We show that vanilla MDM lags behind ARMs in terms of accuracy, although it improves recognition efficiency. To bridge this gap, we propose MDiff4STR, a Mask Diffusion model enhanced with two key improvement strategies tailored for STR. Specifically, we identify two key challenges in applying MDMs to STR: noising gap between training and inference, and overconfident predictions during inference. Both significantly hinder the performance of MDMs. To mitigate the first issue, we develop six noising strategies that better align training with inference behavior. For the second, we propose a token-replacement noise mechanism that provides a non-mask noise type, encouraging the model to reconsider and revise overly confident but incorrect predictions. We conduct extensive evaluations of MDiff4STR on both standard and challenging STR benchmarks, covering diverse scenarios including irregular, artistic, occluded, and Chinese text, as well as whether the use of pretraining. Across these settings, MDiff4STR consistently outperforms popular STR models, surpassing state-of-the-art ARMs in accuracy, while maintaining fast inference with only three denoising steps. Code: https://github.com/Topdu/OpenOCR.
[326] \textit{ViRectify}: A Challenging Benchmark for Video Reasoning Correction with Multimodal Large Language Models
Xusen Hei, Jiali Chen, Jinyu Yang, Mengchen Zhao, Yi Cai
Main category: cs.CV
TL;DR: ViRectify is a benchmark for evaluating multimodal LLMs’ ability to identify and correct video reasoning errors across perception, scientific reasoning, and embodied decision-making domains.
Details
Motivation: MLLMs frequently make errors in complex video reasoning scenarios, but existing benchmarks lack systematic evaluation of their error correction capabilities. There's a need to uncover model weaknesses and improve performance through better error identification and correction.Method: Created ViRectify benchmark with 30K+ instances via AI-assisted annotation with human verification. Proposed trajectory evidence-driven correction framework with step-wise error trajectory and reward modeling on visual evidence-grounded correction to focus on error propagation and key timestamps.
Result: GPT-5 achieves only 31.94% correction accuracy on ViRectify, showing it’s a challenging testbed. The framework enables Qwen2.5-VL-7B to outperform 72B variants, demonstrating effectiveness. Analysis reveals systematic asymmetries in error correction across models.
Conclusion: ViRectify provides a comprehensive benchmark for evaluating MLLMs’ video reasoning error correction capabilities, offers a valuable data resource for reflection learning, and introduces a new direction for assessing advanced MLLMs in video reasoning.
Abstract: As multimodal large language models (MLLMs) frequently exhibit errors in complex video reasoning scenarios, correcting these errors is critical for uncovering their weaknesses and improving performance. However, existing benchmarks lack systematic evaluation of MLLMs’ ability to identify and correct these video reasoning errors. To bridge this gap, we propose \textit{ViRectify}, a comprehensive benchmark to evaluate their fine-grained correction capability. Through an AI-assisted annotation pipeline with human verification, we construct a dataset of over 30\textit{K} instances spanning dynamic perception, scientific reasoning, and embodied decision-making domains. In \textit{ViRectify}, we challenge MLLMs to perform step-wise error identification and generate rationales with key video evidence grounding. In addition, we further propose the trajectory evidence-driven correction framework, comprising step-wise error trajectory and reward modeling on visual evidence-grounded correction. It encourages the model to explicitly concentrate on error propagation and key timestamps for correction. Extensive evaluation across 16 advanced MLLMs demonstrates that our \textit{ViRectify} serves as a challenging testbed, where GPT-5 achieves only 31.94% correction accuracy. Our framework enables a Qwen2.5-VL-7B to consistently outperform the variants of 72B on \textit{ViRectify}, showing the effectiveness of our approach. Further analysis uncovers systematic asymmetries in error correction across models, and our dataset is also a valuable data resource to perform reflection learning. We believe \textit{ViRectify} provides a new direction for comprehensively evaluating the advanced MLLMs in video reasoning.
[327] ResDiT: Evoking the Intrinsic Resolution Scalability in Diffusion Transformers
Yiyang Ma, Feng Zhou, Xuedan Yin, Pu Cao, Yonghao Dang, Jianqin Yin
Main category: cs.CV
TL;DR: ResDiT: A training-free method that enables high-resolution image synthesis with Diffusion Transformers by fixing position embedding scaling and adding local enhancement mechanisms.
Details
Motivation: Pre-trained Diffusion Transformers (DiTs) suffer from spatial layout collapse and degraded texture fidelity when scaled to high-resolution image synthesis. Existing solutions require complex multi-stage pipelines, so the authors seek a simpler, more efficient approach that leverages DiTs' intrinsic generative mechanisms.Method: 1) PE scaling technique to correct positional encoding errors when extrapolating to higher resolutions, 2) Local-enhancement mechanism using base-resolution local attention, 3) Patch-level fusion module to aggregate global and local cues, 4) Gaussian-weighted splicing strategy to eliminate grid artifacts.
Result: ResDiT consistently delivers high-fidelity, high-resolution image synthesis without additional training, integrates seamlessly with downstream tasks including spatially controlled generation, and outperforms previous complex pipeline approaches.
Conclusion: ResDiT provides an efficient, training-free solution for high-resolution image synthesis with Diffusion Transformers by addressing core issues in position embeddings and local detail generation, enabling practical high-resolution applications.
Abstract: Leveraging pre-trained Diffusion Transformers (DiTs) for high-resolution (HR) image synthesis often leads to spatial layout collapse and degraded texture fidelity. Prior work mitigates these issues with complex pipelines that first perform a base-resolution (i.e., training-resolution) denoising process to guide HR generation. We instead explore the intrinsic generative mechanisms of DiTs and propose ResDiT, a training-free method that scales resolution efficiently. We identify the core factor governing spatial layout, position embeddings (PEs), and show that the original PEs encode incorrect positional information when extrapolated to HR, which triggers layout collapse. To address this, we introduce a PE scaling technique that rectifies positional encoding under resolution changes. To further remedy low-fidelity details, we develop a local-enhancement mechanism grounded in base-resolution local attention. We design a patch-level fusion module that aggregates global and local cues, together with a Gaussian-weighted splicing strategy that eliminates grid artifacts. Comprehensive evaluations demonstrate that ResDiT consistently delivers high-fidelity, high-resolution image synthesis and integrates seamlessly with downstream tasks, including spatially controlled generation.
[328] Language-Guided Open-World Anomaly Segmentation
Klara Reichard, Nikolas Brasch, Nassir Navab, Federico Tombari
Main category: cs.CV
TL;DR: Clipomaly: A CLIP-based zero-shot method for open-world and anomaly segmentation that detects unknown objects and assigns human-interpretable names without anomaly-specific training.
Details
Motivation: Existing open-world/anomaly segmentation methods fail to assign semantically meaningful labels to unknown regions and struggle to distinguish/learn representations for unknown classes. Open-vocabulary methods require fixed inference vocabularies and cannot handle unconstrained unknown classes in anomaly segmentation.Method: Zero-shot approach leveraging CLIP’s shared image-text embedding space to segment unknown objects and assign human-interpretable names. Unlike open-vocabulary methods, it dynamically extends vocabulary at inference time without retraining, enabling detection and naming of anomalies beyond common class definitions.
Result: Achieves state-of-the-art performance on established anomaly segmentation benchmarks while providing interpretability and flexibility essential for practical deployment in autonomous driving systems.
Conclusion: Clipomaly is the first CLIP-based open-world and anomaly segmentation method that successfully addresses the limitations of existing approaches by enabling zero-shot detection and semantic labeling of unknown objects without requiring anomaly-specific training data.
Abstract: Open-world and anomaly segmentation methods seek to enable autonomous driving systems to detect and segment both known and unknown objects in real-world scenes. However, existing methods do not assign semantically meaningful labels to unknown regions, and distinguishing and learning representations for unknown classes remains difficult. While open-vocabulary segmentation methods show promise in generalizing to novel classes, they require a fixed inference vocabulary and thus cannot be directly applied to anomaly segmentation where unknown classes are unconstrained. We propose Clipomaly, the first CLIP-based open-world and anomaly segmentation method for autonomous driving. Our zero-shot approach requires no anomaly-specific training data and leverages CLIP’s shared image-text embedding space to both segment unknown objects and assign human-interpretable names to them. Unlike open-vocabulary methods, our model dynamically extends its vocabulary at inference time without retraining, enabling robust detection and naming of anomalies beyond common class definitions such as those in Cityscapes. Clipomaly achieves state-of-the-art performance on established anomaly segmentation benchmarks while providing interpretability and flexibility essential for practical deployment.
[329] FastAnimate: Towards Learnable Template Construction and Pose Deformation for Fast 3D Human Avatar Animation
Jian Shu, Nanjie Yao, Gangjian Zhang, Junlong Ren, Yu Feng, Hao Wang
Main category: cs.CV
TL;DR: Proposes a unified learning framework for 3D human avatar animation that addresses template construction inefficiencies and deformation artifacts through a U-Net-based template generator and data-driven refinement.
Details
Motivation: Existing 3D human avatar animation approaches have two main problems: 1) template construction requires extensive skeletal rigging and produces artifacts for specific poses, and 2) target pose deformation suffers from structural distortions caused by Linear Blend Skinning (LBS), which undermines animation realism.Method: A unified learning-based framework with two phases: 1) Uses a U-Net architecture to decouple texture and pose information in a feed-forward process for fast template generation, overcoming inefficiencies and artifacts. 2) Proposes a data-driven refinement technique to enhance structural integrity during target pose deformation.
Result: Extensive experiments show the model delivers consistent performance across diverse poses with optimal balance between efficiency and quality, surpassing state-of-the-art methods.
Conclusion: The proposed framework effectively addresses both template construction inefficiencies and deformation artifacts in 3D human avatar animation, achieving superior performance compared to existing approaches.
Abstract: 3D human avatar animation aims at transforming a human avatar from an arbitrary initial pose to a specified target pose using deformation algorithms. Existing approaches typically divide this task into two stages: canonical template construction and target pose deformation. However, current template construction methods demand extensive skeletal rigging and often produce artifacts for specific poses. Moreover, target pose deformation suffers from structural distortions caused by Linear Blend Skinning (LBS), which significantly undermines animation realism. To address these problems, we propose a unified learning-based framework to address both challenges in two phases. For the former phase, to overcome the inefficiencies and artifacts during template construction, we leverage a U-Net architecture that decouples texture and pose information in a feed-forward process, enabling fast generation of a human template. For the latter phase, we propose a data-driven refinement technique that enhances structural integrity. Extensive experiments show that our model delivers consistent performance across diverse poses with an optimal balance between efficiency and quality,surpassing state-of-the-art (SOTA) methods.
[330] GrndCtrl: Grounding World Models via Self-Supervised Reward Alignment
Haoyang He, Jay Patrikar, Dong-Ki Kim, Max Smith, Daniel McGann, Ali-akbar Agha-mohammadi, Shayegan Omidshafiei, Sebastian Scherer
Main category: cs.CV
TL;DR: RLWG is a self-supervised framework that aligns pretrained video world models with geometric grounding using verifiable rewards, enabling better spatial coherence and navigation stability.
Details
Motivation: Current video world models lack geometric grounding despite high visual fidelity, limiting their use in navigation tasks that require spatial coherence and long-horizon stability.Method: Reinforcement Learning with World Grounding (RLWG) uses self-supervised post-training with geometric and perceptual rewards (pose cycle-consistency, depth reprojection, temporal coherence) to align pretrained world models with physically verifiable structure. Implemented as GrndCtrl using Group Relative Policy Optimization (GRPO).
Result: The framework yields world models that maintain stable trajectories, consistent geometry, and reliable rollouts for embodied navigation, achieving superior spatial coherence and navigation stability over supervised fine-tuning in outdoor environments.
Conclusion: RLWG bridges generative pretraining and grounded behavior through verifiable rewards, analogous to RLVR in language models, enabling geometrically grounded world models for navigation tasks.
Abstract: Recent advances in video world modeling have enabled large-scale generative models to simulate embodied environments with high visual fidelity, providing strong priors for prediction, planning, and control. Yet, despite their realism, these models often lack geometric grounding, limiting their use in navigation tasks that require spatial coherence and long-horizon stability. We introduce Reinforcement Learning with World Grounding (RLWG), a self-supervised post-training framework that aligns pretrained world models with a physically verifiable structure through geometric and perceptual rewards. Analogous to reinforcement learning from verifiable feedback (RLVR) in language models, RLWG can use multiple rewards that measure pose cycle-consistency, depth reprojection, and temporal coherence. We instantiate this framework with GrndCtrl, a reward-aligned adaptation method based on Group Relative Policy Optimization (GRPO), yielding world models that maintain stable trajectories, consistent geometry, and reliable rollouts for embodied navigation. Like post-training alignment in large language models, GrndCtrl leverages verifiable rewards to bridge generative pretraining and grounded behavior, achieving superior spatial coherence and navigation stability over supervised fine-tuning in outdoor environments.
[331] ChronosObserver: Taming 4D World with Hyperspace Diffusion Sampling
Qisen Wang, Yifan Zhao, Peisen Shen, Jialu Li, Jia Li
Main category: cs.CV
TL;DR: ChronosObserver is a training-free method that generates 3D-consistent, time-synchronized multi-view videos using World State Hyperspace representation and guided sampling, without requiring diffusion model training or fine-tuning.
Details
Motivation: Existing camera-controlled video generation models struggle with 3D-consistent, time-synchronized multi-view video generation, which is essential for 4D world creation. Current approaches using data augmentation or test-time optimization face limitations in generalization and scalability.Method: Proposes ChronosObserver with two key components: 1) World State Hyperspace to represent spatiotemporal constraints of 4D scenes, and 2) Hyperspace Guided Sampling to synchronize diffusion sampling trajectories across multiple views using the hyperspace representation.
Result: The method achieves high-fidelity, 3D-consistent, time-synchronized multi-view video generation without requiring training or fine-tuning of diffusion models.
Conclusion: ChronosObserver provides an effective training-free solution for generating synchronized multi-view videos with 3D consistency, addressing limitations of existing approaches while maintaining scalability and generalization.
Abstract: Although prevailing camera-controlled video generation models can produce cinematic results, lifting them directly to the generation of 3D-consistent and high-fidelity time-synchronized multi-view videos remains challenging, which is a pivotal capability for taming 4D worlds. Some works resort to data augmentation or test-time optimization, but these strategies are constrained by limited model generalization and scalability issues. To this end, we propose ChronosObserver, a training-free method including World State Hyperspace to represent the spatiotemporal constraints of a 4D world scene, and Hyperspace Guided Sampling to synchronize the diffusion sampling trajectories of multiple views using the hyperspace. Experimental results demonstrate that our method achieves high-fidelity and 3D-consistent time-synchronized multi-view videos generation without training or fine-tuning for diffusion models.
[332] Enhancing OCR for Sino-Vietnamese Language Processing via Fine-tuned PaddleOCRv5
Minh Hoang Nguyen, Su Nguyen Thiet
Main category: cs.CV
TL;DR: Fine-tuning PaddleOCRv5 improves Classical Chinese (Han-Nom) text recognition accuracy from 37.5% to 50.0% on degraded historical Vietnamese documents.
Details
Motivation: Existing OCR systems struggle with degraded scans, non-standard glyphs, and handwriting variations in ancient Vietnamese Chinese manuscripts, hindering digitization and cross-lingual semantic research.Method: Fine-tuning approach for PaddleOCRv5 using curated ancient Vietnamese Chinese manuscripts, with full training pipeline including preprocessing, LMDB conversion, evaluation, and visualization.
Result: Significant improvement over base model with exact accuracy increasing from 37.5% to 50.0%, particularly effective under noisy image conditions.
Conclusion: The fine-tuned OCR model enables better digitization of historical documents and supports downstream applications like Han-Vietnamese semantic alignment, machine translation, and historical linguistics research, with an interactive demo available.
Abstract: Recognizing and processing Classical Chinese (Han-Nom) texts play a vital role in digitizing Vietnamese historical documents and enabling cross-lingual semantic research. However, existing OCR systems struggle with degraded scans, non-standard glyphs, and handwriting variations common in ancient sources. In this work, we propose a fine-tuning approach for PaddleOCRv5 to improve character recognition on Han-Nom texts. We retrain the text recognition module using a curated subset of ancient Vietnamese Chinese manuscripts, supported by a full training pipeline covering preprocessing, LMDB conversion, evaluation, and visualization. Experimental results show a significant improvement over the base model, with exact accuracy increasing from 37.5 percent to 50.0 percent, particularly under noisy image conditions. Furthermore, we develop an interactive demo that visually compares pre- and post-fine-tuning recognition results, facilitating downstream applications such as Han-Vietnamese semantic alignment, machine translation, and historical linguistics research. The demo is available at https://huggingface.co/spaces/MinhDS/Fine-tuned-PaddleOCRv5
[333] A variational method for curve extraction with curvature-dependent energies
Majid Arthaud, Antonin Chambolle, Vincent Duval
Main category: cs.CV
TL;DR: Variational method for extracting curves between endpoints using energy discretization and Smirnov’s theorem, extended to curvature-dependent energies via lifting to position-orientation space.
Details
Motivation: To develop an automated, mostly unsupervised method for extracting curves and 1D structures from images, particularly handling complex cases with curvature dependencies.Method: Uses variational approach based on energy discretization and Smirnov’s decomposition theorem for vector fields. Implements bi-level minimization for curve extraction. Extends to curvature-dependent energies by lifting curves to position-orientation space with sub-Riemannian/Finslerian metrics.
Result: Developed a framework for automated curve extraction from images that can handle both simple and curvature-dependent cases through mathematical lifting techniques.
Conclusion: The variational approach provides an effective, mostly unsupervised method for extracting 1D structures from images, with extensions to curvature-dependent cases through geometric lifting techniques.
Abstract: We introduce a variational approach for extracting curves between a list of possible endpoints, based on the discretization of an energy and Smirnov’s decomposition theorem for vector fields. It is used to design a bi-level minimization approach to automatically extract curves and 1D structures from an image, which is mostly unsupervised. We extend then the method to curvature-dependent energies, using a now classical lifting of the curves in the space of positions and orientations equipped with an appropriate sub-Riemanian or Finslerian metric.
[334] ELVIS: Enhance Low-Light for Video Instance Segmentation in the Dark
Joanne Lin, Ruirui Lin, Yini Li, David Bull, Nantheera Anantrasirichai
Main category: cs.CV
TL;DR: ELVIS is a framework for adapting video instance segmentation models to low-light conditions, featuring synthetic low-light video generation and enhancement components.
Details
Motivation: Video instance segmentation in low-light conditions is challenging due to noise, blur, low contrast, lack of annotated datasets, and poor performance of existing VIS methods on low-light data.Method: ELVIS includes: 1) unsupervised synthetic low-light video pipeline modeling spatial/temporal degradations, 2) calibration-free degradation profile synthesis network (VDP-Net), and 3) enhancement decoder head that disentangles degradations from content features.
Result: Improves performance by up to +3.7AP on synthetic low-light YouTube-VIS 2019 dataset.
Conclusion: ELVIS enables effective domain adaptation of state-of-the-art VIS models to low-light scenarios, addressing key challenges in low-light video instance segmentation.
Abstract: Video instance segmentation (VIS) for low-light content remains highly challenging for both humans and machines alike, due to adverse imaging conditions including noise, blur and low-contrast. The lack of large-scale annotated datasets and the limitations of current synthetic pipelines, particularly in modeling temporal degradations, further hinder progress. Moreover, existing VIS methods are not robust to the degradations found in low-light videos and, as a result, perform poorly even when finetuned on low-light data. In this paper, we introduce \textbf{ELVIS} (\textbf{E}nhance \textbf{L}ow-light for \textbf{V}ideo \textbf{I}nstance \textbf{S}egmentation), a novel framework that enables effective domain adaptation of state-of-the-art VIS models to low-light scenarios. ELVIS comprises an unsupervised synthetic low-light video pipeline that models both spatial and temporal degradations, a calibration-free degradation profile synthesis network (VDP-Net) and an enhancement decoder head that disentangles degradations from content features. ELVIS improves performances by up to \textbf{+3.7AP} on the synthetic low-light YouTube-VIS 2019 dataset. Code will be released upon acceptance.
[335] Semantic-aware Random Convolution and Source Matching for Domain Generalization in Medical Image Segmentation
Franz Thaler, Martin Urschler, Mateusz Kozinski, Matthias AF Gsell, Gernot Plank, Darko Stern
Main category: cs.CV
TL;DR: SRCSM is a novel single-source domain generalization method for medical image segmentation that uses semantic-aware random convolution for training and intensity mapping at test-time to handle cross-modality and cross-center domain shifts.
Details
Motivation: The paper addresses the challenging problem of single-source domain generalization (DG) for medical image segmentation, where models trained on one domain (e.g., CT) need to generalize to different domains (e.g., MR) without adaptation or access to target domain data during training.Method: SRCSM uses semantic-aware random convolution during training to diversify the source domain by applying different augmentations to different image regions based on annotation labels. At test-time, it maps target domain image intensities to resemble source domain data.
Result: The method outperforms previous DG techniques in most experiments across abdominal, whole-heart, and prostate segmentation tasks in cross-modality and cross-center settings. It even matches in-domain baseline performance in several settings and shows promising results for more challenging cine MR data.
Conclusion: SRCSM establishes a new state-of-the-art in domain generalization for medical image segmentation, demonstrating robust performance across diverse domain shifts and showing potential to close domain gaps even in challenging clinical scenarios.
Abstract: We tackle the challenging problem of single-source domain generalization (DG) for medical image segmentation. To this end, we aim for training a network on one domain (e.g., CT) and directly apply it to a different domain (e.g., MR) without adapting the model and without requiring images or annotations from the new domain during training. We propose a novel method for promoting DG when training deep segmentation networks, which we call SRCSM. During training, our method diversifies the source domain through semantic-aware random convolution, where different regions of a source image are augmented differently, based on their annotation labels. At test-time, we complement the randomization of the training domain via mapping the intensity of target domain images, making them similar to source domain data. We perform a comprehensive evaluation on a variety of cross-modality and cross-center generalization settings for abdominal, whole-heart and prostate segmentation, where we outperform previous DG techniques in a vast majority of experiments. Additionally, we also investigate our method when training on whole-heart CT or MR data and testing on the diastolic and systolic phase of cine MR data captured with different scanner hardware, where we make a step towards closing the domain gap in this even more challenging setting. Overall, our evaluation shows that SRCSM can be considered a new state-of-the-art in DG for medical image segmentation and, moreover, even achieves a segmentation performance that matches the performance of the in-domain baseline in several settings.
[336] Small Drafts, Big Verdict: Information-Intensive Visual Reasoning via Speculation
Yuhan Liu, Lianhui Qin, Shengjie Wang
Main category: cs.CV
TL;DR: SV is a training-free framework that uses multiple lightweight draft VLMs to generate diverse reasoning paths, then a strong verdict VLM synthesizes them for final answers, improving accuracy on information-intensive visual tasks while reducing computational costs.
Details
Motivation: Large VLMs struggle with information-intensive images that densely interleave text and graphical elements, facing challenges in precise localization of critical cues and multi-hop reasoning to integrate dispersed evidence.Method: Two-stage framework: 1) Draft stage: multiple small VLMs act as draft experts to generate diverse reasoning paths with localization candidates; 2) Verdict stage: a strong VLM synthesizes these paths to produce final answer. Includes consensus expert selection to forward only high-agreement paths for efficiency.
Result: Achieves consistent gains on challenging information-intensive and high-resolution VQA benchmarks (InfographicVQA, ChartMuseum, ChartQAPro, HR-Bench 4K). Enables error correction and cost-efficiency compared to large proprietary models or training pipelines.
Conclusion: SV framework successfully addresses challenges in information-intensive visual reasoning by combining multiple draft experts with a verdict model, achieving both improved accuracy and computational efficiency without requiring training.
Abstract: Large Vision-Language Models (VLMs) have achieved remarkable progress in multimodal understanding, yet they struggle when reasoning over information-intensive images that densely interleave textual annotations with fine-grained graphical elements. The main challenges lie in precisely localizing critical cues in dense layouts and multi-hop reasoning to integrate dispersed evidence. We propose Speculative Verdict (SV), a training-free framework inspired by speculative decoding that combines multiple lightweight draft experts with a large verdict model. In the draft stage, small VLMs act as draft experts to generate reasoning paths that provide diverse localization candidates; in the verdict stage, a strong VLM synthesizes these paths to produce the final answer, minimizing computational cost while recovering correct answers. To further improve efficiency and accuracy, SV introduces a consensus expert selection mechanism that forwards only high-agreement reasoning paths to the verdict. Empirically, SV achieves consistent gains on challenging information-intensive and high-resolution visual question answering benchmarks, including InfographicVQA, ChartMuseum, ChartQAPro, and HR-Bench 4K. By synthesizing correct insights from multiple partially accurate reasoning paths, SV achieves both error correction and cost-efficiency compared to large proprietary models or training pipelines. Code is available at https://github.com/Tinaliu0123/speculative-verdict.
[337] QuantumCanvas: A Multimodal Benchmark for Visual Learning of Atomic Interactions
Can Polat, Erchin Serpedin, Mustafa Kurban, Hasan Kurban
Main category: cs.CV
TL;DR: QuantumCanvas is a multimodal benchmark dataset focusing on two-body quantum systems as fundamental building blocks of matter, featuring 2,850 element-element pairs with 18 properties and visual representations of orbital densities for transferable quantum interaction learning.
Details
Motivation: Current molecular and materials machine learning models lack physical transferability because they focus on whole-molecule correlations rather than learning the fundamental quantum interactions between atomic pairs that govern bonding, charge redistribution, orbital hybridization, and electronic coupling.Method: Created QuantumCanvas dataset spanning 2,850 element-element pairs with 18 electronic, thermodynamic, and geometric properties, paired with ten-channel image representations derived from orbital densities, angular field transforms, co-occupancy maps, and charge-density projections. Benchmarking eight architectures across 18 targets.
Result: Achieved mean absolute errors of 0.201 eV on energy gap (GATv2), 0.265 eV on HOMO and 0.274 eV on LUMO (EGNN). DimeNet attained 2.27 eV total-energy MAE and 0.132 eV repulsive-energy MAE. Multimodal fusion achieved 2.15 eV Mermin free-energy MAE. Pretraining improves convergence and generalization on larger datasets.
Conclusion: QuantumCanvas provides a principled, interpretable basis for learning transferable quantum interactions through coupled visual and numerical modalities, unifying orbital physics with vision-based representation learning to address the physical transferability gap in molecular machine learning.
Abstract: Despite rapid advances in molecular and materials machine learning, most models still lack physical transferability: they fit correlations across whole molecules or crystals rather than learning the quantum interactions between atomic pairs. Yet bonding, charge redistribution, orbital hybridization, and electronic coupling all emerge from these two-body interactions that define local quantum fields in many-body systems. We introduce QuantumCanvas, a large-scale multimodal benchmark that treats two-body quantum systems as foundational units of matter. The dataset spans 2,850 element-element pairs, each annotated with 18 electronic, thermodynamic, and geometric properties and paired with ten-channel image representations derived from l- and m-resolved orbital densities, angular field transforms, co-occupancy maps, and charge-density projections. These physically grounded images encode spatial, angular, and electrostatic symmetries without explicit coordinates, providing an interpretable visual modality for quantum learning. Benchmarking eight architectures across 18 targets, we report mean absolute errors of 0.201 eV on energy gap using GATv2, 0.265 eV on HOMO and 0.274 eV on LUMO using EGNN. For energy-related quantities, DimeNet attains 2.27 eV total-energy MAE and 0.132 eV repulsive-energy MAE, while a multimodal fusion model achieves a 2.15 eV Mermin free-energy MAE. Pretraining on QuantumCanvas further improves convergence stability and generalization when fine-tuned on larger datasets such as QM9, MD17, and CrysMTM. By unifying orbital physics with vision-based representation learning, QuantumCanvas provides a principled and interpretable basis for learning transferable quantum interactions through coupled visual and numerical modalities. Dataset and model implementations are available at https://github.com/KurbanIntelligenceLab/QuantumCanvas.
[338] Visual Sync: Multi-Camera Synchronization via Cross-View Object Motion
Shaowei Liu, David Yifan Yao, Saurabh Gupta, Shenlong Wang
Main category: cs.CV
TL;DR: VisualSync is an optimization framework that synchronizes unposed, unsynchronized videos from multiple consumer cameras with millisecond accuracy by exploiting epipolar constraints of moving 3D points.
Details
Motivation: People frequently record events with multiple consumer cameras, but synchronizing these cross-camera streams is challenging. Existing methods require controlled settings, specific targets, manual correction, or expensive hardware, limiting practical use.Method: VisualSync uses multi-view dynamics optimization that leverages off-the-shelf 3D reconstruction, feature matching, and dense tracking to extract tracklets, relative poses, and cross-view correspondences. It then jointly minimizes epipolar error to estimate each camera’s time offset.
Result: Experiments on four diverse, challenging datasets show VisualSync outperforms baseline methods, achieving median synchronization error below 50 ms.
Conclusion: VisualSync provides an effective framework for synchronizing unposed, unsynchronized videos from consumer cameras with high accuracy, overcoming limitations of existing methods that require controlled settings or specialized hardware.
Abstract: Today, people can easily record memorable moments, ranging from concerts, sports events, lectures, family gatherings, and birthday parties with multiple consumer cameras. However, synchronizing these cross-camera streams remains challenging. Existing methods assume controlled settings, specific targets, manual correction, or costly hardware. We present VisualSync, an optimization framework based on multi-view dynamics that aligns unposed, unsynchronized videos at millisecond accuracy. Our key insight is that any moving 3D point, when co-visible in two cameras, obeys epipolar constraints once properly synchronized. To exploit this, VisualSync leverages off-the-shelf 3D reconstruction, feature matching, and dense tracking to extract tracklets, relative poses, and cross-view correspondences. It then jointly minimizes the epipolar error to estimate each camera’s time offset. Experiments on four diverse, challenging datasets show that VisualSync outperforms baseline methods, achieving an median synchronization error below 50 ms.
[339] FlashVGGT: Efficient and Scalable Visual Geometry Transformers with Compressed Descriptor Attention
Zipeng Wang, Dan Xu
Main category: cs.CV
TL;DR: FlashVGGT introduces an efficient descriptor-based attention mechanism for 3D reconstruction from multi-view images, reducing computational overhead while maintaining accuracy comparable to state-of-the-art VGGT.
Details
Motivation: Current state-of-the-art 3D reconstruction methods like VGGT use full self-attention over all image tokens, which suffers from poor scalability due to quadratic complexity and large token counts in long image sequences.Method: FlashVGGT compresses spatial information from each frame into compact descriptor tokens, then computes global attention as cross-attention between full image tokens and this smaller descriptor set. It also enables online inference via chunk-recursive mechanism that reuses cached descriptors.
Result: FlashVGGT achieves reconstruction accuracy competitive with VGGT while reducing inference time to just 9.3% of VGGT for 1,000 images, and scales efficiently to sequences exceeding 3,000 images.
Conclusion: FlashVGGT provides an efficient alternative to full self-attention methods for 3D reconstruction, addressing scalability bottlenecks while maintaining competitive performance, enabling practical applications with long image sequences.
Abstract: 3D reconstruction from multi-view images is a core challenge in computer vision. Recently, feed-forward methods have emerged as efficient and robust alternatives to traditional per-scene optimization techniques. Among them, state-of-the-art models like the Visual Geometry Grounding Transformer (VGGT) leverage full self-attention over all image tokens to capture global relationships. However, this approach suffers from poor scalability due to the quadratic complexity of self-attention and the large number of tokens generated in long image sequences. In this work, we introduce FlashVGGT, an efficient alternative that addresses this bottleneck through a descriptor-based attention mechanism. Instead of applying dense global attention across all tokens, FlashVGGT compresses spatial information from each frame into a compact set of descriptor tokens. Global attention is then computed as cross-attention between the full set of image tokens and this smaller descriptor set, significantly reducing computational overhead. Moreover, the compactness of the descriptors enables online inference over long sequences via a chunk-recursive mechanism that reuses cached descriptors from previous chunks. Experimental results show that FlashVGGT achieves reconstruction accuracy competitive with VGGT while reducing inference time to just 9.3% of VGGT for 1,000 images, and scaling efficiently to sequences exceeding 3,000 images. Our project page is available at https://wzpscott.github.io/flashvggt_page/.
[340] Toward Content-based Indexing and Retrieval of Head and Neck CT with Abscess Segmentation
Thao Thi Phuong Dao, Tan-Cong Nguyen, Trong-Le Do, Truong Hoang Viet, Nguyen Chi Thanh, Huynh Nguyen Thuan, Do Vo Cong Nguyen, Minh-Khoi Pham, Mai-Khiem Tran, Viet-Tham Huynh, Trong-Thuan Nguyen, Trung-Nghia Le, Vo Thanh Toan, Tam V. Nguyen, Minh-Triet Tran, Thanh Dinh Le
Main category: cs.CV
TL;DR: AbscessHeNe: A new dataset of 4,926 contrast-enhanced CT slices with head/neck abscess annotations for semantic segmentation model development, achieving baseline Dice of 0.39.
Details
Motivation: Head and neck abscesses require prompt diagnosis and treatment to prevent sepsis or mortality. Accurate imaging detection and delineation are crucial for clinical decision-making, but there's a lack of comprehensive annotated datasets for developing segmentation models.Method: Created AbscessHeNe dataset with 4,926 contrast-enhanced CT slices of clinically confirmed head/neck abscesses, featuring pixel-level annotations and clinical metadata. Evaluated CNN, Transformer, and Mamba-based segmentation architectures to establish performance baselines.
Result: Best model achieved Dice Similarity Coefficient of 0.39, Intersection-over-Union of 0.27, and Normalized Surface Distance of 0.67, highlighting the difficulty of the segmentation task. Dataset structured for future multimedia indexing and case-based retrieval applications.
Conclusion: AbscessHeNe provides a valuable resource for developing abscess segmentation models and supporting clinical workflows. The modest baseline performance indicates significant challenges and need for further research. Dataset will be publicly available for community use.
Abstract: Abscesses in the head and neck represent an acute infectious process that can potentially lead to sepsis or mortality if not diagnosed and managed promptly. Accurate detection and delineation of these lesions on imaging are essential for diagnosis, treatment planning, and surgical intervention. In this study, we introduce AbscessHeNe, a curated and comprehensively annotated dataset comprising 4,926 contrast-enhanced CT slices with clinically confirmed head and neck abscesses. The dataset is designed to facilitate the development of robust semantic segmentation models that can accurately delineate abscess boundaries and evaluate deep neck space involvement, thereby supporting informed clinical decision-making. To establish performance baselines, we evaluate several state-of-the-art segmentation architectures, including CNN, Transformer, and Mamba-based models. The highest-performing model achieved a Dice Similarity Coefficient of 0.39, Intersection-over-Union of 0.27, and Normalized Surface Distance of 0.67, indicating the challenges of this task and the need for further research. Beyond segmentation, AbscessHeNe is structured for future applications in content-based multimedia indexing and case-based retrieval. Each CT scan is linked with pixel-level annotations and clinical metadata, providing a foundation for building intelligent retrieval systems and supporting knowledge-driven clinical workflows. The dataset will be made publicly available at https://github.com/drthaodao3101/AbscessHeNe.git.
[341] Depth Matching Method Based on ShapeDTW for Oil-Based Mud Imager
Fengfeng Li, Zhou Feng, Hongliang Wu, Hao Zhang, Han Tian, Peng Liu, Lixin Yuan
Main category: cs.CV
TL;DR: A ShapeDTW-based depth matching method for OBM microresistivity images that addresses depth misalignment issues by using local shape features for precise alignment of complex borehole textures.
Details
Motivation: Depth misalignment persists between upper and lower pad images in OBM microresistivity imagers even after velocity correction, requiring more accurate alignment methods for proper geological interpretation.Method: Shape Dynamic Time Warping (ShapeDTW) algorithm with local shape features, using combined HOG1D and original signal as shape descriptors to construct morphologically sensitive distance matrix for sequence alignment.
Result: Field tests show precise alignment for images with complex textures, depth shifts, or local scaling, providing better structural similarity preservation during alignment.
Conclusion: The ShapeDTW-based method effectively solves depth misalignment in borehole imaging and offers a flexible framework for integrating other descriptors tailored to specific geological features.
Abstract: In well logging operations using the oil-based mud (OBM) microresistivity imager, which employs an interleaved design with upper and lower pad sets, depth misalignment issues persist between the pad images even after velocity correction. This paper presents a depth matching method for borehole images based on the Shape Dynamic Time Warping (ShapeDTW) algorithm. The method extracts local shape features to construct a morphologically sensitive distance matrix, better preserving structural similarity between sequences during alignment. We implement this by employing a combined feature set of the one-dimensional Histogram of Oriented Gradients (HOG1D) and the original signal as the shape descriptor. Field test examples demonstrate that our method achieves precise alignment for images with complex textures, depth shifts, or local scaling. Furthermore, it provides a flexible framework for feature extension, allowing the integration of other descriptors tailored to specific geological features.
[342] SPARK: Sim-ready Part-level Articulated Reconstruction with VLM Knowledge
Yumeng He, Ying Jiang, Jiayin Lu, Yin Yang, Chenfanfu Jiang
Main category: cs.CV
TL;DR: SPARK reconstructs physically consistent, kinematic articulated 3D objects from single RGB images using VLMs, diffusion transformers, and differentiable optimization.
Details
Motivation: Creating simulation-ready articulated 3D assets is labor-intensive and requires expert modeling of part hierarchies and motion structures, limiting applications in embodied AI, robotics, and interactive scene understanding.Method: 1) Use VLMs to extract coarse URDF parameters and generate part-level reference images; 2) Integrate part-image guidance and inferred structure graph into generative diffusion transformer for consistent part and complete shape synthesis; 3) Refine URDF parameters using differentiable forward kinematics and rendering optimization under VLM-generated open-state supervision.
Result: Extensive experiments show SPARK produces high-quality, simulation-ready articulated assets across diverse categories, enabling downstream applications like robotic manipulation and interaction modeling.
Conclusion: SPARK provides an effective framework for reconstructing physically consistent, kinematic part-level articulated objects from single images, addressing the labor-intensive nature of creating simulation-ready assets.
Abstract: Articulated 3D objects are critical for embodied AI, robotics, and interactive scene understanding, yet creating simulation-ready assets remains labor-intensive and requires expert modeling of part hierarchies and motion structures. We introduce SPARK, a framework for reconstructing physically consistent, kinematic part-level articulated objects from a single RGB image. Given an input image, we first leverage VLMs to extract coarse URDF parameters and generate part-level reference images. We then integrate the part-image guidance and the inferred structure graph into a generative diffusion transformer to synthesize consistent part and complete shapes of articulated objects. To further refine the URDF parameters, we incorporate differentiable forward kinematics and differentiable rendering to optimize joint types, axes, and origins under VLM-generated open-state supervision. Extensive experiments show that SPARK produces high-quality, simulation-ready articulated assets across diverse categories, enabling downstream applications such as robotic manipulation and interaction modeling.
[343] Generative Editing in the Joint Vision-Language Space for Zero-Shot Composed Image Retrieval
Xin Wang, Haipeng Zhang, Mang Li, Zhaohui Xia, Yueguo Chen, Yu Zhang, Chunyu Wei
Main category: cs.CV
TL;DR: Fusion-Diff is a novel zero-shot composed image retrieval framework that uses multimodal fusion feature editing and a lightweight Control-Adapter to achieve state-of-the-art performance with minimal synthetic data.
Details
Motivation: Supervised CIR methods require costly triplet annotations, motivating zero-shot solutions. Existing zero-shot approaches struggle with the vision-language modality gap, creating a need for more effective and data-efficient methods.Method: Proposes Fusion-Diff with two key innovations: 1) multimodal fusion feature editing in joint vision-language space to narrow modality gap, and 2) lightweight Control-Adapter fine-tuned on only 200K synthetic samples for data efficiency.
Result: Significantly outperforms prior zero-shot approaches on standard CIR benchmarks (CIRR, FashionIQ, CIRCO). Achieves state-of-the-art performance with enhanced interpretability through visualization of fused multimodal representations.
Conclusion: Fusion-Diff effectively addresses the modality gap challenge in zero-shot CIR while maintaining high data efficiency, offering a promising direction for fine-grained visual search without expensive annotations.
Abstract: Composed Image Retrieval (CIR) enables fine-grained visual search by combining a reference image with a textual modification. While supervised CIR methods achieve high accuracy, their reliance on costly triplet annotations motivates zero-shot solutions. The core challenge in zero-shot CIR (ZS-CIR) stems from a fundamental dilemma: existing text-centric or diffusion-based approaches struggle to effectively bridge the vision-language modality gap. To address this, we propose Fusion-Diff, a novel generative editing framework with high effectiveness and data efficiency designed for multimodal alignment. First, it introduces a multimodal fusion feature editing strategy within a joint vision-language (VL) space, substantially narrowing the modality gap. Second, to maximize data efficiency, the framework incorporates a lightweight Control-Adapter, enabling state-of-the-art performance through fine-tuning on only a limited-scale synthetic dataset of 200K samples. Extensive experiments on standard CIR benchmarks (CIRR, FashionIQ, and CIRCO) demonstrate that Fusion-Diff significantly outperforms prior zero-shot approaches. We further enhance the interpretability of our model by visualizing the fused multimodal representations.
[344] Does Understanding Inform Generation in Unified Multimodal Models? From Analysis to Path Forward
Yuwei Niu, Weiyang Jin, Jiaqi Liao, Chaoran Feng, Peng Jin, Bin Lin, Zongjian Li, Bin Zhu, Weihao Yu, Li Yuan
Main category: cs.CV
TL;DR: UniSandbox is a decoupled evaluation framework that reveals a significant understanding-generation gap in multimodal models, showing that explicit Chain-of-Thought reasoning bridges this gap and can be internalized through self-training.
Details
Motivation: Despite progress in Unified Multimodal Models, it's unclear whether understanding truly informs generation. The paper aims to investigate this fundamental question by developing a framework to analyze the understanding-generation gap.Method: Introduces UniSandbox, a decoupled evaluation framework with controlled synthetic datasets to avoid data leakage. Uses explicit Chain-of-Thought (CoT) in understanding modules and self-training approaches to internalize reasoning abilities.
Result: Reveals significant understanding-generation gap in two dimensions: reasoning generation and knowledge transfer. Shows explicit CoT effectively bridges the gap, and self-training can internalize this ability for implicit reasoning during generation. Also finds CoT helps retrieve newly learned knowledge and query-based architectures have latent CoT-like properties.
Conclusion: UniSandbox provides preliminary insights for designing future unified architectures and training strategies to bridge the understanding-generation gap. The framework enables detailed analysis of how understanding informs generation in multimodal models.
Abstract: Recent years have witnessed significant progress in Unified Multimodal Models, yet a fundamental question remains: Does understanding truly inform generation? To investigate this, we introduce UniSandbox, a decoupled evaluation framework paired with controlled, synthetic datasets to avoid data leakage and enable detailed analysis. Our findings reveal a significant understanding-generation gap, which is mainly reflected in two key dimensions: reasoning generation and knowledge transfer. Specifically, for reasoning generation tasks, we observe that explicit Chain-of-Thought (CoT) in the understanding module effectively bridges the gap, and further demonstrate that a self-training approach can successfully internalize this ability, enabling implicit reasoning during generation. Additionally, for knowledge transfer tasks, we find that CoT assists the generative process by helping retrieve newly learned knowledge, and also discover that query-based architectures inherently exhibit latent CoT-like properties that affect this transfer. UniSandbox provides preliminary insights for designing future unified architectures and training strategies that truly bridge the gap between understanding and generation. Code and data are available at https://github.com/PKU-YuanGroup/UniSandBox
[345] ViT$^3$: Unlocking Test-Time Training in Vision
Dongchen Han, Yining Li, Tianyu Li, Zixuan Cao, Ziming Wang, Jun Song, Yu Cheng, Bo Zheng, Gao Huang
Main category: cs.CV
TL;DR: Vision Test-Time Training (ViT³) reformulates attention as online learning at test time, achieving linear complexity and competitive performance across visual tasks.
Details
Motivation: Test-Time Training (TTT) shows promise for efficient sequence modeling but lacks comprehensive design guidelines for visual applications. The paper aims to bridge this gap through systematic empirical study.Method: Conducted systematic empirical study of TTT designs for visual sequence modeling, distilled six practical insights, and developed ViT³ - a pure TTT architecture with linear complexity and parallelizable computation.
Result: ViT³ consistently matches or outperforms advanced linear-complexity models (Mamba, linear attention variants) across image classification, generation, object detection, and segmentation, narrowing the gap to optimized vision Transformers.
Conclusion: The study establishes design principles for effective visual TTT and provides ViT³ as a strong baseline to facilitate future work on efficient visual sequence modeling.
Abstract: Test-Time Training (TTT) has recently emerged as a promising direction for efficient sequence modeling. TTT reformulates attention operation as an online learning problem, constructing a compact inner model from key-value pairs at test time. This reformulation opens a rich and flexible design space while achieving linear computational complexity. However, crafting a powerful visual TTT design remains challenging: fundamental choices for the inner module and inner training lack comprehensive understanding and practical guidelines. To bridge this critical gap, in this paper, we present a systematic empirical study of TTT designs for visual sequence modeling. From a series of experiments and analyses, we distill six practical insights that establish design principles for effective visual TTT and illuminate paths for future improvement. These findings culminate in the Vision Test-Time Training (ViT$^3$) model, a pure TTT architecture that achieves linear complexity and parallelizable computation. We evaluate ViT$^3$ across diverse visual tasks, including image classification, image generation, object detection, and semantic segmentation. Results show that ViT$^3$ consistently matches or outperforms advanced linear-complexity models (e.g., Mamba and linear attention variants) and effectively narrows the gap to highly optimized vision Transformers. We hope this study and the ViT$^3$ baseline can facilitate future work on visual TTT models. Code is available at https://github.com/LeapLabTHU/ViTTT.
[346] DB-KAUNet: An Adaptive Dual Branch Kolmogorov-Arnold UNet for Retinal Vessel Segmentation
Hongyu Xu, Panpan Meng, Meng Wang, Dayu Hu, Liming Liang, Xiaoqi Sheng
Main category: cs.CV
TL;DR: DB-KAUNet: A novel dual-branch UNet architecture combining CNN and Transformer pathways with KANConv/KAT blocks for superior retinal vessel segmentation, achieving state-of-the-art performance on major datasets.
Details
Motivation: Traditional CNN methods for retinal vessel segmentation have limitations in capturing long-range dependencies and complex nonlinear relationships, which are crucial for accurate clinical diagnosis of ophthalmic and systemic diseases.Method: Proposes Adaptive Dual Branch Kolmogorov-Arnold UNet (DB-KAUNet) with Heterogeneous Dual-Branch Encoder (HDBE) featuring parallel CNN and Transformer pathways. Key components include: KANConv and KAT blocks, Cross-Branch Channel Interaction (CCI) module, attention-based Spatial Feature Enhancement (SFE) module, and advanced SFE-GAF module with adaptive sampling for precise vessel morphology focus.
Result: Extensive experiments on DRIVE, STARE, and CHASE_DB1 datasets validate that DB-KAUNet achieves leading segmentation performance and demonstrates exceptional robustness.
Conclusion: DB-KAUNet effectively addresses the limitations of traditional CNN methods by combining CNN and Transformer strengths with novel architectural components, resulting in superior retinal vessel segmentation for clinical applications.
Abstract: Accurate segmentation of retinal vessels is crucial for the clinical diagnosis of numerous ophthalmic and systemic diseases. However, traditional Convolutional Neural Network (CNN) methods exhibit inherent limitations, struggling to capture long-range dependencies and complex nonlinear relationships. To address the above limitations, an Adaptive Dual Branch Kolmogorov-Arnold UNet (DB-KAUNet) is proposed for retinal vessel segmentation. In DB-KAUNet, we design a Heterogeneous Dual-Branch Encoder (HDBE) that features parallel CNN and Transformer pathways. The HDBE strategically interleaves standard CNN and Transformer blocks with novel KANConv and KAT blocks, enabling the model to form a comprehensive feature representation. To optimize feature processing, we integrate several critical components into the HDBE. First, a Cross-Branch Channel Interaction (CCI) module is embedded to facilitate efficient interaction of channel features between the parallel pathways. Second, an attention-based Spatial Feature Enhancement (SFE) module is employed to enhance spatial features and fuse the outputs from both branches. Building upon the SFE module, an advanced Spatial Feature Enhancement with Geometrically Adaptive Fusion (SFE-GAF) module is subsequently developed. In the SFE-GAF module, adaptive sampling is utilized to focus on true vessel morphology precisely. The adaptive process strengthens salient vascular features while significantly reducing background noise and computational overhead. Extensive experiments on the DRIVE, STARE, and CHASE_DB1 datasets validate that DB-KAUNet achieves leading segmentation performance and demonstrates exceptional robustness.
[347] Bridging the Scale Gap: Balanced Tiny and General Object Detection in Remote Sensing Imagery
Zhicheng Zhao, Yin Huang, Lingma Sun, Chenglong Li, Jin Tang
Main category: cs.CV
TL;DR: ScaleBridge-Det: First large detection framework for tiny objects in remote sensing using scale-adaptive expert routing and density-guided query allocation to achieve balanced performance across diverse object scales.
Details
Motivation: Tiny object detection in remote sensing faces challenges with balanced performance across diverse scales, especially when dense tiny objects coexist with large objects. Existing foundation models haven't addressed the extreme scale variation and density distribution in remote sensing imagery.Method: Proposes ScaleBridge-Det with two key modules: 1) Routing-Enhanced Mixture Attention (REM) that dynamically selects/fuses scale-specific expert features via adaptive routing, and 2) Density-Guided Dynamic Query (DGQ) that predicts object density to adaptively adjust query positions/numbers for efficient resource allocation.
Result: Achieves state-of-the-art performance on AI-TOD-V2 and DTOD datasets, while showing superior cross-domain robustness on VisDrone dataset. The framework optimizes performance for both dense tiny and general objects without trade-offs.
Conclusion: ScaleBridge-Det successfully bridges the scale gap in remote sensing object detection, providing the first large detection framework specifically designed for tiny objects that achieves balanced performance across diverse scales through innovative scale-adaptive mechanisms.
Abstract: Tiny object detection in remote sensing imagery has attracted significant research interest in recent years. Despite recent progress, achieving balanced detection performance across diverse object scales remains a formidable challenge, particularly in scenarios where dense tiny objects and large objects coexist. Although large foundation models have revolutionized general vision tasks, their application to tiny object detection remains unexplored due to the extreme scale variation and density distribution inherent to remote sensing imagery. To bridge this scale gap, we propose ScaleBridge-Det, to the best of our knowledge, the first large detection framework designed for tiny objects, which could achieve balanced performance across diverse scales through scale-adaptive expert routing and density-guided query allocation. Specifically, we introduce a Routing-Enhanced Mixture Attention (REM) module that dynamically selects and fuses scale-specific expert features via adaptive routing to address the tendency of standard MoE models to favor dominant scales. REM generates complementary and discriminative multi-scale representations suitable for both tiny and large objects. Furthermore, we present a Density-Guided Dynamic Query (DGQ) module that predicts object density to adaptively adjust query positions and numbers, enabling efficient resource allocation for objects of varying scales. The proposed framework allows ScaleBridge-Det to simultaneously optimize performance for both dense tiny and general objects without trade-offs. Extensive experiments on benchmark and cross-domain datasets demonstrate that ScaleBridge-Det achieves state-of-the-art performance on AI-TOD-V2 and DTOD, while exhibiting superior cross-domain robustness on VisDrone.
[348] GRASP: Guided Residual Adapters with Sample-wise Partitioning
Felix Nützel, Mischa Dombrowski, Bernhard Kainz
Main category: cs.CV
TL;DR: GRASP addresses mode collapse in diffusion models for long-tail data by using guided residual adapters with sample-wise partitioning to resolve gradient conflicts between frequent and rare classes.
Details
Motivation: Text-to-image diffusion models struggle with long-tail distributions like medical imaging, where rare pathologies have limited data, causing mode collapse that reduces quality and diversity of synthetic data for underrepresented conditions.Method: GRASP uses external priors to partition samples into clusters minimizing intra-group gradient conflicts, then fine-tunes pre-trained models by injecting cluster-specific residual adapters into transformer feedforward layers, bypassing learned gating for stability and efficiency.
Result: On MIMIC-CXR-LT dataset, GRASP achieves superior FID and diversity metrics, especially for rare classes, outperforming baselines like vanilla fine-tuning and Mixture of Experts variants. Downstream classification on NIH-CXR-LT improves for tail labels, and generalization to ImageNet-LT confirms broad applicability.
Conclusion: GRASP provides a lightweight, scalable solution to gradient conflicts in long-tail diffusion models, enabling better synthetic data generation for underrepresented classes while maintaining efficiency and integration with existing diffusion pipelines.
Abstract: Recent advances in text-to-image diffusion models enable high-fidelity generation across diverse prompts. However, these models falter in long-tail settings, such as medical imaging, where rare pathologies comprise a small fraction of data. This results in mode collapse: tail-class outputs lack quality and diversity, undermining the goal of synthetic data augmentation for underrepresented conditions. We pinpoint gradient conflicts between frequent head and rare tail classes as the primary culprit, a factor unaddressed by existing sampling or conditioning methods that mainly steer inference without altering the learned distribution. To resolve this, we propose GRASP: Guided Residual Adapters with Sample-wise Partitioning. GRASP uses external priors to statically partition samples into clusters that minimize intra-group gradient clashes. It then fine-tunes pre-trained models by injecting cluster-specific residual adapters into transformer feedforward layers, bypassing learned gating for stability and efficiency. On the long-tail MIMIC-CXR-LT dataset, GRASP yields superior FID and diversity metrics, especially for rare classes, outperforming baselines like vanilla fine-tuning and Mixture of Experts variants. Downstream classification on NIH-CXR-LT improves considerably for tail labels. Generalization to ImageNet-LT confirms broad applicability. Our method is lightweight, scalable, and readily integrates with diffusion pipelines.
[349] Open-world Hand-Object Interaction Video Generation Based on Structure and Contact-aware Representation
Haodong Yan, Hang Yu, Zhide Zhong, Weilin Yuan, Xin Gong, Zehang Luo, Chengxi Heyu, Junfeng Li, Wenxuan Song, Shunbo Zhou, Haoang Li
Main category: cs.CV
TL;DR: Proposes SCAR, a structure and contact-aware representation for generating realistic hand-object interaction videos without 3D annotations, using a joint-generation paradigm.
Details
Motivation: Current methods for hand-object interaction video generation face a dilemma between 2D and 3D representations - 2D lacks interaction fidelity while 3D lacks scalability. There's a need for representation that captures physical constraints (contact, occlusion) without requiring 3D annotations.Method: Proposes SCAR (Structure and Contact-Aware Representation) that captures hand-object contact, occlusion, and holistic structure context without 3D annotations. Introduces joint-generation paradigm with share-and-specialization strategy to generate both interaction-oriented representations and videos.
Result: Outperforms state-of-the-art methods on two real-world datasets in generating physics-realistic and temporally coherent HOI videos. Shows strong generalization to challenging open-world scenarios.
Conclusion: The proposed SCAR representation provides scalable yet interaction-faithful supervision for HOI video generation, enabling learning of fine-grained interaction physics and generalization to open-world scenarios without 3D annotations.
Abstract: Generating realistic hand-object interactions (HOI) videos is a significant challenge due to the difficulty of modeling physical constraints (e.g., contact and occlusion between hands and manipulated objects). Current methods utilize HOI representation as an auxiliary generative objective to guide video synthesis. However, there is a dilemma between 2D and 3D representations that cannot simultaneously guarantee scalability and interaction fidelity. To address this limitation, we propose a structure and contact-aware representation that captures hand-object contact, hand-object occlusion, and holistic structure context without 3D annotations. This interaction-oriented and scalable supervision signal enables the model to learn fine-grained interaction physics and generalize to open-world scenarios. To fully exploit the proposed representation, we introduce a joint-generation paradigm with a share-and-specialization strategy that generates interaction-oriented representations and videos. Extensive experiments demonstrate that our method outperforms state-of-the-art methods on two real-world datasets in generating physics-realistic and temporally coherent HOI videos. Furthermore, our approach exhibits strong generalization to challenging open-world scenarios, highlighting the benefit of our scalable design. Our project page is https://hgzn258.github.io/SCAR/.
[350] Cross-Domain Validation of a Resection-Trained Self-Supervised Model on Multicentre Mesothelioma Biopsies
Farzaneh Seyedshahi, Francesca Damiola, Sylvie Lantuejoul, Ke Yuan, John Le Quesne
Main category: cs.CV
TL;DR: Self-supervised AI model trained on mesothelioma resection specimens can be applied to small biopsies to predict survival and classify tumor subtypes.
Details
Motivation: Most computational pathology models require large resection specimens, but real-world clinical practice often involves small biopsies, creating a gap between research models and practical application.Method: Train a self-supervised encoder on large tissue images from resection specimens, then apply it to smaller biopsy material to capture meaningful morphological patterns for analysis.
Result: The model successfully captures morphological patterns from biopsy material and can predict patient survival and classify tumor subtypes using these patterns.
Conclusion: AI-driven tools trained on resection tissue can be effectively applied to biopsy material, demonstrating potential to support mesothelioma diagnosis and treatment planning in real-world clinical settings.
Abstract: Accurate subtype classification and outcome prediction in mesothelioma are essential for guiding therapy and patient care. Most computational pathology models are trained on large tissue images from resection specimens, limiting their use in real-world settings where small biopsies are common. We show that a self-supervised encoder trained on resection tissue can be applied to biopsy material, capturing meaningful morphological patterns. Using these patterns, the model can predict patient survival and classify tumor subtypes. This approach demonstrates the potential of AI-driven tools to support diagnosis and treatment planning in mesothelioma.
[351] DreamingComics: A Story Visualization Pipeline via Subject and Layout Customized Generation using Video Models
Patrick Kwon, Chen Chen
Main category: cs.CV
TL;DR: DreamingComics is a layout-aware story visualization framework that improves character consistency and artistic style by integrating layout control with a pretrained video diffusion-transformer model.
Details
Motivation: Current story visualization methods position subjects solely by text and struggle with maintaining artistic consistency, lacking proper layout control for comic-style storytelling.Method: Builds on pretrained video DiT model; introduces RegionalRoPE for layout-based position control via region-aware positional encoding; adds masked condition loss for spatial constraints; integrates LLM-based layout generator for comic-style layouts from natural language scripts.
Result: Achieves 29.2% increase in character consistency and 36.2% increase in style similarity compared to previous methods, with high spatial accuracy.
Conclusion: DreamingComics effectively addresses layout control and consistency issues in story visualization, enabling flexible comic-style generation with improved artistic quality.
Abstract: Current story visualization methods tend to position subjects solely by text and face challenges in maintaining artistic consistency. To address these limitations, we introduce DreamingComics, a layout-aware story visualization framework. We build upon a pretrained video diffusion-transformer (DiT) model, leveraging its spatiotemporal priors to enhance identity and style consistency. For layout-based position control, we propose RegionalRoPE, a region-aware positional encoding scheme that re-indexes embeddings based on the target layout. Additionally, we introduce a masked condition loss to further constrain each subject’s visual features to their designated region. To infer layouts from natural language scripts, we integrate an LLM-based layout generator trained to produce comic-style layouts, enabling flexible and controllable layout conditioning. We present a comprehensive evaluation of our approach, showing a 29.2% increase in character consistency and a 36.2% increase in style similarity compared to previous methods, while displaying high spatial accuracy. Our project page is available at https://yj7082126.github.io/dreamingcomics/
[352] SSR: Semantic and Spatial Rectification for CLIP-based Weakly Supervised Segmentation
Xiuli Bi, Die Xiao, Junchao Fan, Bin Xiao
Main category: cs.CV
TL;DR: SSR method addresses CLIP-based WSSS limitations using semantic and spatial rectification to reduce over-activation in non-target foreground and background regions.
Details
Motivation: Existing CLIP-based weakly supervised semantic segmentation approaches suffer from over-activation in non-target foreground regions and background areas, limiting their segmentation accuracy.Method: Proposes Semantic and Spatial Rectification (SSR) with two components: 1) Cross-Modal Prototype Alignment (CMPA) for semantic-level contrastive learning to align feature spaces across modalities, and 2) Superpixel-Guided Correction (SGC) for spatial-level filtering using superpixel-based priors during affinity propagation.
Result: Achieves state-of-the-art performance on PASCAL VOC (79.5% mIoU) and MS COCO (50.6% mIoU), outperforming both single-stage and more complex multi-stage approaches.
Conclusion: The SSR method effectively addresses over-activation issues in CLIP-based WSSS through semantic and spatial rectification, demonstrating superior segmentation performance across benchmark datasets.
Abstract: In recent years, Contrastive Language-Image Pretraining (CLIP) has been widely applied to Weakly Supervised Semantic Segmentation (WSSS) tasks due to its powerful cross-modal semantic understanding capabilities. This paper proposes a novel Semantic and Spatial Rectification (SSR) method to address the limitations of existing CLIP-based weakly supervised semantic segmentation approaches: over-activation in non-target foreground regions and background areas. Specifically, at the semantic level, the Cross-Modal Prototype Alignment (CMPA) establishes a contrastive learning mechanism to enforce feature space alignment across modalities, reducing inter-class overlap while enhancing semantic correlations, to rectify over-activation in non-target foreground regions effectively; at the spatial level, the Superpixel-Guided Correction (SGC) leverages superpixel-based spatial priors to precisely filter out interference from non-target regions during affinity propagation, significantly rectifying background over-activation. Extensive experiments on the PASCAL VOC and MS COCO datasets demonstrate that our method outperforms all single-stage approaches, as well as more complex multi-stage approaches, achieving mIoU scores of 79.5% and 50.6%, respectively.
[353] FreqEdit: Preserving High-Frequency Features for Robust Multi-Turn Image Editing
Yucheng Liao, Jiajun Liang, Kaiqian Cui, Baoquan Zhao, Haoran Xie, Wei Liu, Qing Li, Xudong Mao
Main category: cs.CV
TL;DR: FreqEdit is a training-free framework that addresses quality degradation in multi-turn image editing by preserving high-frequency information through reference velocity injection, adaptive spatial modulation, and trajectory compensation.
Details
Motivation: Current instruction-based image editing models work well for single edits but suffer severe quality degradation in multi-turn editing scenarios due to progressive loss of high-frequency information.Method: FreqEdit uses three components: (1) high-frequency feature injection from reference velocity fields to preserve details, (2) adaptive injection strategy for spatial modulation of injection strength, and (3) path compensation mechanism to periodically recalibrate editing trajectory.
Result: FreqEdit enables stable editing across 10+ consecutive iterations and outperforms seven state-of-the-art baselines in both identity preservation and instruction following.
Conclusion: The training-free FreqEdit framework effectively addresses multi-turn editing degradation by preserving high-frequency information, offering a practical solution for stable iterative image editing.
Abstract: Instruction-based image editing through natural language has emerged as a powerful paradigm for intuitive visual manipulation. While recent models achieve impressive results on single edits, they suffer from severe quality degradation under multi-turn editing. Through systematic analysis, we identify progressive loss of high-frequency information as the primary cause of this quality degradation. We present FreqEdit, a training-free framework that enables stable editing across 10+ consecutive iterations. Our approach comprises three synergistic components: (1) high-frequency feature injection from reference velocity fields to preserve fine-grained details, (2) an adaptive injection strategy that spatially modulates injection strength for precise region-specific control, and (3) a path compensation mechanism that periodically recalibrates the editing trajectory to prevent over-constraint. Extensive experiments demonstrate that FreqEdit achieves superior performance in both identity preservation and instruction following compared to seven state-of-the-art baselines.
[354] HiconAgent: History Context-aware Policy Optimization for GUI Agents
Xurui Zhou, Gongwei Chen, Yuquan Xie, Zaijing Li, Kaiwen Zhou, Shuai Wang, Shuo Yang, Zhuotao Tian, Rui Shao
Main category: cs.CV
TL;DR: HiconAgent is a GUI agent that efficiently uses historical context through History Context-aware Policy Optimization (HCPO), achieving better performance with reduced computational costs.
Details
Motivation: GUI agents need historical context for sequential navigation tasks, but naive use of full history causes computational overhead and distraction from irrelevant information.Method: HCPO with two components: (1) Dynamic Context Sampling (DCS) presents variable length histories during sampling for adaptive context use; (2) Anchor-guided History Compression (AHC) uses a dual branch strategy with compressed branch removing history observations while keeping actions as anchors, coupled through history-enhanced alignment loss.
Result: HiconAgent-3B outperforms GUI-R1-7B by +8.46% grounding accuracy and +11.32% step success rate on GUI-Odyssey, achieves comparable results on AndroidControl and AITW with up to 2.47x speedup and 60% FLOPs reduction.
Conclusion: HiconAgent demonstrates efficient and effective historical context utilization for GUI agents through HCPO, achieving strong performance with reduced computational requirements.
Abstract: Graphical User Interface (GUI) agents require effective use of historical context to perform sequential navigation tasks. While incorporating past actions and observations can improve decision making, naive use of full history leads to excessive computational overhead and distraction from irrelevant information. To address this, we introduce HiconAgent, a GUI agent trained with History Context-aware Policy Optimization (HCPO) for efficient and effective utilization of historical information. HCPO optimizes history usage in both sampling and policy updates through two complementary components: (1) Dynamic Context Sampling (DCS) presents the agent with variable length histories during sampling, enabling adaptive use of the most relevant context; (2) Anchor-guided History Compression (AHC) refines the policy update phase with a dual branch strategy where the compressed branch removes history observations while keeping history actions as information flow anchors. The compressed and uncompressed branches are coupled through a history-enhanced alignment loss to enforce consistent history usage while maintaining efficiency. Experiments on mainstream GUI navigation benchmarks demonstrate strong performance. Despite being smaller, HiconAgent-3B outperforms GUI-R1-7B by +8.46 percent grounding accuracy and +11.32 percent step success rate on GUI-Odyssey, while achieving comparable results on AndroidControl and AITW with up to 2.47x computational speedup and 60 percent FLOPs reduction.
[355] VideoScoop: A Non-Traditional Domain-Independent Framework For Video Analysis
Hafsa Billah
Main category: cs.CV
TL;DR: A general-purpose Video Situation Analysis framework that combines relational and graph models with continuous query processing for automated video understanding across multiple domains.
Details
Motivation: Current video situation analysis is either manual (error-prone, labor-intensive) or uses custom algorithms specific to particular video types/situations, lacking general-purpose solutions that work across different domains.Method: Extracts video content using state-of-the-art techniques, represents it using extended relational model (R++) and graph models, enables continuous query processing via a specialized query language, and uses parameterized templates for domain independence.
Result: Extensive experiments across Assisted Living, Civic Monitoring, and Surveillance domains show the approach achieves accuracy, efficiency, and robustness with videos of varying lengths.
Conclusion: The proposed framework overcomes limitations of manual analysis and domain-specific algorithms by providing a general-purpose solution for automated video situation analysis across multiple application domains.
Abstract: Automatically understanding video contents is important for several applications in Civic Monitoring (CM), general Surveillance (SL), Assisted Living (AL), etc. Decades of Image and Video Analysis (IVA) research have advanced tasks such as content extraction (e.g., object recognition and tracking). Identifying meaningful activities or situations (e.g., two objects coming closer) remains difficult and cannot be achieved by content extraction alone. Currently, Video Situation Analysis (VSA) is done manually with a human in the loop, which is error-prone and labor-intensive, or through custom algorithms designed for specific video types or situations. These algorithms are not general-purpose and require a new algorithm/software for each new situation or video from a new domain. This report proposes a general-purpose VSA framework that overcomes the above limitations. Video contents are extracted once using state-of-the-art Video Content Extraction technologies. They are represented using two alternative models – the extended relational model (R++) and graph models. When represented using R++, the extracted contents can be used as data streams, enabling Continuous Query Processing via the proposed Continuous Query Language for Video Analysis. The graph models complement this by enabling the detection of situations that are difficult or impossible to detect using the relational model alone. Existing graph algorithms and newly developed algorithms support a wide variety of situation detection. To support domain independence, primitive situation variants across domains are identified and expressed as parameterized templates. Extensive experiments were conducted across several interesting situations from three domains – AL, CM, and SL– to evaluate the accuracy, efficiency, and robustness of the proposed approach using a dataset of videos of varying lengths from these domains.
[356] Robust Rigid and Non-Rigid Medical Image Registration Using Learnable Edge Kernels
Ahsan Raza Siyal, Markus Haltmeier, Ruth Steiger, Malik Galijasevic, Elke Ruth Gizewski, Astrid Ellen Grams
Main category: cs.CV
TL;DR: Proposes medical image registration method using learnable edge kernels with rigid/non-rigid registration, outperforming state-of-the-art across multiple datasets.
Details
Motivation: Medical image registration faces challenges with contrast differences, spatial distortions, and modality variations. Traditional methods struggle with these issues, requiring better approaches for multi-modal alignment in clinical applications.Method: Integrates learnable edge kernels with learning-based rigid and non-rigid registration. Starts with predefined edge detection kernel perturbed with random noise, then learns optimal edge features during training. Introduces four variant models each for rigid and non-rigid registration to analyze component contributions.
Result: Method consistently outperformed state-of-the-art techniques across all experiments: rigid registration without/with skull removal, non-rigid registration on Medical University dataset, and two public datasets.
Conclusion: The adaptive edge detection approach enhances medical image registration by capturing diverse structural features, demonstrating potential to improve multi-modal image alignment and anatomical structure analysis.
Abstract: Medical image registration is crucial for various clinical and research applications including disease diagnosis or treatment planning which require alignment of images from different modalities, time points, or subjects. Traditional registration techniques often struggle with challenges such as contrast differences, spatial distortions, and modality-specific variations. To address these limitations, we propose a method that integrates learnable edge kernels with learning-based rigid and non-rigid registration techniques. Unlike conventional layers that learn all features without specific bias, our approach begins with a predefined edge detection kernel, which is then perturbed with random noise. These kernels are learned during training to extract optimal edge features tailored to the task. This adaptive edge detection enhances the registration process by capturing diverse structural features critical in medical imaging. To provide clearer insight into the contribution of each component in our design, we introduce four variant models for rigid registration and four variant models for non-rigid registration. We evaluated our approach using a dataset provided by the Medical University across three setups: rigid registration without skull removal, with skull removal, and non-rigid registration. Additionally, we assessed performance on two publicly available datasets. Across all experiments, our method consistently outperformed state-of-the-art techniques, demonstrating its potential to improve multi-modal image alignment and anatomical structure analysis.
[357] Evaluating SAM2 for Video Semantic Segmentation
Syed Hesham Syed Ariff, Yun Liu, Guolei Sun, Jing Yang, Henghui Ding, Xue Geng, Xudong Jiang
Main category: cs.CV
TL;DR: This paper explores extending SAM2 (Segmentation Anything Model 2) from video object segmentation to dense Video Semantic Segmentation (VSS), proposing two approaches that leverage SAM2’s object boundary precision to improve VSS performance.
Details
Motivation: While SAM2 excels at promptable visual object segmentation in images and videos, extending it to dense Video Semantic Segmentation (VSS) presents challenges including spatial accuracy, temporal consistency, and tracking multiple objects with complex boundaries and varying scales. The paper aims to address these challenges by adapting SAM2's capabilities to the VSS task.Method: Two primary approaches: 1) Using SAM2 to extract unique objects as masks from images, with a segmentation network generating and refining initial predictions in parallel. 2) Using predicted masks to extract unique feature vectors, feeding them into a simple network for classification, then combining classifications and masks for final segmentation.
Result: Experiments suggest that leveraging SAM2 enhances overall performance in VSS, primarily due to its precise predictions of object boundaries.
Conclusion: SAM2 can be effectively extended to Video Semantic Segmentation, with its object boundary precision providing significant performance improvements despite the challenges of spatial accuracy, temporal consistency, and multi-object tracking in complex video scenes.
Abstract: The Segmentation Anything Model 2 (SAM2) has proven to be a powerful foundation model for promptable visual object segmentation in both images and videos, capable of storing object-aware memories and transferring them temporally through memory blocks. While SAM2 excels in video object segmentation by providing dense segmentation masks based on prompts, extending it to dense Video Semantic Segmentation (VSS) poses challenges due to the need for spatial accuracy, temporal consistency, and the ability to track multiple objects with complex boundaries and varying scales. This paper explores the extension of SAM2 for VSS, focusing on two primary approaches and highlighting firsthand observations and common challenges faced during this process. The first approach involves using SAM2 to extract unique objects as masks from a given image, with a segmentation network employed in parallel to generate and refine initial predictions. The second approach utilizes the predicted masks to extract unique feature vectors, which are then fed into a simple network for classification. The resulting classifications and masks are subsequently combined to produce the final segmentation. Our experiments suggest that leveraging SAM2 enhances overall performance in VSS, primarily due to its precise predictions of object boundaries.
[358] Learned Image Compression for Earth Observation: Implications for Downstream Segmentation Tasks
Christian Mollière, Iker Cumplido, Marco Zeulner, Lukas Liesenhoff, Matthias Schubert, Julia Gottfriedsen
Main category: cs.CV
TL;DR: Learned compression outperforms JPEG 2000 for large-scale EO imagery but traditional codecs remain competitive for smaller thermal datasets; joint optimization doesn’t improve over standalone approaches.
Details
Motivation: Address the data transmission and storage challenges posed by rapidly growing satellite Earth observation data by evaluating task-specific learned compression algorithms.Method: Compare traditional compression (JPEG 2000) versus learned compression (Discretized Mixed Gaussian Likelihood) on three EO segmentation tasks: fire, cloud, and building detection. Also test joint end-to-end optimization of compression and segmentation models.
Result: Learned compression significantly outperforms JPEG 2000 for large-scale, multi-channel optical imagery in both reconstruction quality (PSNR) and segmentation accuracy. Traditional codecs remain competitive on smaller, single-channel thermal infrared datasets. Joint optimization doesn’t improve performance over standalone optimization.
Conclusion: Learned compression is promising for large-scale EO data but context matters; traditional methods still have value for certain datasets, and joint optimization doesn’t provide additional benefits.
Abstract: The rapid growth of data from satellite-based Earth observation (EO) systems poses significant challenges in data transmission and storage. We evaluate the potential of task-specific learned compression algorithms in this context to reduce data volumes while retaining crucial information. In detail, we compare traditional compression (JPEG 2000) versus a learned compression approach (Discretized Mixed Gaussian Likelihood) on three EO segmentation tasks: Fire, cloud, and building detection. Learned compression notably outperforms JPEG 2000 for large-scale, multi-channel optical imagery in both reconstruction quality (PSNR) and segmentation accuracy. However, traditional codecs remain competitive on smaller, single-channel thermal infrared datasets due to limited data and architectural constraints. Additionally, joint end-to-end optimization of compression and segmentation models does not improve performance over standalone optimization.
[359] SAM3-UNet: Simplified Adaptation of Segment Anything Model 3
Xinyu Xiong, Zihuang Wu, Lei Lu, Yufa Xia
Main category: cs.CV
TL;DR: SAM3-UNet is a simplified, efficient variant of Segment Anything Model 3 designed for low-cost adaptation to downstream segmentation tasks, outperforming previous methods while using less than 6GB GPU memory.
Details
Motivation: To create a cost-effective adaptation of the powerful SAM3 foundation model for downstream segmentation tasks without requiring extensive computational resources or parameter tuning.Method: Three-component architecture: 1) SAM3 image encoder for feature extraction, 2) simple adapter for parameter-efficient fine-tuning, and 3) lightweight U-Net-style decoder for segmentation output.
Result: Outperforms prior SAM2-UNet and other state-of-the-art methods on multiple tasks (mirror detection, salient object detection) while requiring less than 6GB GPU memory during training with batch size 12.
Conclusion: SAM3-UNet provides an effective, resource-efficient solution for adapting large foundation models to downstream segmentation tasks, making advanced segmentation capabilities more accessible.
Abstract: In this paper, we introduce SAM3-UNet, a simplified variant of Segment Anything Model 3 (SAM3), designed to adapt SAM3 for downstream tasks at a low cost. Our SAM3-UNet consists of three components: a SAM3 image encoder, a simple adapter for parameter-efficient fine-tuning, and a lightweight U-Net-style decoder. Preliminary experiments on multiple tasks, such as mirror detection and salient object detection, demonstrate that the proposed SAM3-UNet outperforms the prior SAM2-UNet and other state-of-the-art methods, while requiring less than 6 GB of GPU memory during training with a batch size of 12. The code is publicly available at https://github.com/WZH0120/SAM3-UNet.
[360] Generative Action Tell-Tales: Assessing Human Motion in Synthesized Videos
Xavier Thomas, Youngsun Lim, Ananya Srinivasan, Audrey Zheng, Deepti Ghadiyaram
Main category: cs.CV
TL;DR: A new evaluation metric for video generation that measures human action plausibility by comparing generated videos to real-world action distributions using fused skeletal and appearance features.
Details
Motivation: Existing video evaluation metrics are appearance-biased and lack temporal understanding, struggling to assess complex human action dynamics and anatomical plausibility in generated videos.Method: Learns a latent space from real-world human actions by fusing appearance-agnostic skeletal geometry features with appearance-based features. Measures action quality by computing distance between generated video representations and this learned real-world action distribution.
Result: Achieves 68% improvement over state-of-the-art methods on their new benchmark, performs competitively on established external benchmarks, and shows stronger correlation with human perception.
Conclusion: The metric reveals critical limitations in current video generative models and establishes a new standard for advanced video generation research by providing robust evaluation of human action fidelity.
Abstract: Despite rapid advances in video generative models, robust metrics for evaluating visual and temporal correctness of complex human actions remain elusive. Critically, existing pure-vision encoders and Multimodal Large Language Models (MLLMs) are strongly appearance-biased, lack temporal understanding, and thus struggle to discern intricate motion dynamics and anatomical implausibilities in generated videos. We tackle this gap by introducing a novel evaluation metric derived from a learned latent space of real-world human actions. Our method first captures the nuances, constraints, and temporal smoothness of real-world motion by fusing appearance-agnostic human skeletal geometry features with appearance-based features. We posit that this combined feature space provides a robust representation of action plausibility. Given a generated video, our metric quantifies its action quality by measuring the distance between its underlying representations and this learned real-world action distribution. For rigorous validation, we develop a new multi-faceted benchmark specifically designed to probe temporally challenging aspects of human action fidelity. Through extensive experiments, we show that our metric achieves substantial improvement of more than 68% compared to existing state-of-the-art methods on our benchmark, performs competitively on established external benchmarks, and has a stronger correlation with human perception. Our in-depth analysis reveals critical limitations in current video generative models and establishes a new standard for advanced research in video generation.
[361] Seeing through Imagination: Learning Scene Geometry via Implicit Spatial World Modeling
Meng Cao, Haokun Lin, Haoyuan Li, Haoran Tang, Rongtao Xu, Dong An, Xue Liu, Ian Reid, Xiaodan Liang
Main category: cs.CV
TL;DR: MILO introduces an implicit spatial world modeling paradigm with visual generator feedback and relative positional encoding to improve MLLMs’ spatial reasoning by grounding symbolic reasoning in perceptual experience.
Details
Motivation: Current MLLMs suffer from visual illiteracy in spatial reasoning - they learn spatial concepts through textual symbols alone without connection to visual manifestations, limiting their ability to understand 3D structure.Method: 1) MILO paradigm integrates visual generator for geometry-aware feedback to implicitly ground symbolic reasoning; 2) RePE encoding captures relative camera-pose transformations instead of absolute coordinates; 3) GeoGen dataset with 2,241 videos and 67,827 observation-action-outcome triplets for training.
Result: Experiments show significant enhancement of spatial reasoning capabilities across multiple baselines and benchmarks, offering more holistic understanding of 3D space.
Conclusion: MILO successfully bridges the gap between symbolic reasoning and visual perception in spatial understanding, providing a more human-like spatial imagination capability for MLLMs.
Abstract: Spatial reasoning, the ability to understand and interpret the 3D structure of the world, is a critical yet underdeveloped capability in Multimodal Large Language Models (MLLMs). Current methods predominantly rely on verbal descriptive tuning, which suffers from visual illiteracy, i.e., they learn spatial concepts through textual symbols alone, devoid of connection to their visual manifestations. To bridge this gap, this paper introduces MILO, an Implicit spatIaL wOrld modeling paradigm that simulates human-like spatial imagination. MILO integrates a visual generator to provide geometry-aware feedback, thereby implicitly grounding the MLLM’s symbolic reasoning in perceptual experience. Complementing this paradigm, we propose RePE (Relative Positional Encoding), a novel encoding scheme that captures relative camera-pose transformations, offering superior performance over absolute coordinate systems. To support the training, we construct GeoGen, a large-scale Geometry-aware Generative dataset with approximately 2,241 videos and 67,827 observation-action-outcome triplets. Experiments demonstrate that our approach significantly enhances spatial reasoning capabilities across multiple baselines and benchmarks, offering a more holistic understanding of 3D space.
[362] CauSight: Learning to Supersense for Visual Causal Discovery
Yize Zhang, Meiqi Chen, Sirui Chen, Bo Peng, Yanxi Zhang, Tianyu Li, Chaochao Lu
Main category: cs.CV
TL;DR: The paper introduces visual causal discovery as a new AI task, creates a large-scale VCG-32K dataset with entity-level causal graphs, and develops CauSight model that outperforms GPT-4.1 with a threefold performance boost.
Details
Motivation: To replicate human causal thinking in AI systems, enabling models to understand why things happen rather than just what is seen, moving beyond mere perception to infer cause-and-effect relations among visual entities.Method: Three-component training approach: (1) data curation from VCG-32K dataset (32K+ images with causal graphs), (2) Tree-of-Causal-Thought for synthesizing reasoning trajectories, and (3) reinforcement learning with causal reward to refine reasoning policy.
Result: CauSight outperforms GPT-4.1 on visual causal discovery, achieving over a threefold performance boost with 21% absolute gain. The model, code, and dataset are fully open-sourced.
Conclusion: The work successfully introduces visual causal discovery as a new AI task, demonstrates that specialized models can significantly outperform general-purpose LLMs on causal reasoning in visual domains, and provides open resources to advance the field.
Abstract: Causal thinking enables humans to understand not just what is seen, but why it happens. To replicate this capability in modern AI systems, we introduce the task of visual causal discovery. It requires models to infer cause-and-effect relations among visual entities across diverse scenarios instead of merely perceiving their presence. To this end, we first construct the Visual Causal Graph dataset (VCG-32K), a large-scale collection of over 32,000 images annotated with entity-level causal graphs, and further develop CauSight, a novel vision-language model to perform visual causal discovery through causally aware reasoning. Our training recipe integrates three components: (1) training data curation from VCG-32K, (2) Tree-of-Causal-Thought (ToCT) for synthesizing reasoning trajectories, and (3) reinforcement learning with a designed causal reward to refine the reasoning policy. Experiments show that CauSight outperforms GPT-4.1 on visual causal discovery, achieving over a threefold performance boost (21% absolute gain). Our code, model, and dataset are fully open-sourced at project page: https://github.com/OpenCausaLab/CauSight.
[363] OpenREAD: Reinforced Open-Ended Reasoing for End-to-End Autonomous Driving with LLM-as-Critic
Songyan Zhang, Wenhui Huang, Zhan Chen, Chua Jiahao Collister, Qihang Huang, Chen Lv
Main category: cs.CV
TL;DR: OpenREAD is an end-to-end reinforcement fine-tuning framework for autonomous driving that uses LLM-based reward modeling to enhance reasoning and planning across both upstream and downstream tasks.
Details
Motivation: Current two-stage fine-tuning approaches for autonomous driving have limitations: SFT lacks generalization in reasoning, and RFT is only applied to downstream tasks due to difficulty quantifying rewards for open-ended scene understanding problems.Method: Proposes OpenREAD framework with: 1) Large-scale Chain-of-Thought annotations on driving datasets, 2) Using Qwen3 LLM as critic in RFT to quantify reasoning quality for open-ended questions, 3) End-to-end RFT across both high-level reasoning and low-level trajectory planning.
Result: Extensive experiments show joint end-to-end RFT yields substantial improvements in both upstream and downstream tasks, achieving state-of-the-art performance on reasoning and planning benchmarks.
Conclusion: OpenREAD successfully addresses limitations of current approaches by enabling comprehensive reinforcement fine-tuning across the full spectrum of autonomous driving tasks, demonstrating the effectiveness of LLM-based reward modeling for open-ended reasoning problems.
Abstract: Recently, two-stage fine-tuning strategies, e.g., acquiring essential driving knowledge through supervised fine-tuning (SFT) and further enhancing decision-making and planning via reinforcement fine-tuning (RFT), have shown strong potential in advancing the knowledge-driven autonomous driving (AD) paradigm. However, the learning nature of SFT still limits the generalization of reasoning, thereby constraining the full potential of driving performance. Meanwhile, current RFT approaches are primarily applied to downstream tasks, since scene understanding is an open-ended problem where corresponding rewards are difficult to quantify. To address these limitations, we propose OpenREAD, an OPEN-ended REasoning reinforced vision-language model (VLM)-based autonomous driving (AD) framework that enables end-to-end RFT across the full spectrum from high-level reasoning to low-level trajectory planning. Specifically, we begin by constructing large-scale Chain-of-Thought (CoT) annotations on open-source driving-related knowledge datasets, and employ the powerful Qwen3 large language model (LLM) as the critic in RFT to quantify reasoning quality for open-ended questions during reward modeling. Extensive experiments confirm that joint end-to-end RFT yields substantial improvements in both upstream and downstream tasks, enabling OpenREAD to achieve state-of-the-art performance on reasoning and planning benchmarks.
[364] PhyDetEx: Detecting and Explaining the Physical Plausibility of T2V Models
Zeqing Wang, Keze Wang, Lei Zhang
Main category: cs.CV
TL;DR: The paper introduces PhyDetEx, a fine-tuned VLM for detecting physical implausibility in T2V generated videos and explaining violated physical principles, along with a PID dataset for training and benchmarking T2V models on physical plausibility.
Details
Motivation: While T2V models have improved in quality and scale, their ability to generate physically plausible videos remains questionable. Current VLMs struggle to identify physically impossible content, creating a need for better evaluation methods.Method: Constructed a PID dataset with test/train splits of annotated videos where implausible videos are created by rewriting captions to induce physical violations. Fine-tuned VLMs to detect implausibility and generate explanations (PhyDetEx).
Result: PhyDetEx effectively detects physical implausibility and explains violated principles. Benchmarking shows T2V models still struggle with physical laws, especially open-source models, despite recent progress.
Conclusion: Understanding and adhering to physical laws remains challenging for T2V models. The proposed PhyDetEx provides a valuable tool for evaluating physical plausibility, with dataset and code publicly available.
Abstract: Driven by the growing capacity and training scale, Text-to-Video (T2V) generation models have recently achieved substantial progress in video quality, length, and instruction-following capability. However, whether these models can understand physics and generate physically plausible videos remains a question. While Vision-Language Models (VLMs) have been widely used as general-purpose evaluators in various applications, they struggle to identify the physically impossible content from generated videos. To investigate this issue, we construct a \textbf{PID} (\textbf{P}hysical \textbf{I}mplausibility \textbf{D}etection) dataset, which consists of a \textit{test split} of 500 manually annotated videos and a \textit{train split} of 2,588 paired videos, where each implausible video is generated by carefully rewriting the caption of its corresponding real-world video to induce T2V models producing physically implausible content. With the constructed dataset, we introduce a lightweight fine-tuning approach, enabling VLMs to not only detect physically implausible events but also generate textual explanations on the violated physical principles. Taking the fine-tuned VLM as a physical plausibility detector and explainer, namely \textbf{PhyDetEx}, we benchmark a series of state-of-the-art T2V models to assess their adherence to physical laws. Our findings show that although recent T2V models have made notable progress toward generating physically plausible content, understanding and adhering to physical laws remains a challenging issue, especially for open-source models. Our dataset, training code, and checkpoints are available at \href{https://github.com/Zeqing-Wang/PhyDetEx}{https://github.com/Zeqing-Wang/PhyDetEx}.
[365] Register Any Point: Scaling 3D Point Cloud Registration by Flow Matching
Yue Pan, Tao Sun, Liyuan Zhu, Lucas Nunes, Iro Armeni, Jens Behley, Cyrill Stachniss
Main category: cs.CV
TL;DR: RAP casts point cloud registration as conditional generation using learned velocity fields to transport noisy points to registered scenes, achieving SOTA results on benchmarks.
Details
Motivation: Traditional registration methods rely on correspondence matching and pairwise transformation optimization, which can be complex and may not handle low-overlap cases well. The authors aim to simplify registration by directly generating registered point clouds through conditional generation.Method: Proposes a conditional generation approach where a learned continuous, point-wise velocity field transports noisy points to a registered scene. Uses a lightweight local feature extractor and test-time rigidity enforcement. Unlike previous methods that do correspondence matching and pairwise optimization, this directly generates the registered point cloud.
Result: Achieves state-of-the-art results on pairwise and multi-view registration benchmarks, particularly with low overlap. Generalizes across scales and sensor modalities. Supports downstream tasks including relocalization, multi-robot SLAM, and multi-session map merging.
Conclusion: The conditional generation approach to point cloud registration is effective, achieving superior performance on challenging benchmarks and demonstrating practical utility for various 3D vision and robotics applications.
Abstract: Point cloud registration aligns multiple unposed point clouds into a common frame, and is a core step for 3D reconstruction and robot localization. In this work, we cast registration as conditional generation: a learned continuous, point-wise velocity field transports noisy points to a registered scene, from which the pose of each view is recovered. Unlike previous methods that conduct correspondence matching to estimate the transformation between a pair of point clouds and then optimize the pairwise transformations to realize multi-view registration, our model directly generates the registered point cloud. With a lightweight local feature extractor and test-time rigidity enforcement, our approach achieves state-of-the-art results on pairwise and multi-view registration benchmarks, particularly with low overlap, and generalizes across scales and sensor modalities. It further supports downstream tasks including relocalization, multi-robot SLAM, and multi-session map merging. Source code available at: https://github.com/PRBonn/RAP.
[366] COACH: Collaborative Agents for Contextual Highlighting - A Multi-Agent Framework for Sports Video Analysis
Tsz-To Wong, Ching-Chun Huang, Hong-Han Shuai
Main category: cs.CV
TL;DR: A reconfigurable Multi-Agent System (MAS) framework for sports video analysis that uses specialized agents as “cognitive tools” to handle temporal hierarchies from micro-actions to macro-strategies, enabling adaptive pipelines for both short-term reasoning and long-term summarization.
Details
Motivation: Existing end-to-end models struggle with temporal hierarchy in sports video analysis, lacking generalization, having high development costs for new tasks, and suffering from poor interpretability.Method: Proposes a reconfigurable Multi-Agent System where each agent specializes in specific analysis aspects; uses iterative invocation and flexible composition of agents to construct adaptive pipelines for different temporal scales and tasks.
Result: Demonstrated adaptability using two representative tasks in badminton analysis, showing ability to bridge fine-grained event detection and global semantic organization.
Conclusion: Presents a paradigm shift toward flexible, scalable, and interpretable systems for robust, cross-task sports video intelligence.
Abstract: Intelligent sports video analysis demands a comprehensive understanding of temporal context, from micro-level actions to macro-level game strategies. Existing end-to-end models often struggle with this temporal hierarchy, offering solutions that lack generalization, incur high development costs for new tasks, and suffer from poor interpretability. To overcome these limitations, we propose a reconfigurable Multi-Agent System (MAS) as a foundational framework for sports video understanding. In our system, each agent functions as a distinct “cognitive tool” specializing in a specific aspect of analysis. The system’s architecture is not confined to a single temporal dimension or task. By leveraging iterative invocation and flexible composition of these agents, our framework can construct adaptive pipelines for both short-term analytic reasoning (e.g., Rally QA) and long-term generative summarization (e.g., match summaries). We demonstrate the adaptability of this framework using two representative tasks in badminton analysis, showcasing its ability to bridge fine-grained event detection and global semantic organization. This work presents a paradigm shift towards a flexible, scalable, and interpretable system for robust, cross-task sports video intelligence.The project homepage is available at https://aiden1020.github.io/COACH-project-page
[367] TransientTrack: Advanced Multi-Object Tracking and Classification of Cancer Cells with Transient Fluorescent Signals
Florian Bürger, Martim Dias Gomes, Nica Gutu, Adrián E. Granada, Noémie Moreau, Katarzyna Bozek
Main category: cs.CV
TL;DR: TransientTrack is a deep learning framework for cell tracking in multi-channel microscopy videos with transient fluorescent signals, capable of detecting cell division and death events to build complete cell lineages.
Details
Motivation: Current cell tracking methods are limited to videos with constant signals and cannot detect critical events like cell death, creating a need for methods that can handle transient fluorescent signals and capture complete cell dynamics including division and death events.Method: A lightweight deep learning framework that performs matching on cell detection embeddings without tracking-specific feature quantification, integrates Transformer Networks, uses multi-stage matching with all detection boxes, and interpolates missing tracklets with Kalman Filter.
Result: The framework achieves strong performance across diverse conditions, effectively tracks cells while capturing division and death events, and demonstrates practical utility in analyzing chemotherapeutic drug efficacy at single-cell level.
Conclusion: TransientTrack advances quantitative cancer cell dynamics studies by enabling detailed characterization of treatment response and resistance mechanisms, with code publicly available for broader use.
Abstract: Tracking cells in time-lapse videos is an essential technique for monitoring cell population dynamics at a single-cell level. Current methods for cell tracking are developed on videos with mostly single, constant signals and do not detect pivotal events such as cell death. Here, we present TransientTrack, a deep learning-based framework for cell tracking in multi-channel microscopy video data with transient fluorescent signals that fluctuate over time following processes such as the circadian rhythm of cells. By identifying key cellular events - mitosis (cell division) and apoptosis (cell death) our method allows us to build complete trajectories, including cell lineage information. TransientTrack is lightweight and performs matching on cell detection embeddings directly, without the need for quantification of tracking-specific cell features. Furthermore, our approach integrates Transformer Networks, multi-stage matching using all detection boxes, and the interpolation of missing tracklets with the Kalman Filter. This unified framework achieves strong performance across diverse conditions, effectively tracking cells and capturing cell division and death. We demonstrate the use of TransientTrack in an analysis of the efficacy of a chemotherapeutic drug at a single-cell level. The proposed framework could further advance quantitative studies of cancer cell dynamics, enabling detailed characterization of treatment response and resistance mechanisms. The code is available at https://github.com/bozeklab/TransientTrack.
[368] KM-ViPE: Online Tightly Coupled Vision-Language-Geometry Fusion for Open-Vocabulary Semantic SLAM
Zaid Nasser, Mikhail Iumanov, Tianhao Li, Maxim Popov, Jaafar Mahmoud, Malik Mohrat, Ilya Obrubov, Ekaterina Derevyanka, Ivan Sosin, Sergey Kolyubin
Main category: cs.CV
TL;DR: KM-ViPE is a real-time open-vocabulary SLAM system that works with uncalibrated monocular cameras in dynamic environments, using DINO features and geometric constraints for robust online localization and semantic mapping.
Details
Motivation: Existing SLAM systems often require depth sensors, offline calibration, or lack robustness in dynamic scenes. There's a need for systems that can operate in real-time with uncalibrated monocular cameras while handling moving objects and movable static objects, especially for ego-centric applications and harvesting internet-scale video data.Method: KM-ViPE tightly couples DINO visual features with geometric constraints using a high-level features based adaptive robust kernel. It performs simultaneous online localization and open-vocabulary semantic mapping by fusing geometric and deep visual features aligned with language embeddings, operating directly on raw RGB streams without depth sensors or calibration.
Result: The system achieves competitive results with state-of-the-art approaches while operating online with uncalibrated monocular input. It uniquely combines online operation, uncalibrated monocular input, and robust handling of dynamic scenes, outperforming solutions that require offline processing, depth data, or odometry estimation.
Conclusion: KM-ViPE advances practical spatial intelligence for embodied AI, making it suitable for autonomous robotics and AR/VR applications. Its ability to operate in real-time with uncalibrated monocular cameras while handling dynamic scenes represents significant progress toward practical spatial understanding systems.
Abstract: We present KM-ViPE (Knowledge Mapping Video Pose Engine), a real-time open-vocabulary SLAM framework for uncalibrated monocular cameras in dynamic environments. Unlike systems requiring depth sensors and offline calibration, KM-ViPE operates directly on raw RGB streams, making it ideal for ego-centric applications and harvesting internet-scale video data for training. KM-ViPE tightly couples DINO visual features with geometric constraints through a high-level features based adaptive robust kernel that handles both moving objects and movable static objects (e.g., moving furniture in ego-centric views). The system performs simultaneous online localization and open-vocabulary semantic mapping by fusing geometric and deep visual features aligned with language embeddings. Our results are competitive with state-of-the-art approaches, while existing solutions either operate offline, need depth data and/or odometry estimation, or lack dynamic scene robustness. KM-ViPE benefits from internet-scale training and uniquely combines online operation, uncalibrated monocular input, and robust handling of dynamic scenes, which makes it a good fit for autonomous robotics and AR/VR applications and advances practical spatial intelligence capabilities for embodied AI.
[369] StyleYourSmile: Cross-Domain Face Retargeting Without Paired Multi-Style Data
Avirup Dey, Vinay Namboodiri
Main category: cs.CV
TL;DR: StyleYourSmile is a one-shot cross-domain face retargeting method that disentangles identity, expressions, and domain-specific styles without needing curated multi-style paired data.
Details
Motivation: Existing methods for cross-domain face retargeting either fail to generalize across domains, require test-time optimizations, or need fine-tuning with carefully curated multi-style datasets to achieve domain-invariant identity representations.Method: Proposes an efficient data augmentation strategy alongside a dual-encoder framework to extract domain-invariant identity cues and capture domain-specific stylistic variations, then conditions a diffusion model to retarget facial expressions across domains.
Result: Extensive experiments demonstrate that StyleYourSmile achieves superior identity preservation and retargeting fidelity across a wide range of visual domains.
Conclusion: StyleYourSmile enables effective cross-domain face retargeting with disentangled control over identity, expressions, and domain-specific styles without requiring curated multi-style paired data.
Abstract: Cross-domain face retargeting requires disentangled control over identity, expressions, and domain-specific stylistic attributes. Existing methods, typically trained on real-world faces, either fail to generalize across domains, need test-time optimizations, or require fine-tuning with carefully curated multi-style datasets to achieve domain-invariant identity representations. In this work, we introduce \textit{StyleYourSmile}, a novel one-shot cross-domain face retargeting method that eliminates the need for curated multi-style paired data. We propose an efficient data augmentation strategy alongside a dual-encoder framework, for extracting domain-invariant identity cues and capturing domain-specific stylistic variations. Leveraging these disentangled control signals, we condition a diffusion model to retarget facial expressions across domains. Extensive experiments demonstrate that \textit{StyleYourSmile} achieves superior identity preservation and retargeting fidelity across a wide range of visual domains.
[370] SARL: Spatially-Aware Self-Supervised Representation Learning for Visuo-Tactile Perception
Gurmeher Khurana, Lan Wei, Dandan Zhang
Main category: cs.CV
TL;DR: SARL is a spatially-aware self-supervised learning framework for fused visual-tactile data that preserves spatial structure through map-level objectives, achieving 30% improvement over baselines on geometry-sensitive tasks.
Details
Motivation: Contact-rich robotic manipulation needs representations that encode local geometry. While fused visuo-tactile sensors provide aligned visual and tactile data, most SSL frameworks compress spatial information into global vectors, losing the spatial structure crucial for manipulation tasks.Method: SARL augments BYOL architecture with three map-level objectives: Saliency Alignment (SAL) to maintain attentional focus, Patch-Prototype Distribution Alignment (PPDA) to preserve part composition, and Region Affinity Matching (RAM) to keep geometric relations consistent across views. These losses act on intermediate feature maps alongside the global objective.
Result: SARL consistently outperforms nine SSL baselines across six downstream tasks. On edge-pose regression (geometry-sensitive task), SARL achieves MAE of 0.3955, a 30% relative improvement over next-best SSL method (0.5682 MAE), approaching supervised upper bound performance.
Conclusion: For fused visual-tactile data, structured spatial equivariance (features varying predictably with object geometry) is the most effective signal, enabling more capable robotic perception. Preserving spatial structure in SSL frameworks is crucial for manipulation tasks.
Abstract: Contact-rich robotic manipulation requires representations that encode local geometry. Vision provides global context but lacks direct measurements of properties such as texture and hardness, whereas touch supplies these cues. Modern visuo-tactile sensors capture both modalities in a single fused image, yielding intrinsically aligned inputs that are well suited to manipulation tasks requiring visual and tactile information. Most self-supervised learning (SSL) frameworks, however, compress feature maps into a global vector, discarding spatial structure and misaligning with the needs of manipulation. To address this, we propose SARL, a spatially-aware SSL framework that augments the Bootstrap Your Own Latent (BYOL) architecture with three map-level objectives, including Saliency Alignment (SAL), Patch-Prototype Distribution Alignment (PPDA), and Region Affinity Matching (RAM), to keep attentional focus, part composition, and geometric relations consistent across views. These losses act on intermediate feature maps, complementing the global objective. SARL consistently outperforms nine SSL baselines across six downstream tasks with fused visual-tactile data. On the geometry-sensitive edge-pose regression task, SARL achieves a Mean Absolute Error (MAE) of 0.3955, a 30% relative improvement over the next-best SSL method (0.5682 MAE) and approaching the supervised upper bound. These findings indicate that, for fused visual-tactile data, the most effective signal is structured spatial equivariance, in which features vary predictably with object geometry, which enables more capable robotic perception.
[371] Med-VCD: Mitigating Hallucination for Medical Large Vision Language Models through Visual Contrastive Decoding
Zahra Mahdavi, Zahra Khodakaramimaghsoud, Hooman Khaloo, Sina Bakhshandeh Taleshani, Erfan Hashemi, Javad Mirzapour Kaleybar, Omid Nejati Manzari
Main category: cs.CV
TL;DR: Med-VCD is a sparse visual-contrastive decoding method that reduces hallucinations in medical LVLMs without slowing inference, improving factual accuracy by 13% and hallucination accuracy by 6% across medical domains.
Details
Motivation: Medical LVLMs suffer from hallucination problems where outputs appear plausible but are incorrect. Existing solutions from natural image domains rely on secondary decoding or rollback procedures that slow inference and may introduce modality misalignment, making them unsuitable for medical applications.Method: Med-VCD uses sparse visual-contrastive decoding with a novel token-sparsification strategy that selects visually informed tokens on the fly. This approach trims redundancy while retaining critical visual context, balancing efficiency with reliability without secondary decoding overhead.
Result: Evaluations across eight medical datasets spanning ophthalmology, radiology, and pathology tasks (visual question answering, report generation, hallucination benchmarks) show Med-VCD improves factual accuracy by average 13% and hallucination accuracy by 6% relative to baseline medical LVLMs.
Conclusion: Med-VCD effectively mitigates hallucinations in medical LVLMs while maintaining inference efficiency, making it suitable for healthcare applications where both accuracy and speed are critical.
Abstract: Large vision-language models (LVLMs) are now central to healthcare applications such as medical visual question answering and imaging report generation. Yet, these models remain vulnerable to hallucination outputs that appear plausible but are in fact incorrect. In the natural image domain, several decoding strategies have been proposed to mitigate hallucinations by reinforcing visual evidence, but most rely on secondary decoding or rollback procedures that substantially slow inference. Moreover, existing solutions are often domain-specific and may introduce misalignment between modalities or between generated and ground-truth content. We introduce Med-VCD, a sparse visual-contrastive decoding method that mitigates hallucinations in medical LVLMs without the time overhead of secondary decoding. Med-VCD incorporates a novel token-sparsification strategy that selects visually informed tokens on the fly, trimming redundancy while retaining critical visual context and thus balancing efficiency with reliability. Evaluations on eight medical datasets, spanning ophthalmology, radiology, and pathology tasks in visual question answering, report generation, and dedicated hallucination benchmarks, show that Med-VCD raises factual accuracy by an average of 13% and improves hallucination accuracy by 6% relative to baseline medical LVLMs.
[372] Physical ID-Transfer Attacks against Multi-Object Tracking via Adversarial Trajectory
Chenyi Wang, Yanmao Man, Raymond Muller, Ming Li, Z. Berkay Celik, Ryan Gerdes, Jonathan Petit
Main category: cs.CV
TL;DR: AdvTraj: First online physical ID-manipulation attack against MOT systems using adversarial trajectories to transfer attacker’s ID to target objects without attacking object detection.
Details
Motivation: Previous MOT attacks were limited - either focused on individual object trackers, or manipulated IDs by attacking object detection in digital domain. These were model-specific, non-robust, and only affected offline datasets. There's a need to explore vulnerabilities in the object association phase of MOT systems.Method: AdvTraj uses adversarial trajectories to transfer attacker’s ID to targeted objects, confusing tracking systems without attacking object detection. The attack works in online and physical settings, with simulation in CARLA. The paper characterizes trajectory patterns and proposes two universal adversarial maneuvers that can be performed by humans in daily scenarios.
Result: White-box attacks against SORT achieve 100% success rate in various scenarios. High transferability (up to 93% success rate) against SOTA MOT algorithms due to common design principles. Identifies patterns in adversarial trajectories and proposes practical adversarial maneuvers.
Conclusion: Reveals under-explored weaknesses in object association phase of SOTA MOT systems. Provides insights for enhancing robustness of tracking systems against such ID-manipulation attacks.
Abstract: Multi-Object Tracking (MOT) is a critical task in computer vision, with applications ranging from surveillance systems to autonomous driving. However, threats to MOT algorithms have yet been widely studied. In particular, incorrect association between the tracked objects and their assigned IDs can lead to severe consequences, such as wrong trajectory predictions. Previous attacks against MOT either focused on hijacking the trackers of individual objects, or manipulating the tracker IDs in MOT by attacking the integrated object detection (OD) module in the digital domain, which are model-specific, non-robust, and only able to affect specific samples in offline datasets. In this paper, we present AdvTraj, the first online and physical ID-manipulation attack against tracking-by-detection MOT, in which an attacker uses adversarial trajectories to transfer its ID to a targeted object to confuse the tracking system, without attacking OD. Our simulation results in CARLA show that AdvTraj can fool ID assignments with 100% success rate in various scenarios for white-box attacks against SORT, which also have high attack transferability (up to 93% attack success rate) against state-of-the-art (SOTA) MOT algorithms due to their common design principles. We characterize the patterns of trajectories generated by AdvTraj and propose two universal adversarial maneuvers that can be performed by a human walker/driver in daily scenarios. Our work reveals under-explored weaknesses in the object association phase of SOTA MOT systems, and provides insights into enhancing the robustness of such systems.
[373] Script: Graph-Structured and Query-Conditioned Semantic Token Pruning for Multimodal Large Language Models
Zhongyu Yang, Dannong Xu, Wei Pang, Yingfang Yuan
Main category: cs.CV
TL;DR: Script is a plug-and-play token pruning method for MLLMs that removes redundant visual tokens without retraining, achieving up to 6.8x speedup and 10x FLOP reduction while maintaining 96.88% of original performance.
Details
Motivation: Multimodal large language models suffer from excessive memory consumption and inference latency due to rapid growth of visual tokens, especially with high-resolution images and videos. Existing token pruning methods ignore query relevance or suffer from attention mechanism limitations.Method: Script uses two modules: 1) graph-structured pruning module that removes visually redundant tokens, and 2) query-conditioned semantic pruning module that preserves query-relevant visual information. It’s plug-and-play, requires no retraining, and generalizes across diverse MLLMs.
Result: Experiments on 14 benchmarks across image and video understanding tasks show Script consistently achieves higher efficiency and accuracy than existing pruning methods. On LLaVA-NeXT-7B, it achieves up to 6.8x prefill speedup, 10x FLOP reduction, while retaining 96.88% of original performance.
Conclusion: Script effectively addresses the computational challenges of MLLMs by combining visual redundancy removal with query-aware semantic preservation, offering a practical solution for efficient multimodal inference without sacrificing performance.
Abstract: The rapid growth of visual tokens in multimodal large language models (MLLMs) leads to excessive memory consumption and inference latency, especially when handling high-resolution images and videos. Token pruning is a technique used to mitigate this issue by removing redundancy, but existing methods often ignore relevance to the user query or suffer from the limitations of attention mechanisms, reducing their adaptability and effectiveness. To address these challenges, we propose Script, a plug-and-play pruning method that requires no retraining and generalizes across diverse MLLMs. Script comprises two modules: a graph-structured pruning module that removes visually redundant tokens, and a query-conditioned semantic pruning module that preserves query-relevant visual information. Together, they enhance performance on multimodal tasks. Experiments on fourteen benchmarks across image and video understanding tasks show that Script consistently achieves higher model efficiency and predictive accuracy compared to existing pruning methods. On LLaVA-NeXT-7B, it achieves up to 6.8x prefill speedup and 10x FLOP reduction, while retaining 96.88% of the original performance.
[374] MoH: Multi-Head Attention as Mixture-of-Head Attention
Peng Jin, Bo Zhu, Li Yuan, Shuicheng Yan
Main category: cs.CV
TL;DR: MoH (Mixture-of-Head attention) upgrades multi-head attention by treating attention heads as experts in a Mixture-of-Experts framework, enabling token-specific head selection and weighted summation for improved efficiency and performance.
Details
Motivation: Multi-head attention is core to Transformers but inefficient - not all attention heads are equally important for all tokens. Current architecture lacks flexibility in head utilization and uses simple summation that may not be optimal.Method: Propose Mixture-of-Head attention (MoH) that: 1) Expresses multi-head attention in summation form, 2) Treats attention heads as experts in MoE framework, 3) Enables tokens to select appropriate heads via routing, 4) Replaces standard summation with weighted summation.
Result: MoH outperforms multi-head attention using only 50%-90% of attention heads across ViT, DiT, and LLMs. MoH-LLaMA3-8B achieves 64.0% average accuracy across 14 benchmarks (2.4% improvement) using only 75% of attention heads. Pre-trained models can be fine-tuned into MoH.
Conclusion: MoH is a promising alternative to multi-head attention that improves efficiency without compromising accuracy, providing flexibility through token-specific head selection and weighted summation, enabling more advanced and efficient attention-based models.
Abstract: In this work, we upgrade the multi-head attention mechanism, the core of the Transformer model, to improve efficiency while maintaining or surpassing the previous accuracy level. We show that multi-head attention can be expressed in the summation form. Drawing on the insight that not all attention heads hold equal significance, we propose Mixture-of-Head attention (MoH), a new architecture that treats attention heads as experts in the Mixture-of-Experts (MoE) mechanism. MoH has two significant advantages: First, MoH enables each token to select the appropriate attention heads, enhancing inference efficiency without compromising accuracy or increasing the number of parameters. Second, MoH replaces the standard summation in multi-head attention with a weighted summation, introducing flexibility to the attention mechanism and unlocking extra performance potential. Extensive experiments on ViT, DiT, and LLMs demonstrate that MoH outperforms multi-head attention by using only 50%-90% of the attention heads. Moreover, we demonstrate that pre-trained multi-head attention models, such as LLaMA3-8B, can be further continue-tuned into our MoH models. Notably, MoH-LLaMA3-8B achieves an average accuracy of 64.0% across 14 benchmarks, outperforming LLaMA3-8B by 2.4% by utilizing only 75% of the attention heads. We believe the proposed MoH is a promising alternative to multi-head attention and provides a strong foundation for developing advanced and efficient attention-based models.
[375] SpriteHand: Real-Time Versatile Hand-Object Interaction with Autoregressive Video Generation
Zisu Li, Hengye Lyu, Jiaxin Shi, Yufeng Zeng, Mingming Fan, Hanwang Zhang, Chen Liang
Main category: cs.CV
TL;DR: SpriteHand is an autoregressive video generation framework for real-time synthesis of diverse hand-object interactions across various object types, using static object images and hand video streams as input.
Details
Motivation: Existing simulation-based approaches rely on rigid object models and pre-scripted hand gestures, making them inadequate for capturing dynamic interactions with non-rigid, articulated, or complex objects like fabrics, elastic materials, furry surfaces, or living creatures.Method: SpriteHand uses an autoregressive video generation framework with causal inference architecture for real-time synthesis. It takes a static object image and hand video stream as input, generates hand-object interaction effects, and employs hybrid post-training to enhance visual realism and temporal coherence.
Result: The 1.3B model achieves real-time streaming generation at ~18 FPS and 640x368 resolution with ~150 ms latency on a single NVIDIA RTX 5090 GPU, supporting over a minute of continuous output. It demonstrates superior visual quality, physical plausibility, and interaction fidelity compared to generative and engine-based baselines.
Conclusion: SpriteHand provides an effective solution for synthesizing versatile hand-object interactions across diverse object types, overcoming limitations of conventional simulation approaches and enabling real-time generation of physically plausible interactions.
Abstract: Modeling and synthesizing complex hand-object interactions remains a significant challenge, even for state-of-the-art physics engines. Conventional simulation-based approaches rely on explicitly defined rigid object models and pre-scripted hand gestures, making them inadequate for capturing dynamic interactions with non-rigid or articulated entities such as deformable fabrics, elastic materials, hinge-based structures, furry surfaces, or even living creatures. In this paper, we present SpriteHand, an autoregressive video generation framework for real-time synthesis of versatile hand-object interaction videos across a wide range of object types and motion patterns. SpriteHand takes as input a static object image and a video stream in which the hands are imagined to interact with the virtual object embedded in a real-world scene, and generates corresponding hand-object interaction effects in real time. Our model employs a causal inference architecture for autoregressive generation and leverages a hybrid post-training approach to enhance visual realism and temporal coherence. Our 1.3B model supports real-time streaming generation at around 18 FPS and 640x368 resolution, with an approximate 150 ms latency on a single NVIDIA RTX 5090 GPU, and more than a minute of continuous output. Experiments demonstrate superior visual quality, physical plausibility, and interaction fidelity compared to both generative and engine-based baselines.
[376] SGDiff: Scene Graph Guided Diffusion Model for Image Collaborative SegCaptioning
Xu Zhang, Jin Yuan, Hanwang Zhang, Guojin Zhong, Yongsheng Zang, Jiacheng Lin, Zhiyong Li
Main category: cs.CV
TL;DR: SegCaptioning is a new task that generates diverse (caption, mask) pairs from simple prompts like bounding boxes, using a Scene Graph Guided Diffusion Model to produce correlated semantic interpretations.
Details
Motivation: Current controllable image understanding tasks require expensive prompt inputs (e.g., detailed text) and produce limited single outputs. There's a need for a system that accepts minimal prompts but generates diverse, flexible semantic interpretations for user selection.Method: Proposes Scene Graph Guided Diffusion Model (SGDiff) with: 1) Prompt-Centric Scene Graph Adaptor to map user prompts to scene graphs capturing intention; 2) Diffusion process with Scene Graph Guided Bimodal Transformer to predict correlated caption-mask pairs; 3) Multi-Entities Contrastive Learning loss for explicit visual-textual alignment.
Result: Extensive experiments on two datasets show SGDiff achieves superior performance in SegCaptioning, producing promising results for both captioning and segmentation tasks with minimal prompt input.
Conclusion: SegCaptioning enables flexible semantic interpretation from simple prompts, and SGDiff effectively addresses the challenges of capturing user intention while generating correlated, well-aligned caption-mask pairs through structured scene graph guidance.
Abstract: Controllable image semantic understanding tasks, such as captioning or segmentation, necessitate users to input a prompt (e.g., text or bounding boxes) to predict a unique outcome, presenting challenges such as high-cost prompt input or limited information output. This paper introduces a new task ``Image Collaborative Segmentation and Captioning’’ (SegCaptioning), which aims to translate a straightforward prompt, like a bounding box around an object, into diverse semantic interpretations represented by (caption, masks) pairs, allowing flexible result selection by users. This task poses significant challenges, including accurately capturing a user’s intention from a minimal prompt while simultaneously predicting multiple semantically aligned caption words and masks. Technically, we propose a novel Scene Graph Guided Diffusion Model that leverages structured scene graph features for correlated mask-caption prediction. Initially, we introduce a Prompt-Centric Scene Graph Adaptor to map a user’s prompt to a scene graph, effectively capturing his intention. Subsequently, we employ a diffusion process incorporating a Scene Graph Guided Bimodal Transformer to predict correlated caption-mask pairs by uncovering intricate correlations between them. To ensure accurate alignment, we design a Multi-Entities Contrastive Learning loss to explicitly align visual and textual entities by considering inter-modal similarity, resulting in well-aligned caption-mask pairs. Extensive experiments conducted on two datasets demonstrate that SGDiff achieves superior performance in SegCaptioning, yielding promising results for both captioning and segmentation tasks with minimal prompt input.
[377] Adversarial Exploitation of Data Diversity Improves Visual Localization
Sihang Li, Siqi Tan, Bowen Chang, Jing Zhang, Chen Feng, Yiming Li
Main category: cs.CV
TL;DR: RAP introduces appearance diversity as key to robust visual localization, using 3D Gaussian Splats to generate appearance-varied training data and a two-branch adversarial training pipeline to bridge synthetic-to-real gap.
Details
Motivation: Absolute pose regression (APR) methods for visual localization struggle with generalization despite viewpoint data augmentation. The paper identifies that appearance variation (lighting, weather, etc.) is the critical missing factor for robust localization, especially in dynamic real-world scenarios.Method: 1) Lift real 2D images into 3D Gaussian Splats with varying appearance and deblurring ability to synthesize diverse training data with pose AND appearance variations. 2) Build a two-branch joint training pipeline with an adversarial discriminator to bridge the synthetic-to-real gap and fully leverage appearance-diverse data.
Result: Significantly outperforms state-of-the-art methods: reduces translation errors by 50% and rotation errors by 41% on indoor datasets, and 38% and 44% on outdoor datasets. Shows remarkable robustness in dynamic driving scenarios under varying weather conditions and day-to-night scenarios where previous APR methods fail.
Conclusion: Appearance variation is crucial for robust visual localization. The proposed approach effectively addresses generalization limitations of APR methods by incorporating appearance diversity through 3D Gaussian Splat synthesis and adversarial training, achieving superior performance across diverse real-world conditions.
Abstract: Visual localization, which estimates a camera’s pose within a known scene, is a fundamental capability for autonomous systems. While absolute pose regression (APR) methods have shown promise for efficient inference, they often struggle with generalization. Recent approaches attempt to address this through data augmentation with varied viewpoints, yet they overlook a critical factor: appearance diversity. In this work, we identify appearance variation as the key to robust localization. Specifically, we first lift real 2D images into 3D Gaussian Splats with varying appearance and deblurring ability, enabling the synthesis of diverse training data that varies not just in poses but also in environmental conditions such as lighting and weather. To fully unleash the potential of the appearance-diverse data, we build a two-branch joint training pipeline with an adversarial discriminator to bridge the syn-to-real gap. Extensive experiments demonstrate that our approach significantly outperforms state-of-the-art methods, reducing translation and rotation errors by 50% and 41% on indoor datasets, and 38% and 44% on outdoor datasets. Most notably, our method shows remarkable robustness in dynamic driving scenarios under varying weather conditions and in day-to-night scenarios, where previous APR methods fail. Project Page: https://ai4ce.github.io/RAP/
[378] Artemis: Structured Visual Reasoning for Perception Policy Learning
Wei Tang, Yanpeng Sun, Shan Zhang, Xiaofan Li, Piotr Koniusz, Wei Li, Na Zhao, Zechao Li
Main category: cs.CV
TL;DR: Artemis is a perception-policy learning framework that replaces ambiguous natural language reasoning with structured proposal-based reasoning using (label, bounding-box) pairs, improving performance on visual perception tasks.
Details
Motivation: Current reinforcement learning frameworks use natural language intermediate reasoning which reduces performance on perception tasks because language reasoning happens in unstructured semantic space while visual perception requires spatial and object-centric reasoning.Method: Artemis performs structured proposal-based reasoning where each intermediate step is represented as a (label, bounding-box) pair capturing verifiable visual states, enabling explicit state tracking, direct supervision for proposal quality, and avoiding language ambiguity.
Result: Artemis, built on Qwen2.5-VL-3B, achieves strong performance on grounding and detection tasks, shows substantial generalization to counting and geometric-perception tasks, and achieves competitive performance on general MLLM benchmarks.
Conclusion: Aligning reasoning with spatial representations enhances perception-policy learning, and spatially grounded reasoning provides a principled route toward scalable and general perception policies.
Abstract: Recent reinforcement-learning frameworks for visual perception policy have begun to incorporate intermediate reasoning chains expressed in natural language. Empirical observations indicate that such purely linguistic intermediate reasoning often reduces performance on perception tasks. We argue that the core issue lies not in reasoning per se but in the form of reasoning: while these chains perform semantic reasoning in an unstructured linguistic space, visual perception requires reasoning in a spatial and object-centric space. In response, we introduce Artemis, a perception-policy learning framework that performs structured proposal-based reasoning, where each intermediate step is represented as a (label, bounding-box) pair capturing a verifiable visual state. This design enables explicit tracking of intermediate states, direct supervision for proposal quality, and avoids ambiguity introduced by language-based reasoning. Artemis is built on Qwen2.5-VL-3B, achieves strong performance on grounding and detection task and exhibits substantial generalization to counting and geometric-perception tasks. The consistent improvements across these diverse settings confirm that aligning reasoning with spatial representations enhances perception-policy learning. Owing to its strengthened visual reasoning, Artemis also achieves competitive performance on general MLLM benchmarks, illustrating that spatially grounded reasoning provides a principled route toward scalable and general perception policies.
[379] Blind Inverse Problem Solving Made Easy by Text-to-Image Latent Diffusion
Michail Dontas, Yutong He, Naoki Murata, Yuki Mitsufuji, J. Zico Kolter, Ruslan Salakhutdinov
Main category: cs.CV
TL;DR: LADiBI is a training-free method that uses text-to-image diffusion models to solve blind inverse image restoration problems without needing to know the degradation operator or make restrictive assumptions about linearity or image distributions.
Details
Motivation: Existing blind inverse image restoration methods have impractical limitations: they require assumptions about operator linearity, need curated training data, or only work with narrow image distributions. There's a need for more flexible solutions that can handle diverse real-world scenarios.Method: LADiBI uses large-scale text-to-image diffusion models within a Bayesian framework. It employs text prompts to encode priors for both target images and degradation operators. The method features a novel diffusion posterior sampling algorithm that combines strategic operator initialization with iterative refinement of both image and operator parameters.
Result: Experiments demonstrate that LADiBI effectively handles both linear and challenging nonlinear image restoration problems across various image distributions, all without task-specific assumptions or retraining.
Conclusion: LADiBI represents a significant advancement in blind inverse image restoration by eliminating restrictive assumptions, offering unprecedented flexibility through text-guided priors, and working without training or specific operator constraints.
Abstract: This paper considers blind inverse image restoration, the task of predicting a target image from a degraded source when the degradation (i.e. the forward operator) is unknown. Existing solutions typically rely on restrictive assumptions such as operator linearity, curated training data or narrow image distributions limiting their practicality. We introduce LADiBI, a training-free method leveraging large-scale text-to-image diffusion to solve diverse blind inverse problems with minimal assumptions. Within a Bayesian framework, LADiBI uses text prompts to jointly encode priors for both target images and operators, unlocking unprecedented flexibility compared to existing methods. Additionally, we propose a novel diffusion posterior sampling algorithm that combines strategic operator initialization with iterative refinement of image and operator parameters, eliminating the need for highly constrained operator forms. Experiments show that LADiBI effectively handles both linear and challenging nonlinear image restoration problems across various image distributions, all without task-specific assumptions or retraining.
[380] PAI-Bench: A Comprehensive Benchmark For Physical AI
Fengzhe Zhou, Jiannan Huang, Jialuo Li, Deva Ramanan, Humphrey Shi
Main category: cs.CV
TL;DR: PAI-Bench is a comprehensive benchmark with 2,808 real-world cases that evaluates video generation and understanding models on physical plausibility and reasoning, revealing current models struggle with physically coherent dynamics and causal interpretation.
Details
Motivation: Current multi-modal large language models and video generative models lack sufficient understanding of their capabilities in perceiving and predicting real-world physical dynamics, necessitating a unified benchmark to systematically assess these abilities.Method: Introduces PAI-Bench, a unified benchmark evaluating perception and prediction capabilities across three areas: video generation, conditional video generation, and video understanding, using 2,808 real-world cases with task-aligned metrics focused on physical plausibility and domain-specific reasoning.
Result: Video generative models show strong visual fidelity but struggle with physically coherent dynamics, while multi-modal large language models have limited performance in forecasting and causal interpretation, indicating current systems are at early stages for Physical AI demands.
Conclusion: PAI-Bench establishes a realistic foundation for evaluating Physical AI and highlights key gaps in physical coherence and causal reasoning that future systems must address to advance physical intelligence capabilities.
Abstract: Physical AI aims to develop models that can perceive and predict real-world dynamics; yet, the extent to which current multi-modal large language models and video generative models support these abilities is insufficiently understood. We introduce Physical AI Bench (PAI-Bench), a unified and comprehensive benchmark that evaluates perception and prediction capabilities across video generation, conditional video generation, and video understanding, comprising 2,808 real-world cases with task-aligned metrics designed to capture physical plausibility and domain-specific reasoning. Our study provides a systematic assessment of recent models and shows that video generative models, despite strong visual fidelity, often struggle to maintain physically coherent dynamics, while multi-modal large language models exhibit limited performance in forecasting and causal interpretation. These observations suggest that current systems are still at an early stage in handling the perceptual and predictive demands of Physical AI. In summary, PAI-Bench establishes a realistic foundation for evaluating Physical AI and highlights key gaps that future systems must address.
[381] PRIMA: Multi-Image Vision-Language Models for Reasoning Segmentation
Muntasir Wahed, Kiet A. Nguyen, Adheesh Sunil Juvekar, Xinzhuo Li, Xiaona Zhou, Vedant Shah, Tianjiao Yu, Pinar Yanardag, Ismini Lourentzou
Main category: cs.CV
TL;DR: PRIMA introduces multi-image pixel-grounded reasoning, combining pixel-level grounding with multi-image understanding using SQuARE vision module and M4SEG benchmark.
Details
Motivation: Existing LVLMs have limitations: pixel-grounding models work only with single images, while multi-image models lack pixel-level grounding. This gap prevents fine-grained comparisons across multiple images.Method: PRIMA integrates pixel-level grounding with multi-image reasoning using SQuARE vision module, which injects cross-image relational context into compact query-based visual tokens before fusing with language backbone. Trained on M4SEG benchmark.
Result: PRIMA outperforms state-of-the-art baselines with 7.83% improvement in Recall and 11.25% improvement in S-IoU. Ablation studies confirm SQuARE’s effectiveness in capturing cross-image relationships.
Conclusion: PRIMA successfully bridges the gap between pixel-level grounding and multi-image reasoning, enabling detailed, fine-grained comparisons across multiple images with pixel-grounded explanations.
Abstract: Despite significant advancements in Large Vision-Language Models (LVLMs)’ capabilities, existing pixel-grounding models operate in single-image settings, limiting their ability to perform detailed, fine-grained comparisons across multiple images. Conversely, current multi-image understanding models lack pixel-level grounding. Our work addresses this gap by introducing the task of multi-image pixel-grounded reasoning alongside PRIMA, an LVLM that integrates pixel-level grounding with robust multi-image reasoning to produce contextually rich, pixel-grounded explanations. Central to PRIMA is SQuARE, a vision module that injects cross-image relational context into compact query-based visual tokens before fusing them with the language backbone. To support training and evaluation, we curate M4SEG, a new multi-image reasoning segmentation benchmark consisting of $\sim$744K question-answer pairs that require fine-grained visual understanding across multiple images. PRIMA outperforms state-of-the-art baselines with $7.83%$ and $11.25%$ improvements in Recall and S-IoU, respectively. Ablation studies further demonstrate the effectiveness of the proposed SQuARE module in capturing cross-image relationships.
[382] Learning Visual Affordance from Audio
Lidong Lu, Guo Chen, Zhu Wei, Yicheng Liu, Tong Lu
Main category: cs.CV
TL;DR: Audio-Visual Affordance Grounding (AV-AG) segments object interaction regions from action sounds using audio-visual fusion, outperforming text/video-based methods.
Details
Motivation: Existing affordance grounding methods rely on textual instructions or demonstration videos, which suffer from ambiguity or occlusion. Audio provides real-time, semantically rich, and visually independent cues that enable more intuitive understanding of interaction regions.Method: Propose AVAGFormer with semantic-conditioned cross-modal mixer and dual-head decoder to effectively fuse audio and visual signals for mask prediction. Construct first AV-AG dataset with action sounds, object images, and pixel-level affordance annotations including unseen subset for zero-shot evaluation.
Result: AVAGFormer achieves state-of-the-art performance on AV-AG, surpassing baselines from related tasks. Comprehensive analyses show distinctions between AV-AG and AVS, benefits of end-to-end modeling, and contributions of each component.
Conclusion: Audio provides valuable cues for affordance grounding, and the proposed AVAGFormer effectively leverages audio-visual fusion for interaction region segmentation. The released dataset and code support further research in this direction.
Abstract: We introduce Audio-Visual Affordance Grounding (AV-AG), a new task that segments object interaction regions from action sounds. Unlike existing approaches that rely on textual instructions or demonstration videos, which often limited by ambiguity or occlusion, audio provides real-time, semantically rich, and visually independent cues for affordance grounding, enabling more intuitive understanding of interaction regions. To support this task, we construct the first AV-AG dataset, comprising a large collection of action sounds, object images, and pixel-level affordance annotations. The dataset also includes an unseen subset to evaluate zero-shot generalization. Furthermore, we propose AVAGFormer, a model equipped with a semantic-conditioned cross-modal mixer and a dual-head decoder that effectively fuses audio and visual signals for mask prediction. Experiments show that AVAGFormer achieves state-of-the-art performance on AV-AG, surpassing baselines from related tasks. Comprehensive analyses highlight the distinctions between AV-AG and AVS, the benefits of end-to-end modeling, and the contribution of each component. Code and dataset have been released on https://jscslld.github.io/AVAGFormer/.
[383] MV-TAP: Tracking Any Point in Multi-View Videos
Jahyeok Koo, Inès Hyeonsu Kim, Mungyeom Kim, Junghyun Park, Seohyun Park, Jaeyeong Kim, Jung Yi, Seokju Cho, Seungryong Kim
Main category: cs.CV
TL;DR: MV-TAP is a novel point tracker for multi-view videos that uses camera geometry and cross-view attention to track points across multiple camera views of dynamic scenes.
Details
Motivation: Multi-view camera systems provide rich observations of complex scenes, but understanding dynamic objects in multi-view settings is challenging and crucial for various applications. Existing methods need better ways to leverage cross-view information for more complete and reliable point tracking.Method: MV-TAP uses camera geometry and a cross-view attention mechanism to aggregate spatio-temporal information across different views. The approach is supported by a large-scale synthetic training dataset and real-world evaluation sets specifically designed for multi-view tracking.
Result: Extensive experiments show that MV-TAP outperforms existing point-tracking methods on challenging benchmarks, establishing an effective baseline for advancing research in multi-view point tracking.
Conclusion: MV-TAP provides a novel and effective solution for point tracking in multi-view videos, demonstrating the value of leveraging cross-view information through attention mechanisms and geometric constraints for improved trajectory estimation in dynamic scenes.
Abstract: Multi-view camera systems enable rich observations of complex real-world scenes, and understanding dynamic objects in multi-view settings has become central to various applications. In this work, we present MV-TAP, a novel point tracker that tracks points across multi-view videos of dynamic scenes by leveraging cross-view information. MV-TAP utilizes camera geometry and a cross-view attention mechanism to aggregate spatio-temporal information across views, enabling more complete and reliable trajectory estimation in multi-view videos. To support this task, we construct a large-scale synthetic training dataset and real-world evaluation sets tailored for multi-view tracking. Extensive experiments demonstrate that MV-TAP outperforms existing point-tracking methods on challenging benchmarks, establishing an effective baseline for advancing research in multi-view point tracking.
[384] Towards Balanced Multi-Modal Learning in 3D Human Pose Estimation
Mengshi Qi, Jiaxuan Peng, Xianlin Zhang, Huadong Ma
Main category: cs.CV
TL;DR: A balanced multi-modal learning method for 3D human pose estimation using RGB, LiDAR, mmWave, and WiFi with Shapley value-based contribution assessment and modality learning regulation to address modality imbalance.
Details
Motivation: RGB-based 3D HPE faces occlusion and privacy issues, while multi-modal approaches suffer from modality imbalance problems that need to be addressed for better performance.Method: Proposes a Shapley value-based algorithm to assess modality contributions and detect imbalance, plus a modality learning regulation strategy that slows early-stage training to balance learning across modalities.
Result: Extensive experiments on MM-Fi dataset show superior performance in enhancing 3D pose estimation under complex conditions compared to existing methods.
Conclusion: The balanced multi-modal learning approach effectively addresses modality imbalance and improves 3D human pose estimation performance across multiple sensing modalities.
Abstract: 3D human pose estimation (3D HPE) has emerged as a prominent research topic, particularly in the realm of RGB-based methods. However, the use of RGB images is often limited by issues such as occlusion and privacy constraints. Consequently, multi-modal sensing, which leverages non-intrusive sensors, is gaining increasing attention. Nevertheless, multi-modal 3D HPE still faces challenges, including modality imbalance. In this work, we introduce a novel balanced multi-modal learning method for 3D HPE, which harnesses the power of RGB, LiDAR, mmWave, and WiFi. Specifically, we propose a Shapley value-based contribution algorithm to assess the contribution of each modality and detect modality imbalance. To address this imbalance, we design a modality learning regulation strategy that decelerates the learning process during the early stages of training. We conduct extensive experiments on the widely adopted multi-modal dataset, MM-Fi, demonstrating the superiority of our approach in enhancing 3D pose estimation under complex conditions. We will release our codes soon.
[385] AirSim360: A Panoramic Simulation Platform within Drone View
Xian Ge, Yuling Pan, Yuhang Zhang, Xiang Li, Weijun Zhang, Dizhe Zhang, Zhaoliang Wan, Xin Lin, Xiangkai Zhang, Juntao Liang, Jason Li, Wenjie Jiang, Bo Du, Ming-Hsuan Yang, Lu Qi
Main category: cs.CV
TL;DR: AirSim360 is a simulation platform for omnidirectional aerial data collection using drones, featuring render-aligned labeling, pedestrian-aware systems, and automated trajectory generation for navigation tasks.
Details
Motivation: The field of 360-degree omnidirectional understanding lacks large-scale and diverse data, which limits advancement in spatial intelligence. Existing simulators don't systematically model the 4D real world under omnidirectional settings.Method: 1) Render-aligned data and labeling paradigm for pixel-level geometric, semantic, and entity-level understanding; 2) Interactive pedestrian-aware system for modeling human behavior; 3) Automated trajectory generation paradigm to support navigation tasks.
Result: Collected more than 60K panoramic samples and conducted extensive experiments across various tasks demonstrating the simulator’s effectiveness. The platform is the first to systematically model the 4D real world under omnidirectional setting.
Conclusion: AirSim360 addresses the data scarcity problem in omnidirectional understanding by providing a comprehensive simulation platform with diverse aerial data collection capabilities, which will be made publicly available to advance spatial intelligence research.
Abstract: The field of 360-degree omnidirectional understanding has been receiving increasing attention for advancing spatial intelligence. However, the lack of large-scale and diverse data remains a major limitation. In this work, we propose AirSim360, a simulation platform for omnidirectional data from aerial viewpoints, enabling wide-ranging scene sampling with drones. Specifically, AirSim360 focuses on three key aspects: a render-aligned data and labeling paradigm for pixel-level geometric, semantic, and entity-level understanding; an interactive pedestrian-aware system for modeling human behavior; and an automated trajectory generation paradigm to support navigation tasks. Furthermore, we collect more than 60K panoramic samples and conduct extensive experiments across various tasks to demonstrate the effectiveness of our simulator. Unlike existing simulators, our work is the first to systematically model the 4D real world under an omnidirectional setting. The entire platform, including the toolkit, plugins, and collected datasets, will be made publicly available at https://insta360-research-team.github.io/AirSim360-website.
[386] Improved Mean Flows: On the Challenges of Fastforward Generative Models
Zhengyang Geng, Yiyang Lu, Zongze Wu, Eli Shechtman, J. Zico Kolter, Kaiming He
Main category: cs.CV
TL;DR: Improved MeanFlow (iMF) enhances one-step generative modeling with better training stability and flexible guidance, achieving 1.72 FID on ImageNet 256×256 with single function evaluation.
Details
Motivation: Original MeanFlow has two key challenges: 1) training objective depends on the network itself causing instability, and 2) fixed classifier-free guidance scale during training sacrifices flexibility at test time.Method: Reformulate training objective as loss on instantaneous velocity parameterized by network predicting average velocity. Formulate guidance as explicit conditioning variables processed through in-context conditioning. Train entirely from scratch.
Result: Achieves 1.72 FID on ImageNet 256×256 with single function evaluation (1-NFE), substantially outperforming prior one-step methods and closing gap with multi-step methods without distillation.
Conclusion: Improved MeanFlow (iMF) advances fastforward generative modeling as a standalone paradigm with better training stability and flexible guidance mechanisms.
Abstract: MeanFlow (MF) has recently been established as a framework for one-step generative modeling. However, its ``fastforward’’ nature introduces key challenges in both the training objective and the guidance mechanism. First, the original MF’s training target depends not only on the underlying ground-truth fields but also on the network itself. To address this issue, we recast the objective as a loss on the instantaneous velocity $v$, re-parameterized by a network that predicts the average velocity $u$. Our reformulation yields a more standard regression problem and improves the training stability. Second, the original MF fixes the classifier-free guidance scale during training, which sacrifices flexibility. We tackle this issue by formulating guidance as explicit conditioning variables, thereby retaining flexibility at test time. The diverse conditions are processed through in-context conditioning, which reduces model size and benefits performance. Overall, our $\textbf{improved MeanFlow}$ ($\textbf{iMF}$) method, trained entirely from scratch, achieves $\textbf{1.72}$ FID with a single function evaluation (1-NFE) on ImageNet 256$\times$256. iMF substantially outperforms prior methods of this kind and closes the gap with multi-step methods while using no distillation. We hope our work will further advance fastforward generative modeling as a stand-alone paradigm.
[387] TUNA: Taming Unified Visual Representations for Native Unified Multimodal Models
Zhiheng Liu, Weiming Ren, Haozhe Liu, Zijian Zhou, Shoufa Chen, Haonan Qiu, Xiaoke Huang, Zhaochong An, Fanny Yang, Aditya Patel, Viktar Atliha, Tony Ng, Xiao Han, Chuyan Zhu, Chenyang Zhang, Ding Liu, Juan-Manuel Perez-Rua, Sen He, Jürgen Schmidhuber, Wenhu Chen, Ping Luo, Wei Liu, Tao Xiang, Jonas Schult, Yuren Cong
Main category: cs.CV
TL;DR: TUNA is a unified multimodal model that creates a single continuous visual representation space for both understanding and generation tasks, outperforming decoupled approaches and achieving SOTA across image/video understanding and generation.
Details
Motivation: Prior unified multimodal models use decoupled representations that cause format mismatches between understanding and generation tasks. The authors aim to create a truly unified framework that avoids these issues and allows both tasks to benefit from each other.Method: TUNA builds a unified continuous visual representation by cascading a VAE encoder with a representation encoder. This creates a single representation space that enables end-to-end processing of images and videos for both understanding and generation tasks.
Result: TUNA achieves state-of-the-art results across multiple benchmarks: image/video understanding, image/video generation, and image editing. The unified representation outperforms decoupled alternatives, and stronger pretrained representation encoders consistently improve performance.
Conclusion: The unified representation design is effective and scalable. Joint training on understanding and generation data allows the tasks to benefit from each other rather than interfere, demonstrating the advantages of truly unified multimodal modeling.
Abstract: Unified multimodal models (UMMs) aim to jointly perform multimodal understanding and generation within a single framework. We present TUNA, a native UMM that builds a unified continuous visual representation by cascading a VAE encoder with a representation encoder. This unified representation space allows end-to-end processing of images and videos for both understanding and generation tasks. Compared to prior UMMs with decoupled representations, TUNA’s unified visual space avoids representation format mismatches introduced by separate encoders, outperforming decoupled alternatives in both understanding and generation. Moreover, we observe that stronger pretrained representation encoders consistently yield better performance across all multimodal tasks, highlighting the importance of the representation encoder. Finally, in this unified setting, jointly training on both understanding and generation data allows the two tasks to benefit from each other rather than interfere. Our extensive experiments on multimodal understanding and generation benchmarks show that TUNA achieves state-of-the-art results in image and video understanding, image and video generation, and image editing, demonstrating the effectiveness and scalability of its unified representation design.
[388] Generative Video Motion Editing with 3D Point Tracks
Yao-Chih Lee, Zhoutong Zhang, Jiahui Huang, Jui-Hsien Wang, Joon-Young Lee, Jia-Bin Huang, Eli Shechtman, Zhengqi Li
Main category: cs.CV
TL;DR: Track-conditioned video-to-video framework for joint editing of camera and object motion using 3D point tracks to transfer context while preserving coherence.
Details
Motivation: Current motion editing methods lack full-scene context for consistent video editing, with I2V approaches missing scene context and V2V methods offering limited fine-grained object motion control.Method: Track-conditioned V2V framework that conditions video generation on source video and paired 3D point tracks representing source/target motions, using 3D tracks for depth cues and occlusion handling, trained in two stages on synthetic and real data.
Result: Enables diverse motion edits including joint camera/object manipulation, motion transfer, and non-rigid deformation with precise motion control and spatiotemporal coherence.
Conclusion: The framework unlocks new creative potential in video editing by providing precise control over both camera and object motions using 3D track conditioning.
Abstract: Camera and object motions are central to a video’s narrative. However, precisely editing these captured motions remains a significant challenge, especially under complex object movements. Current motion-controlled image-to-video (I2V) approaches often lack full-scene context for consistent video editing, while video-to-video (V2V) methods provide viewpoint changes or basic object translation, but offer limited control over fine-grained object motion. We present a track-conditioned V2V framework that enables joint editing of camera and object motion. We achieve this by conditioning a video generation model on a source video and paired 3D point tracks representing source and target motions. These 3D tracks establish sparse correspondences that transfer rich context from the source video to new motions while preserving spatiotemporal coherence. Crucially, compared to 2D tracks, 3D tracks provide explicit depth cues, allowing the model to resolve depth order and handle occlusions for precise motion editing. Trained in two stages on synthetic and real data, our model supports diverse motion edits, including joint camera/object manipulation, motion transfer, and non-rigid deformation, unlocking new creative potential in video editing.
[389] Objects in Generated Videos Are Slower Than They Appear: Models Suffer Sub-Earth Gravity and Don’t Know Galileo’s Principle…for now
Varun Varma Thozhiyoor, Shivam Tripathi, Venkatesh Babu Radhakrishnan, Anand Bhattad
Main category: cs.CV
TL;DR: Video generators fail to accurately represent gravity, showing objects falling too slowly, but targeted fine-tuning with minimal data can partially correct this physical gap.
Details
Motivation: To evaluate video generators as potential world models by testing their understanding of fundamental physical laws like gravity, and to determine if observed physical errors are artifacts of ambiguous metric scale or genuine representation failures.Method: 1) Test out-of-the-box video generators on falling objects; 2) Investigate if errors are due to ambiguous metric scale (e.g., frame rate assumptions); 3) Introduce unit-free, two-object protocol testing timing ratio t₁²/t₂² = h₁/h₂ independent of g, focal length, and scale; 4) Use lightweight low-rank adaptor fine-tuned on only 100 single-ball clips to correct physical representation.
Result: 1) Video generators consistently generate objects falling at effectively slower acceleration (g_eff = 1.81 m/s² vs. Earth’s 9.8 m/s²); 2) Temporal rescaling cannot correct high-variance gravity artifacts; 3) Relative tests reveal violations of Galileo’s equivalence principle; 4) Fine-tuning raises g_eff to 6.43 m/s² (65% of terrestrial gravity); 5) Specialist adaptor generalizes zero-shot to two-ball drops and inclined planes.
Conclusion: Video generators have significant gaps in representing fundamental physical laws like gravity, but these can be partially corrected with targeted specialization using minimal data, suggesting specific physical laws can be learned and corrected in video generation models.
Abstract: Video generators are increasingly evaluated as potential world models, which requires them to encode and understand physical laws. We investigate their representation of a fundamental law: gravity. Out-of-the-box video generators consistently generate objects falling at an effectively slower acceleration. However, these physical tests are often confounded by ambiguous metric scale. We first investigate if observed physical errors are artifacts of these ambiguities (e.g., incorrect frame rate assumptions). We find that even temporal rescaling cannot correct the high-variance gravity artifacts. To rigorously isolate the underlying physical representation from these confounds, we introduce a unit-free, two-object protocol that tests the timing ratio $t_1^2/t_2^2 = h_1/h_2$, a relationship independent of $g$, focal length, and scale. This relative test reveals violations of Galileo’s equivalence principle. We then demonstrate that this physical gap can be partially mitigated with targeted specialization. A lightweight low-rank adaptor fine-tuned on only 100 single-ball clips raises $g_{\mathrm{eff}}$ from $1.81,\mathrm{m/s^2}$ to $6.43,\mathrm{m/s^2}$ (reaching $65%$ of terrestrial gravity). This specialist adaptor also generalizes zero-shot to two-ball drops and inclined planes, offering initial evidence that specific physical laws can be corrected with minimal data.
[390] Data-Centric Visual Development for Self-Driving Labs
Anbang Liu, Guanzhong Hu, Jiayi Wang, Ping Guo, Han Liu
Main category: cs.CV
TL;DR: Hybrid pipeline combining real and virtual data generation enables robust bubble detection for pipetting in self-driving labs, achieving 99.6% accuracy with minimal human effort.
Details
Motivation: Self-driving labs require robust models trained on large annotated datasets, but such data (especially negative samples) is difficult to obtain in practice, particularly for precision-sensitive pipetting actions.Method: Two-track approach: 1) Real track uses human-in-the-loop scheme with automated acquisition and selective human verification; 2) Virtual track augments real data using reference-conditioned, prompt-guided image generation with screening and validation. Combined tracks create class-balanced dataset for bubble detection training.
Result: Model trained on automatically acquired real images achieves 99.6% accuracy on held-out test set. Mixing real and generated data maintains 99.4% accuracy while reducing collection and review workload.
Conclusion: The approach provides scalable, cost-effective strategy for supplying visual feedback data to SDL workflows and offers practical solution to data scarcity in rare event detection and broader vision tasks.
Abstract: Self-driving laboratories offer a promising path toward reducing the labor-intensive, time-consuming, and often irreproducible workflows in the biological sciences. Yet their stringent precision requirements demand highly robust models whose training relies on large amounts of annotated data. However, this kind of data is difficult to obtain in routine practice, especially negative samples. In this work, we focus on pipetting, the most critical and precision sensitive action in SDLs. To overcome the scarcity of training data, we build a hybrid pipeline that fuses real and virtual data generation. The real track adopts a human-in-the-loop scheme that couples automated acquisition with selective human verification to maximize accuracy with minimal effort. The virtual track augments the real data using reference-conditioned, prompt-guided image generation, which is further screened and validated for reliability. Together, these two tracks yield a class-balanced dataset that enables robust bubble detection training. On a held-out real test set, a model trained entirely on automatically acquired real images reaches 99.6% accuracy, and mixing real and generated data during training sustains 99.4% accuracy while reducing collection and review load. Our approach offers a scalable and cost-effective strategy for supplying visual feedback data to SDL workflows and provides a practical solution to data scarcity in rare event detection and broader vision tasks.
[391] GARF: Learning Generalizable 3D Reassembly for Real-World Fractures
Sihang Li, Zeyu Jiang, Grace Chen, Chenyang Xu, Siqi Tan, Xue Wang, Irving Fang, Kristof Zyskowski, Shannon P. McPherron, Radu Iovita, Chen Feng, Jing Zhang
Main category: cs.CV
TL;DR: GARF is a generalizable 3D reassembly framework for real-world fractures that uses fracture-aware pretraining and flow matching for precise fragment alignment, achieving state-of-the-art performance on both synthetic and real-world datasets.
Details
Motivation: Existing learning-based 3D reassembly methods trained on synthetic datasets have limited generalizability to real-world fractures where breakage patterns are more complex. There's a need for models that can handle diverse real-world fracture types across different materials.Method: GARF uses fracture-aware pretraining to learn fracture features from individual fragments, employs flow matching for precise 6-DoF alignments, and introduces one-step preassembly at inference time for robustness to unseen objects and varying numbers of fractures.
Result: The framework achieves 82.87% lower rotation error and 25.15% higher part accuracy compared to state-of-the-art methods. It demonstrates strong generalization across unseen object shapes and diverse fracture types in real-world materials like ceramics, bones, eggshells, and lithics.
Conclusion: GARF successfully bridges the gap between synthetic training and real-world application, showing that training on synthetic data can advance real-world 3D puzzle solving with strong generalization capabilities across diverse fracture types and object shapes.
Abstract: 3D reassembly is a challenging spatial intelligence task with broad applications across scientific domains. While large-scale synthetic datasets have fueled promising learning-based approaches, their generalizability to different domains is limited. Critically, it remains uncertain whether models trained on synthetic datasets can generalize to real-world fractures where breakage patterns are more complex. To bridge this gap, we propose GARF, a generalizable 3D reassembly framework for real-world fractures. GARF leverages fracture-aware pretraining to learn fracture features from individual fragments, with flow matching enabling precise 6-DoF alignments. At inference time, we introduce one-step preassembly, improving robustness to unseen objects and varying numbers of fractures. In collaboration with archaeologists, paleoanthropologists, and ornithologists, we curate Fractura, a diverse dataset for vision and learning communities, featuring real-world fracture types across ceramics, bones, eggshells, and lithics. Comprehensive experiments have shown our approach consistently outperforms state-of-the-art methods on both synthetic and real-world datasets, achieving 82.87% lower rotation error and 25.15% higher part accuracy. This sheds light on training on synthetic data to advance real-world 3D puzzle solving, demonstrating its strong generalization across unseen object shapes and diverse fracture types. GARF’s code, data and demo are available at https://ai4ce.github.io/GARF/.
[392] Securing the Skies: A Comprehensive Survey on Anti-UAV Methods, Benchmarking, and Future Directions
Yifei Dong, Fengyi Wu, Sanjian Zhang, Guangyu Chen, Yuzhi Hu, Masumi Yano, Jingdong Sun, Siyu Huang, Feng Liu, Qi Dai, Zhi-Qi Cheng
Main category: cs.CV
TL;DR: Survey paper examining anti-UAV technologies focusing on classification, detection, and tracking, covering emerging methods like diffusion models, multi-modal fusion, and discussing gaps in real-time performance and swarm detection.
Details
Motivation: UAVs are essential for infrastructure inspection and surveillance but pose significant security challenges, creating a need for comprehensive analysis of anti-UAV technologies to address emerging threats.Method: Systematic survey approach evaluating state-of-the-art solutions across single-modality and multi-sensor pipelines (RGB, infrared, audio, radar, RF), covering emerging methodologies including diffusion-based data synthesis, multi-modal fusion, vision-language modeling, self-supervised learning, and reinforcement learning.
Result: Identifies persistent gaps in real-time performance, stealth detection, and swarm-based scenarios, while providing comprehensive evaluation of existing solutions and benchmarks (including large-scale and adversarial benchmarks).
Conclusion: Highlights urgent need for robust, adaptive anti-UAV systems and outlines open research directions to guide development of next-generation defense strategies in an era of extensive UAV usage.
Abstract: Unmanned Aerial Vehicles (UAVs) are indispensable for infrastructure inspection, surveillance, and related tasks, yet they also introduce critical security challenges. This survey provides a wide-ranging examination of the anti-UAV domain, centering on three core objectives-classification, detection, and tracking-while detailing emerging methodologies such as diffusion-based data synthesis, multi-modal fusion, vision-language modeling, self-supervised learning, and reinforcement learning. We systematically evaluate state-of-the-art solutions across both single-modality and multi-sensor pipelines (spanning RGB, infrared, audio, radar, and RF) and discuss large-scale as well as adversarially oriented benchmarks. Our analysis reveals persistent gaps in real-time performance, stealth detection, and swarm-based scenarios, underscoring pressing needs for robust, adaptive anti-UAV systems. By highlighting open research directions, we aim to foster innovation and guide the development of next-generation defense strategies in an era marked by the extensive use of UAVs.
[393] VoQA: Visual-only Question Answering
Jianing An, Luyang Jiang, Jie Luo, Wenjun Wu, Lei Huang
Main category: cs.CV
TL;DR: The paper introduces Visual-only Question Answering (VoQA), where both scene and question appear within a single image, requiring pure visual reasoning without separate text input, addressing a gap in current VQA benchmarks.
Details
Motivation: Current VQA benchmarks overlook scenarios where questions are visually embedded in scenes, while advanced agents should be able to see questions directly in images like humans do, requiring more realistic visual understanding.Method: Introduces VoQA dataset and evaluates models under pure visual-only zero-shot, prompt-guided, and OCR-assisted settings. Investigates question-alignment fine-tuning strategies to guide models in interpreting visual questions before reasoning.
Result: Current models show clear performance drop in VoQA compared to traditional VQA. Using VoQA dataset with question-alignment fine-tuning yields robust vision-only reasoning while preserving cross-task generalization to traditional VQA.
Conclusion: VoQA enables more realistic visual understanding where questions are embedded in scenes, and the proposed training strategies foster complementary visual and textual reasoning capabilities while maintaining generalization to traditional VQA tasks.
Abstract: Visual understanding requires interpreting both natural scenes and the textual information that appears within them, motivating tasks such as Visual Question Answering (VQA). However, current VQA benchmarks overlook scenarios with visually embedded questions, whereas advanced agents should be able to see the question without separate text input as humans. We introduce Visual-only Question Answering (VoQA), where both the scene and the question appear within a single image, requiring models to perceive and reason purely through vision. This setting supports more realistic visual understanding and interaction in scenarios where questions or instructions are embedded directly in the visual scene. Evaluations under pure visual-only zero-shot, prompt-guided and OCR-assisted settings show that current models exhibit a clear performance drop compared to traditional VQA. To address this, we investigate question-alignment fine-tuning strategies designed to guide models toward interpreting the visual question prior to reasoning. Leveraging VoQA dataset together with these strategies yields robust vision-only reasoning while preserving cross-task generalization to traditional VQA, reflecting the complementary visual and textual reasoning capabilities fostered through VoQA training. The code and data are publicly available.
[394] Rendering-Aware Reinforcement Learning for Vector Graphics Generation
Juan A. Rodriguez, Haotian Zhang, Abhay Puri, Aarash Feizi, Rishav Pramanik, Pascal Wichmann, Arnab Mondal, Mohammad Reza Samsami, Rabiul Awal, Perouz Taslakian, Spandana Gella, Sai Rajeswar, David Vazquez, Christopher Pal, Marco Pedersoli
Main category: cs.CV
TL;DR: RLRF uses reinforcement learning with rendering feedback to improve SVG generation in vision-language models, addressing the gap where models never see rendered outputs during training.
Details
Motivation: Existing VLM approaches for SVG generation struggle with faithful and efficient SVG production because they never observe rendered images during training, lacking visual feedback on their generated code.Method: RLRF (Reinforcement Learning from Rendering Feedback) uses RL to enhance SVG generation by generating SVG roll-outs, rendering them, comparing to original images for reward computation, and using this visual fidelity feedback to guide model improvements.
Result: RLRF significantly outperforms supervised fine-tuning, addressing common failure modes and enabling precise, high-quality SVG generation with strong structural understanding and generalization.
Conclusion: Reinforcement learning with rendering feedback is an effective approach to improve SVG generation in autoregressive VLMs, overcoming limitations of training without visual output observation.
Abstract: Scalable Vector Graphics (SVG) offer a powerful format for representing visual designs as interpretable code. Recent advances in vision-language models (VLMs) have enabled high-quality SVG generation by framing the problem as a code generation task and leveraging large-scale pretraining. VLMs are particularly suitable for this task as they capture both global semantics and fine-grained visual patterns, while transferring knowledge across vision, natural language, and code domains. However, existing VLM approaches often struggle to produce faithful and efficient SVGs because they never observe the rendered images during training. Although differentiable rendering for autoregressive SVG code generation remains unavailable, rendered outputs can still be compared to original inputs, enabling evaluative feedback suitable for reinforcement learning (RL). We introduce RLRF (Reinforcement Learning from Rendering Feedback), an RL method that enhances SVG generation in autoregressive VLMs by leveraging feedback from rendered SVG outputs. Given an input image, the model generates SVG roll-outs that are rendered and compared to the original image to compute a reward. This visual fidelity feedback guides the model toward producing more accurate, efficient, and semantically coherent SVGs. RLRF significantly outperforms supervised fine-tuning, addressing common failure modes and enabling precise, high-quality SVG generation with strong structural understanding and generalization.
[395] DiffProtect: Generate Adversarial Examples with Diffusion Models for Facial Privacy Protection
Jiang Liu, Chun Pong Lau, Zhongliang Guo, Yuxiang Guo, Zhaoyang Wang, Rama Chellappa
Main category: cs.CV
TL;DR: DiffProtect uses diffusion models to generate adversarial face images that protect privacy by fooling facial recognition systems while maintaining high visual quality.
Details
Motivation: Facial recognition systems pose serious privacy concerns for social media users. Existing adversarial protection methods have poor visual quality or low attack success rates, limiting their practical utility.Method: DiffProtect utilizes a diffusion autoencoder to generate semantically meaningful perturbations for facial recognition systems, creating adversarial examples that preserve visual quality while effectively fooling FR systems.
Result: DiffProtect produces more natural-looking encrypted images than state-of-the-art methods while achieving significantly higher attack success rates (24.5% and 25.1% absolute improvements on CelebA-HQ and FFHQ datasets).
Conclusion: Diffusion models can effectively generate high-quality adversarial examples for facial recognition protection, balancing both visual quality and attack performance better than existing methods.
Abstract: The increasingly pervasive facial recognition (FR) systems raise serious concerns about personal privacy, especially for billions of users who have publicly shared their photos on social media. Several attempts have been made to protect individuals from being identified by unauthorized FR systems utilizing adversarial attacks to generate encrypted face images. However, existing methods suffer from poor visual quality or low attack success rates, which limit their utility. Recently, diffusion models have achieved tremendous success in image generation. In this work, we ask: can diffusion models be used to generate adversarial examples to improve both visual quality and attack performance? We propose DiffProtect, which utilizes a diffusion autoencoder to generate semantically meaningful perturbations on FR systems. Extensive experiments demonstrate that DiffProtect produces more natural-looking encrypted images than state-of-the-art methods while achieving significantly higher attack success rates, e.g., 24.5% and 25.1% absolute improvements on the CelebA-HQ and FFHQ datasets.
[396] CraftSVG: Multi-Object Text-to-SVG Synthesis via Layout Guided Diffusion
Ayan Banerjee, Nityanand Mathur, Josep Llados, Umapada Pal, Anjan Dutta
Main category: cs.CV
TL;DR: SVGCraft is a novel end-to-end framework for generating vector graphics scenes from text prompts, surpassing previous single-object approaches through layout generation, masked latents, attention fusion, and SVG optimization.
Details
Motivation: Existing text-to-vector-art research has been limited to generating single objects rather than comprehensive scenes with multiple elements, creating a gap in generating complete vector graphics scenes from textual descriptions.Method: Uses pre-trained LLM for layout generation, produces masked latents in bounding boxes for object placement, employs attention map fusion and diffusion U-Net for composition, optimizes SVG with pre-trained encoder and LPIPS loss with opacity modulation, and explores primitive shapes for canvas completion.
Result: SVGCraft outperforms prior works in abstraction, recognizability, and detail, achieving metrics: CLIP-T: 0.4563, Cosine Similarity: 0.6342, Confusion: 0.66, Aesthetic: 6.7832.
Conclusion: The framework successfully generates comprehensive vector graphics scenes from text prompts, advancing beyond single-object generation and demonstrating superior performance through both qualitative and quantitative assessments.
Abstract: Generating VectorArt from text prompts is a challenging vision task, requiring diverse yet realistic depictions of the seen as well as unseen entities. However, existing research has been mostly limited to the generation of single objects, rather than comprehensive scenes comprising multiple elements. In response, this work introduces SVGCraft, a novel end-to-end framework for the creation of vector graphics depicting entire scenes from textual descriptions. Utilizing a pre-trained LLM for layout generation from text prompts, this framework introduces a technique for producing masked latents in specified bounding boxes for accurate object placement. It introduces a fusion mechanism for integrating attention maps and employs a diffusion U-Net for coherent composition, speeding up the drawing process. The resulting SVG is optimized using a pre-trained encoder and LPIPS loss with opacity modulation to maximize similarity. Additionally, this work explores the potential of primitive shapes in facilitating canvas completion in constrained environments. Through both qualitative and quantitative assessments, SVGCraft is demonstrated to surpass prior works in abstraction, recognizability, and detail, as evidenced by its performance metrics (CLIP-T: 0.4563, Cosine Similarity: 0.6342, Confusion: 0.66, Aesthetic: 6.7832). The code will be available at https://github.com/ayanban011/SVGCraft.
[397] Efficient Generative Adversarial Networks for Color Document Image Enhancement and Binarization Using Multi-scale Feature Extraction
Rui-Yang Ju, KokSheik Wong, Jen-Shiun Chiang
Main category: cs.CV
TL;DR: Proposes an efficient document image binarization method using multi-scale feature extraction with Haar wavelet transformation to reduce GAN training/inference times while maintaining performance.
Details
Motivation: Current GAN-based document enhancement methods require training separate networks for each color channel, leading to long training and inference times that hinder practical application.Method: Uses multi-scale feature extraction incorporating Haar wavelet transformation and normalization to preprocess document images before GAN training, reducing computational complexity.
Result: Achieves 10% reduction in training time and 26% reduction in inference time while maintaining performance at 73.79 Average-Score metric, comparable to state-of-the-art methods.
Conclusion: The proposed efficient method significantly reduces computational costs for document image binarization while preserving enhancement quality, making it more practical for real-world applications.
Abstract: The outcome of text recognition for degraded color documents is often unsatisfactory due to interference from various contaminants. To extract information more efficiently for text recognition, document image enhancement and binarization are often employed as preliminary steps in document analysis. Training independent generative adversarial networks (GANs) for each color channel can generate images where shadows and noise are effectively removed, which subsequently allows for efficient text information extraction. However, employing multiple GANs for different color channels requires long training and inference times. To reduce both the training and inference times of these preliminary steps, we propose an efficient method based on multi-scale feature extraction, which incorporates Haar wavelet transformation and normalization to process document images before submitting them to GANs for training. Experiment results show that our proposed method significantly reduces both the training and inference times while maintaining comparable performances when benchmarked against the state-of-the-art methods. In the best case scenario, a reduction of 10% and 26% are observed for training and inference times, respectively, while maintaining the model performance at 73.79 of Average-Score metric. The implementation of this work is available at https://github.com/RuiyangJu/Efficient_Document_Image_Binarization.
[398] ICAS: Detecting Training Data from Autoregressive Image Generative Models
Hongyao Yu, Yixiang Qiu, Yiheng Yang, Hao Fang, Tianqu Zhuang, Jiaxin Hong, Bin Chen, Hao Wu, Shu-Tao Xia
Main category: cs.CV
TL;DR: First study applying membership inference to autoregressive image generation models, proposing implicit classification with adaptive score aggregation to detect unauthorized training data usage.
Details
Motivation: Address privacy and copyright concerns in autoregressive image generation by developing detection methods for unauthorized training data usage, as current models lack vulnerability assessment to membership inference attacks.Method: Two-component approach: 1) Compute implicit token-wise classification scores within query images, 2) Adaptive score aggregation strategy that emphasizes tokens with lower scores to produce final membership likelihood score.
Result: Method outperforms adapted LLM detection algorithms in class-conditional and text-to-image scenarios, shows strong robustness to data transformations, and reveals two key findings: linear scaling law on membership inference vulnerability and easier detection in scale-wise visual autoregressive models.
Conclusion: Autoregressive image models are vulnerable to membership inference attacks, with larger foundation models showing increased vulnerability; the proposed method effectively detects unauthorized training data usage while revealing important security implications.
Abstract: Autoregressive image generation has witnessed rapid advancements, with prominent models such as scale-wise visual auto-regression pushing the boundaries of visual synthesis. However, these developments also raise significant concerns regarding data privacy and copyright. In response, training data detection has emerged as a critical task for identifying unauthorized data usage in model training. To better understand the vulnerability of autoregressive image generative models to such detection, we conduct the first study applying membership inference to this domain. Our approach comprises two key components: implicit classification and an adaptive score aggregation strategy. First, we compute the implicit token-wise classification score within the query image. Then we propose an adaptive score aggregation strategy to acquire a final score, which places greater emphasis on the tokens with lower scores. A higher final score indicates that the sample is more likely to be involved in the training set. To validate the effectiveness of our method, we adapt existing detection algorithms originally designed for LLMs to visual autoregressive models. Extensive experiments demonstrate the superiority of our method in both class-conditional and text-to-image scenarios. Moreover, our approach exhibits strong robustness and generalization under various data transformations. Furthermore, sufficient experiments suggest two novel key findings: (1) A linear scaling law on membership inference, exposing the vulnerability of large foundation models. (2) Training data from scale-wise visual autoregressive models is easier to detect than other autoregressive paradigms. Our code is available at https://github.com/Chrisqcwx/ImageAR-MIA.
[399] Hi-EF: Benchmarking Emotion Forecasting in Human-interaction
Haoran Wang, Xinji Mai, Zeng Tao, Junxiong Lin, Xuan Tong, Ivy Pan, Shaoqi Yan, Yan Wang, Shuyong Gao
Main category: cs.CV
TL;DR: Paper introduces Human-interaction-based Emotion Forecasting (EF) as a narrower, more structured approach to affective forecasting by focusing on two-party interactions where emotions are influenced by partner’s expressions and cues.
Details
Motivation: Traditional affective forecasting is inaccurate due to reliance on external factors and remains qualitative. There's a need for more structured, feasible approaches to emotion prediction.Method: Narrows affective forecasting scope to Human-interaction-based Emotion Forecasting (EF), focusing on two-party interactions where emotions are influenced by partner’s emotional expressions and informational cues.
Result: Proposes a structured perspective for exploring emotional change patterns, enhancing the feasibility of emotion forecasting compared to traditional approaches.
Conclusion: Human-interaction-based EF provides a more structured and feasible framework for emotion forecasting by focusing on dyadic interactions and partner influence.
Abstract: Affective Forecasting is an psychology task that involves predicting an individual’s future emotional responses, often hampered by reliance on external factors leading to inaccuracies, and typically remains at a qualitative analysis stage. To address these challenges, we narrows the scope of Affective Forecasting by introducing the concept of Human-interaction-based Emotion Forecasting (EF). This task is set within the context of a two-party interaction, positing that an individual’s emotions are significantly influenced by their interaction partner’s emotional expressions and informational cues. This dynamic provides a structured perspective for exploring the patterns of emotional change, thereby enhancing the feasibility of emotion forecasting.
[400] Continuous Perception Matters: Diagnosing Temporal Integration Failures in Multimodal Models
Zeyu Wang, Zhenzhen Weng, Serena Yeung-Levy
Main category: cs.CV
TL;DR: CP-Bench reveals that state-of-the-art multimodal models fail at continuous perception - the ability to integrate visual observations over time - even on a simple cube-counting task with moving camera.
Details
Motivation: Continuous perception (integrating visual observations over time) is essential for real-world understanding but remains largely untested in current multimodal models. The authors aim to isolate and test this specific capability.Method: Created CP-Bench, a minimal controlled benchmark using a simple task: counting identical cubes in synthetic scenes while camera moves, revealing only subsets of objects at any moment. Used static-camera control variant to isolate temporal integration ability.
Result: State-of-the-art models (Qwen-3-VL, InternVL3, GPT-5, Gemini-3-Pro) fail dramatically. Static-camera control confirms failure is due to inability to accumulate evidence across time, not object recognition. Higher FPS, enhanced models, and finetuning don’t help.
Conclusion: Modern multimodal architectures and training paradigms have a fundamental limitation in continuous perception. CP-Bench provides a diagnostic tool and testbed for developing time-consistent visual reasoning models.
Abstract: Continuous perception, the ability to integrate visual observations over time in a continuous stream fashion, is essential for robust real-world understanding, yet remains largely untested in current multimodal models. We introduce CP-Bench, a minimal and fully controlled benchmark designed to isolate this capability using an extremely simple task: counting identical cubes in a synthetic scene while the camera moves and only reveals subsets of objects at any moment. Despite the simplicity of the setting, we find that state-of-the-art open-source and commercial models, including Qwen-3-VL, InternVL3, GPT-5, and Gemini-3-Pro, fail dramatically. A static-camera control variant confirms that the failure arises not from object recognition but from an inability to accumulate evidence across time. Further experiments show that neither higher sampling FPS, perception- or spatial-enhanced models, nor finetuning with additional videos leads to meaningful cross-temporal generalization. Our results reveal a fundamental limitation in modern multimodal architectures and training paradigms. CP-Bench provides a simple yet powerful diagnostic tool and establishes a clean testbed for developing models capable of genuine time-consistent visual reasoning.
[401] VITA: Vision-to-Action Flow Matching Policy
Dechen Gao, Boqi Zhao, Andrew Lee, Ian Chuang, Hanchu Zhou, Hang Wang, Zhe Zhao, Junshan Zhang, Iman Soltani
Main category: cs.CV
TL;DR: VITA is a vision-to-action policy framework that eliminates iterative denoising and visual conditioning during inference, achieving faster execution while maintaining or improving performance.
Details
Motivation: Current flow matching and diffusion policies require iterative denoising from Gaussian noise and repeated visual conditioning during generation, causing substantial time and memory overhead. The authors aim to reduce this complexity by creating a more efficient vision-to-action policy.Method: VITA directly flows from visual representations to latent actions without noise or conditioning modules. It uses an action autoencoder to map raw actions into a structured latent space aligned with visual latents. To prevent latent space collapse, it employs flow latent decoding that anchors generation by backpropagating action reconstruction loss through ODE solving steps.
Result: VITA achieves 1.5x-2x faster inference compared to conventional methods with conditioning modules, while outperforming or matching state-of-the-art policies on 9 simulation and 5 real-world tasks from ALOHA and Robomimic.
Conclusion: VITA demonstrates that eliminating noise and visual conditioning modules in flow matching policies can significantly improve inference speed while maintaining or improving performance, offering a more efficient approach for vision-to-action policy learning.
Abstract: Conventional flow matching and diffusion-based policies sample through iterative denoising from standard noise distributions (e.g., Gaussian), and require conditioning modules to repeatedly incorporate visual information during the generative process, incurring substantial time and memory overhead. To reduce the complexity, we develop VITA(VIsion-To-Action policy), a noise-free and conditioning-free flow matching policy learning framework that directly flows from visual representations to latent actions. Since the source of the flow is visually grounded, VITA eliminates the need of visual conditioning during generation. As expected, bridging vision and action is challenging, because actions are lower-dimensional, less structured, and sparser than visual representations; moreover, flow matching requires the source and target to have the same dimensionality. To overcome this, we introduce an action autoencoder that maps raw actions into a structured latent space aligned with visual latents, trained jointly with flow matching. To further prevent latent space collapse, we propose flow latent decoding, which anchors the latent generation process by backpropagating the action reconstruction loss through the flow matching ODE (ordinary differential equation) solving steps. We evaluate VITA on 9 simulation and 5 real-world tasks from ALOHA and Robomimic. VITA achieves 1.5x-2x faster inference compared to conventional methods with conditioning modules, while outperforming or matching state-of-the-art policies. Codes, datasets, and demos are available at our project page: https://ucd-dare.github.io/VITA/.
[402] SynPlay: Large-Scale Synthetic Human Data with Real-World Diversity for Aerial-View Perception
Jinsub Yim, Hyungtae Lee, Sungmin Eum, Yi-Ting Shen, Yan Zhang, Heesung Kwon, Shuvra S. Bhattacharyya
Main category: cs.CV
TL;DR: SynPlay is a large-scale synthetic human dataset for multi-perspective human localization, especially aerial-view perception, featuring rule-guided motion generation and extensive multi-camera coverage.
Details
Motivation: Existing datasets fail to address the challenge of localizing humans in aerial scenes where subjects occupy only tens of pixels, making fine-grained details irrelevant and shifting focus to motion, behavior, and interactions.Method: SynPlay uses a novel rule-guided motion generation framework combining real-world motion capture with motion evolution graphs, enabling dynamic human actions through high-level game rules rather than predefined scripts. It features extensive multi-camera setup including UAVs, CCTVs, and UGVs.
Result: The dataset contains over 73k images and 6.5M human instances with detailed annotations for detection, segmentation, and keypoint tasks. Training with SynPlay significantly improves human localization performance, especially in few-shot and data-scarce scenarios.
Conclusion: SynPlay addresses critical gaps in existing datasets by focusing on aerial-view human localization with spontaneous behaviors and complex interactions, providing a valuable resource for developing models for long-range human analysis.
Abstract: We introduce SynPlay, a large-scale synthetic human dataset purpose-built for advancing multi-perspective human localization, with a predominant focus on aerial-view perception. SynPlay departs from traditional synthetic datasets by addressing a critical but underexplored challenge: localizing humans in aerial scenes where subjects often occupy only tens of pixels in the image. In such scenarios, fine-grained details like facial features or textures become irrelevant, shifting the burden of recognition to human motion, behavior, and interactions. To meet this need, SynPlay implements a novel rule-guided motion generation framework that combines real-world motion capture with motion evolution graphs. This design enables human actions to evolve dynamically through high-level game rules rather than predefined scripts, resulting in effectively uncountable motion variations. Unlike existing synthetic datasets-which either focus on static visual traits or reuse a limited set of mocap-driven actions-SynPlay captures a wide spectrum of spontaneous behaviors, including complex interactions that naturally emerge from unscripted gameplay scenarios. SynPlay also introduces an extensive multi-camera setup that spans UAVs at random altitudes, CCTVs, and a freely roaming UGV, achieving true near-to-far perspective coverage in a single dataset. The majority of instances are captured from aerial viewpoints at varying scales, directly supporting the development of models for long-range human analysis-a setting where existing datasets fall short. Our data contains over 73k images and 6.5M human instances, with detailed annotations for detection, segmentation, and keypoint tasks. Extensive experiments demonstrate that training with SynPlay significantly improves human localization performance, especially in few-shot and data-scarce scenarios.
[403] HSR-KAN: Efficient Hyperspectral Image Super-Resolution via Kolmogorov-Arnold Networks
Baisong Li, Xingwang Wang, Haixiao Xu
Main category: cs.CV
TL;DR: Proposes HSR-KAN, a hyperspectral image super-resolution model using Kolmogorov-Arnold Networks (KANs) to fuse low-resolution HSI with high-resolution multispectral images, achieving state-of-the-art performance.
Details
Motivation: Hyperspectral images have rich spectral information but obtaining high-resolution HSIs is challenging due to physical imaging limitations. Existing methods need improvement for better spatial-spectral fusion.Method: Uses KANs for HSI super-resolution: 1) KAN-Fusion module integrates spatial information from HR-MSI, 2) KAN Channel Attention Block (KAN-CAB) for post-fusion feature extraction with channel attention, 3) Avoids curse of dimensionality while enhancing fine-grained adjustment.
Result: Extensive experiments show HSR-KAN achieves best performance compared to current state-of-the-art HSI-SR methods in both qualitative and quantitative assessments.
Conclusion: HSR-KAN effectively addresses HSI super-resolution challenges by leveraging KANs for spatial-spectral fusion and attention mechanisms, demonstrating superior performance over existing methods.
Abstract: Hyperspectral images (HSIs) have great potential in various visual tasks due to their rich spectral information. However, obtaining high-resolution hyperspectral images remains challenging due to limitations of physical imaging. Inspired by Kolmogorov-Arnold Networks (KANs), we propose an efficient HSI super-resolution (HSI-SR) model to fuse a low-resolution HSI (LR-HSI) and a high-resolution multispectral image (HR-MSI), yielding a high-resolution HSI (HR-HSI). To achieve the effective integration of spatial information from HR-MSI, we design a fusion module based on KANs, called KAN-Fusion. Further inspired by the channel attention mechanism, we design a spectral channel attention module called KAN Channel Attention Block (KAN-CAB) for post-fusion feature extraction. As a channel attention module integrated with KANs, KAN-CAB not only enhances the fine-grained adjustment ability of deep networks, enabling networks to accurately simulate details of spectral sequences and spatial textures, but also effectively avoid Curse of Dimensionality. Extensive experiments show that, compared to current state-of-the-art HSI-SR methods, proposed HSR-KAN achieves the best performance in terms of both qualitative and quantitative assessments. Our code is available at: https://github.com/Baisonm-Li/HSR-KAN.
[404] Sketch-guided Cage-based 3D Gaussian Splatting Deformation
Tianhao Xie, Noam Aigerman, Eugene Belilovsky, Tiberiu Popa
Main category: cs.CV
TL;DR: A sketch-guided 3D Gaussian Splatting deformation system that enables intuitive geometry editing via silhouette sketches from single viewpoints, combining cage-based deformation with Neural Jacobian Fields and leveraging diffusion priors for semantic plausibility.
Details
Motivation: While 3D Gaussian Splatting has shown promise as a novel 3D representation and some editing capabilities exist, fine-grained control over deformation remains an open challenge. Current systems lack intuitive ways to precisely modify 3D GS geometry.Method: Proposes a sketch-guided deformation system that: 1) Allows users to modify 3D GS geometry by drawing silhouette sketches from single viewpoints, 2) Combines cage-based deformations with a variant of Neural Jacobian Fields for precise control, 3) Leverages large-scale 2D diffusion priors and ControlNet to ensure semantic plausibility of generated deformations.
Result: Through experiments, demonstrates the effectiveness of the method and showcases its ability to animate static 3D GS models as a key application, providing fine-grained deformation control that was previously challenging.
Conclusion: The presented system successfully addresses the challenge of fine-grained deformation control in 3D Gaussian Splatting by introducing an intuitive sketch-guided approach that combines cage-based deformation with Neural Jacobian Fields and diffusion priors, enabling both precise editing and semantic plausibility.
Abstract: 3D Gaussian Splatting (GS) is one of the most promising novel 3D representations that has received great interest in computer graphics and computer vision. While various systems have introduced editing capabilities for 3D GS, such as those guided by text prompts, fine-grained control over deformation remains an open challenge. In this work, we present a novel sketch-guided 3D GS deformation system that allows users to intuitively modify the geometry of a 3D GS model by drawing a silhouette sketch from a single viewpoint. Our approach introduces a new deformation method that combines cage-based deformations with a variant of Neural Jacobian Fields, enabling precise, fine-grained control. Additionally, it leverages large-scale 2D diffusion priors and ControlNet to ensure the generated deformations are semantically plausible. Through a series of experiments, we demonstrate the effectiveness of our method and showcase its ability to animate static 3D GS models as one of its key applications.
[405] Pushing the Limits of Sparsity: A Bag of Tricks for Extreme Pruning
Andy Li, Aiden Durrant, Milan Markovic, Tianjin Huang, Souvik Kundu, Tianlong Chen, Lu Yin, Georgios Leontidis
Main category: cs.CV
TL;DR: EAST enables stable training of neural networks at extreme sparsity levels (99.9-99.99%) through three complementary techniques: Dynamic ReLU phasing, weight sharing, and cyclic sparsity.
Details
Motivation: Existing sparse learning methods work well up to moderate sparsity (95-98%) but fail at extreme sparsity levels due to fragile gradient flow and accuracy collapse, limiting deployment on highly constrained devices.Method: Three complementary techniques: 1) Dynamic ReLU phasing (DyReLU gradually replaced by standard ReLU), 2) weight sharing within residual layers, and 3) cyclic sparsity with dynamic sparsity levels and patterns.
Result: EAST achieves competitive or improved performance at extreme sparsities (99.9-99.99%) on ResNet-34/50 across CIFAR-10, CIFAR-100, and ImageNet, with notable gains at extreme sparsity levels.
Conclusion: The proposed EAST method enables stable training at extreme sparsity levels through complementary mechanisms that address gradient flow and parameter exploration challenges, pushing the boundaries of sparse neural networks.
Abstract: Pruning of deep neural networks has been an effective technique for reducing model size while preserving most of the performance of dense networks, crucial for deploying models on memory and power-constrained devices. While recent sparse learning methods have shown promising performance up to moderate sparsity levels such as 95% and 98%, accuracy quickly deteriorates when pushing sparsities to extreme levels due to unique challenges such as fragile gradient flow. In this work, we explore network performance beyond the commonly studied sparsities, and develop techniques that encourage stable training without accuracy collapse even at extreme sparsities, including 99.90%, 99.95% and 99.99% on ResNet architectures. We propose three complementary techniques that enhance sparse training through different mechanisms: 1) Dynamic ReLU phasing, where DyReLU initially allows for richer parameter exploration before being gradually replaced by standard ReLU, 2) weight sharing which reuses parameters within a residual layer while maintaining the same number of learnable parameters, and 3) cyclic sparsity, where both sparsity levels and sparsity patterns evolve dynamically throughout training to better encourage parameter exploration. We evaluate our method, which we term Extreme Adaptive Sparse Training (EAST) at extreme sparsities using ResNet-34 and ResNet-50 on CIFAR-10, CIFAR-100, and ImageNet, achieving competitive or improved performance compared to existing methods, with notable gains at extreme sparsity levels.
[406] MMIF-AMIN: Adaptive Loss-Driven Multi-Scale Invertible Dense Network for Multimodal Medical Image Fusion
Tao Luo, Weihua Xu
Main category: cs.CV
TL;DR: MMIF-AMIN: A novel multimodal medical image fusion method using invertible networks, multi-scale complementary feature extraction with hybrid attention, and adaptive loss functions to outperform state-of-the-art methods.
Details
Motivation: Multimodal medical image fusion needs to capture both unique and complementary information across different modalities simultaneously for enhanced medical diagnosis, which remains a key research challenge.Method: Proposes MMIF-AMIN with: 1) Invertible Dense Network (IDN) for lossless feature extraction from individual modalities, 2) Multi-scale Complementary Feature Extraction Module (MCFEM) with hybrid attention mechanism, varying convolutional layers, and Transformers, and 3) Adaptive loss function to guide model learning.
Result: Outperforms nine state-of-the-art MMIF methods in both quantitative and qualitative analyses. Ablation experiments confirm effectiveness of each component. Extension to other image fusion tasks also achieves promising performance.
Conclusion: MMIF-AMIN effectively addresses the challenge of capturing both unique and complementary information in multimodal medical image fusion, demonstrating superior performance and generalizability to other fusion tasks.
Abstract: Multimodal medical image fusion (MMIF) aims to integrate images from different modalities to produce a comprehensive image that enhances medical diagnosis by accurately depicting organ structures, tissue textures, and metabolic information. Capturing both the unique and complementary information across multiple modalities simultaneously is a key research challenge in MMIF. To address this challenge, this paper proposes a novel image fusion method, MMIF-AMIN, which features a new architecture that can effectively extract these unique and complementary features. Specifically, an Invertible Dense Network (IDN) is employed for lossless feature extraction from individual modalities. To extract complementary information between modalities, a Multi-scale Complementary Feature Extraction Module (MCFEM) is designed, which incorporates a hybrid attention mechanism, convolutional layers of varying sizes, and Transformers. An adaptive loss function is introduced to guide model learning, addressing the limitations of traditional manually-designed loss functions and enhancing the depth of data mining. Extensive experiments demonstrate that MMIF-AMIN outperforms nine state-of-the-art MMIF methods, delivering superior results in both quantitative and qualitative analyses. Ablation experiments confirm the effectiveness of each component of the proposed method. Additionally, extending MMIF-AMIN to other image fusion tasks also achieves promising performance.
[407] SuperMat: Physically Consistent PBR Material Estimation at Interactive Rates
Yijia Hong, Yuan-Chen Guo, Ran Yi, Yulong Chen, Yan-Pei Cao, Lizhuang Ma
Main category: cs.CV
TL;DR: SuperMat is a single-step framework for decomposing physically-based materials from images into albedo, metallic, and roughness maps with millisecond-scale inference, extended to 3D objects via UV refinement.
Details
Motivation: Current diffusion-based approaches for material decomposition suffer from computational overhead due to multiple denoising steps and separate models for different properties, making them slow and inefficient.Method: A single-step framework that enables end-to-end training with perceptual and re-render losses, decomposing material properties in one inference step, with a UV refinement network for 3D objects to ensure viewpoint consistency.
Result: Achieves state-of-the-art PBR material decomposition quality while reducing inference time from seconds to milliseconds per image, and completes 3D object material estimation in approximately 3 seconds.
Conclusion: SuperMat provides an efficient, high-quality solution for physically-based material decomposition that maintains computational efficiency and physical consistency across both 2D images and 3D objects.
Abstract: Decomposing physically-based materials from images into their constituent properties remains challenging, particularly when maintaining both computational efficiency and physical consistency. While recent diffusion-based approaches have shown promise, they face substantial computational overhead due to multiple denoising steps and separate models for different material properties. We present SuperMat, a single-step framework that achieves high-quality material decomposition with one-step inference. This enables end-to-end training with perceptual and re-render losses while decomposing albedo, metallic, and roughness maps at millisecond-scale speeds. We further extend our framework to 3D objects through a UV refinement network, enabling consistent material estimation across viewpoints while maintaining efficiency. Experiments demonstrate that SuperMat achieves state-of-the-art PBR material decomposition quality while reducing inference time from seconds to milliseconds per image, and completes PBR material estimation for 3D objects in approximately 3 seconds. The project page is at https://hyj542682306.github.io/SuperMat/.
[408] Manual-PA: Learning 3D Part Assembly from Instruction Diagrams
Jiahao Zhang, Anoop Cherian, Cristian Rodriguez, Weijian Deng, Stephen Gould
Main category: cs.CV
TL;DR: Manual-PA: A transformer-based framework that uses furniture assembly manual diagrams to guide 3D part assembly by predicting assembly order and 6D poses, outperforming state-of-the-art methods.
Details
Motivation: Furniture assembly is a challenging discrete-continuous optimization problem with a combinatorially large but sparse solution space. Current machine learning models struggle with this task, but assembly manuals provide valuable diagrammatic instructions that can guide the process.Method: Manual-PA uses a transformer-based framework that leverages assembly manual diagrams. It employs contrastive learning to semantically align 3D parts with their manual illustrations, then splits the problem into discrete (assembly order prediction) and continuous (6D pose estimation) phases by relating parts to the final furniture depicted in manuals.
Result: The method shows significant improvements over state-of-the-art on the PartNet benchmark dataset. Manual-PA also demonstrates strong generalization to real-world IKEA furniture assembly on the IKEA-Manual dataset.
Conclusion: Using assembly manual diagrams and part order information leads to substantial performance gains in furniture assembly tasks. The framework successfully bridges the gap between diagrammatic instructions and 3D assembly, enabling effective learning from sparse solution spaces.
Abstract: Assembling furniture amounts to solving the discrete-continuous optimization task of selecting the furniture parts to assemble and estimating their connecting poses in a physically realistic manner. The problem is hampered by its combinatorially large yet sparse solution space thus making learning to assemble a challenging task for current machine learning models. In this paper, we attempt to solve this task by leveraging the assembly instructions provided in diagrammatic manuals that typically accompany the furniture parts. Our key insight is to use the cues in these diagrams to split the problem into discrete and continuous phases. Specifically, we present Manual-PA, a transformer-based instruction Manual-guided 3D Part Assembly framework that learns to semantically align 3D parts with their illustrations in the manuals using a contrastive learning backbone towards predicting the assembly order and infers the 6D pose of each part via relating it to the final furniture depicted in the manual. To validate the efficacy of our method, we conduct experiments on the benchmark PartNet dataset. Our results show that using the diagrams and the order of the parts lead to significant improvements in assembly performance against the state of the art. Further, Manual-PA demonstrates strong generalization to real-world IKEA furniture assembly on the IKEA-Manual dataset.
[409] A Minimal Subset Approach for Informed Keyframe Sampling in Large-Scale SLAM
Nikolaos Stathoulopoulos, Christoforos Kanellakis, George Nikolakopoulos
Main category: cs.CV
TL;DR: MSA is an online keyframe sampling method for LiDAR SLAM that selects the most impactful keyframes for loop closure by optimizing redundancy minimization and information preservation in feature space, improving place recognition and localization metrics.
Details
Motivation: Loop closure detection in large-scale LiDAR SLAM missions faces computational challenges due to processing numerous candidate pairs. Keyframe sampling bridges front-end odometry and back-end optimization, but current methods may not efficiently select the most impactful keyframes for loop closure.Method: Proposes Minimal Subset Approach (MSA) - an online keyframe sampling method that constructs pose graphs using the most impactful keyframes. MSA optimizes two objectives: redundancy minimization and information preservation, implemented within a sliding window framework. It operates in feature space rather than 3D space to efficiently reduce redundant keyframes while retaining essential information.
Result: Evaluation on diverse public datasets shows MSA outperforms naive methods in reducing false positive rates in place recognition, delivers superior ATE (Absolute Trajectory Error) and RPE (Relative Pose Error) in metric localization, requires no manual parameter tuning, reduces memory usage, and lowers computational overhead during loop closure detection and pose graph optimization.
Conclusion: MSA provides an efficient and scalable keyframe sampling approach for LiDAR SLAM that improves loop closure performance while reducing computational and memory requirements, making it suitable for large-scale missions.
Abstract: Typical LiDAR SLAM architectures feature a front-end for odometry estimation and a back-end for refining and optimizing the trajectory and map, commonly through loop closures. However, loop closure detection in large-scale missions presents significant computational challenges due to the need to identify, verify, and process numerous candidate pairs for pose graph optimization. Keyframe sampling bridges the front-end and back-end by selecting frames for storing and processing during global optimization. This article proposes an online keyframe sampling approach that constructs the pose graph using the most impactful keyframes for loop closure. We introduce the Minimal Subset Approach (MSA), which optimizes two key objectives: redundancy minimization and information preservation, implemented within a sliding window framework. By operating in the feature space rather than 3-D space, MSA efficiently reduces redundant keyframes while retaining essential information. Evaluations on diverse public datasets show that the proposed approach outperforms naive methods in reducing false positive rates in place recognition, while delivering superior ATE and RPE in metric localization, without the need for manual parameter tuning. Additionally, MSA demonstrates efficiency and scalability by reducing memory usage and computational overhead during loop closure detection and pose graph optimization.
[410] MOON: Generative MLLM-based Multimodal Representation Learning for E-commerce Product Understanding
Daoze Zhang, Chenghan Fu, Zhanheng Nie, Jianyu Liu, Wanxian Guan, Yuan Gao, Jun Song, Pengjie Wang, Jian Xu, Bo Zheng
Main category: cs.CV
TL;DR: MOON is the first generative MLLM-based model for product representation learning that addresses multimodal alignment challenges, background noise, and lack of benchmarks through guided MoE modules, semantic region detection, and specialized negative sampling.
Details
Motivation: Existing discriminative dual-flow architectures struggle with many-to-one alignment between multiple product images and texts. Generative MLLMs have potential for better product representation but face challenges: lack of multimodal/aspect-aware modules in LLMs, background noise in product images, and absence of standard evaluation benchmarks.Method: (1) Guided Mixture-of-Experts (MoE) module for targeted multimodal and aspect-specific modeling; (2) Core semantic region detection in product images to reduce background noise interference; (3) Specialized negative sampling strategy to increase difficulty and diversity of negative samples.
Result: MOON demonstrates competitive zero-shot performance on both the proposed MBE benchmark and public datasets, showing strong generalization across downstream tasks including cross-modal retrieval, product classification, and attribute prediction. Case studies and visualizations confirm effectiveness.
Conclusion: MOON successfully addresses key challenges in product representation learning using generative MLLMs, establishing a new approach that outperforms existing discriminative methods and provides a comprehensive benchmark for future research in product understanding.
Abstract: With the rapid advancement of e-commerce, exploring general representations rather than task-specific ones has attracted increasing research attention. For product understanding, although existing discriminative dual-flow architectures drive progress in this field, they inherently struggle to model the many-to-one alignment between multiple images and texts of products. Therefore, we argue that generative Multimodal Large Language Models (MLLMs) hold significant potential for improving product representation learning. Nevertheless, achieving this goal still remains non-trivial due to several key challenges: the lack of multimodal and aspect-aware modeling modules in typical LLMs; the common presence of background noise in product images; and the absence of a standard benchmark for evaluation. To address these issues, we propose the first generative MLLM-based model named MOON for product representation learning. Our method (1) employs a guided Mixture-of-Experts (MoE) module for targeted modeling of multimodal and aspect-specific product content; (2) effectively detects core semantic regions in product images to mitigate the distraction and interference caused by background noise; and (3) introduces the specialized negative sampling strategy to increase the difficulty and diversity of negative samples. In addition, we release a large-scale multimodal benchmark MBE for various product understanding tasks. Experimentally, our model demonstrates competitive zero-shot performance on both our benchmark and the public dataset, showcasing strong generalization across various downstream tasks, including cross-modal retrieval, product classification, and attribute prediction. Furthermore, the case study and visualization illustrate the effectiveness of MOON for product understanding.
[411] Improving Partially Observed Trajectories Forecasting by Target-driven Self-Distillation
Peng Shu, Pengfei Zhu, Mengshi Qi, Liang Liu
Main category: cs.CV
TL;DR: TSD: Target-driven Self-Distillation method for motion forecasting that improves prediction accuracy for both fully and partially observed trajectories through single-stage end-to-end training.
Details
Motivation: Partially observed trajectories degrade performance of state-of-the-art motion forecasting models. Existing knowledge distillation approaches require expensive multi-stage training and can cause performance degradation.Method: Target-driven Self-Distillation (TSD) leverages predicted accurate targets to guide predictions under partial observation. Uses self-distillation to learn from feature distributions of both fully and partially observed trajectories in single end-to-end training.
Result: Significant performance improvements on multiple datasets and state-of-the-art motion forecasting models for both fully observed and partially observed scenarios.
Conclusion: TSD provides an effective single-stage training approach that enhances motion prediction accuracy in both observation conditions, addressing limitations of previous multi-stage distillation methods.
Abstract: Accurate prediction of future trajectories of traffic agents is essential for ensuring safe autonomous driving. However, partially observed trajectories can significantly degrade the performance of even state-of-the-art models. Previous approaches often rely on knowledge distillation to transfer features from fully observed trajectories to partially observed ones. This involves firstly training a fully observed model and then using a distillation process to create the final model. While effective, they require multi-stage training, making the training process very expensive. Moreover, knowledge distillation can lead to a performance degradation of the model. In this paper, we introduce a Target-drivenSelf-Distillation method (TSD) for motion forecasting. Our method leverages predicted accurate targets to guide the model in making predictions under partial observation conditions. By employing self-distillation, the model learns from the feature distributions of both fully observed and partially observed trajectories during a single end-to-end training process. This enhances the model’s ability to predict motion accurately in both fully observed and partially observed scenarios. We evaluate our method on multiple datasets and state-of-the-art motion forecasting models. Extensive experimental results demonstrate that our approach achieves significant performance improvements in both settings. To facilitate further research, we will release our code and model checkpoints.
[412] Contrastive Forward-Forward: A Training Algorithm of Vision Transformer
Hossein Aghagolzadeh, Mehdi Ezoji
Main category: cs.CV
TL;DR: The paper introduces Contrastive Forward-Forward, an improved version of the Forward-Forward algorithm adapted for Vision Transformers, achieving up to 10% accuracy improvement and 5-20x faster convergence compared to baseline Forward-Forward.
Details
Motivation: Researchers seek brain-inspired alternatives to backpropagation. While Forward-Forward is more biologically plausible, it has a performance gap compared to backpropagation and has only been tested on simple MLP networks. The authors aim to extend Forward-Forward to modern architectures like Vision Transformers and improve its performance.Method: The authors extend Forward-Forward to Vision Transformers and propose Contrastive Forward-Forward, inspired by contrastive learning insights. The algorithm modifies the baseline Forward-Forward approach to better suit complex transformer architectures.
Result: Contrastive Forward-Forward achieves up to 10% higher accuracy and 5-20x faster convergence than baseline Forward-Forward. It reduces the performance gap with backpropagation on Vision Transformers and even outperforms backpropagation in certain conditions like inaccurate supervision.
Conclusion: The proposed Contrastive Forward-Forward algorithm successfully extends the brain-inspired Forward-Forward approach to modern Vision Transformers, significantly improving performance over the baseline and narrowing the gap with backpropagation while maintaining biological plausibility.
Abstract: Although backpropagation is widely accepted as a training algorithm for artificial neural networks, researchers are always looking for inspiration from the brain to find ways with potentially better performance. Forward-Forward is a novel training algorithm that is more similar to what occurs in the brain, although there is a significant performance gap compared to backpropagation. In the Forward-Forward algorithm, the loss functions are placed after each layer, and the updating of a layer is done using two local forward passes and one local backward pass. Forward-Forward is in its early stages and has been designed and evaluated on simple multi-layer perceptron networks to solve image classification tasks. In this work, we have extended the use of this algorithm to a more complex and modern network, namely the Vision Transformer. Inspired by insights from contrastive learning, we have attempted to revise this algorithm, leading to the introduction of Contrastive Forward-Forward. Experimental results show that our proposed algorithm performs significantly better than the baseline Forward-Forward leading to an increase of up to 10% in accuracy and accelerating the convergence speed by 5 to 20 times. Furthermore, if we take Cross Entropy as the baseline loss function in backpropagation, it will be demonstrated that the proposed modifications to the baseline Forward-Forward reduce its performance gap compared to backpropagation on Vision Transformer, and even outperforms it in certain conditions, such as inaccurate supervision.
[413] HiMo: High-Speed Objects Motion Compensation in Point Clouds
Qingwen Zhang, Ajinkya Khoche, Yi Yang, Li Ling, Sina Sharif Mansouri, Olov Andersson, Patric Jensfelt
Main category: cs.CV
TL;DR: HiMo is a pipeline that uses scene flow estimation to correct motion distortions in LiDAR point clouds caused by dynamic objects (not just ego motion), improving geometric consistency and visual fidelity for autonomous driving applications.
Details
Motivation: Motion distortions from dynamic objects degrade LiDAR point cloud quality, especially in high-speed environments and multi-LiDAR setups for heavy vehicles. Existing methods focus on ego motion compensation but overlook distortions from other moving objects, leading to errors in object shape and position.Method: HiMo repurposes scene flow estimation for non-ego motion compensation. The authors also developed SeFlow++, a real-time scene flow estimator that addresses limitations of existing self-supervised methods which produce degenerate estimates under high-speed distortion. They introduce two new evaluation metrics: compensation accuracy at point level and shape similarity of objects.
Result: HiMo improves geometric consistency and visual fidelity of dynamic objects in LiDAR point clouds. The method was validated on Argoverse 2, ZOD, and a new real-world dataset with highway driving and multi-LiDAR heavy vehicles. Benefits extend to downstream tasks like semantic segmentation and 3D detection.
Conclusion: HiMo effectively addresses motion distortions from dynamic objects in LiDAR point clouds, particularly important for high-speed environments and multi-LiDAR configurations. The pipeline and accompanying SeFlow++ estimator provide practical solutions for improving autonomous vehicle perception systems.
Abstract: LiDAR point cloud is essential for autonomous vehicles, but motion distortions from dynamic objects degrade the data quality. While previous work has considered distortions caused by ego motion, distortions caused by other moving objects remain largely overlooked, leading to errors in object shape and position. This distortion is particularly pronounced in high-speed environments such as highways and in multi-LiDAR configurations, a common setup for heavy vehicles. To address this challenge, we introduce HiMo, a pipeline that repurposes scene flow estimation for non-ego motion compensation, correcting the representation of dynamic objects in point clouds. During the development of HiMo, we observed that existing self-supervised scene flow estimators often produce degenerate or inconsistent estimates under high-speed distortion. We further propose SeFlow++, a real-time scene flow estimator that achieves state-of-the-art performance on both scene flow and motion compensation. Since well-established motion distortion metrics are absent in the literature, we introduce two evaluation metrics: compensation accuracy at a point level and shape similarity of objects. We validate HiMo through extensive experiments on Argoverse 2, ZOD, and a newly collected real-world dataset featuring highway driving and multi-LiDAR-equipped heavy vehicles. Our findings show that HiMo improves the geometric consistency and visual fidelity of dynamic objects in LiDAR point clouds, benefiting downstream tasks such as semantic segmentation and 3D detection. See https://kin-zhang.github.io/HiMo for more details.
[414] PointNSP: Autoregressive 3D Point Cloud Generation with Next-Scale Level-of-Detail Prediction
Ziqiao Meng, Qichao Wang, Zhiyang Dou, Zixing Song, Zhipeng Zhou, Irwin King, Peilin Zhao
Main category: cs.CV
TL;DR: PointNSP is a coarse-to-fine autoregressive point cloud generation framework that uses multi-scale factorization to overcome the limitations of traditional autoregressive models, achieving SOTA quality while being more efficient than diffusion-based approaches.
Details
Motivation: Autoregressive point cloud generation has lagged behind diffusion-based methods due to artificial ordering constraints that emphasize short-range continuity but fail to capture long-range dependencies, global structure, symmetry, and geometric regularities.Method: PointNSP uses a coarse-to-fine generative framework inspired by level-of-detail principles. It preserves global shape structure at low resolutions and progressively refines fine-grained geometry through next-scale prediction, enabling rich intra-scale interactions without fixed orderings.
Result: PointNSP establishes state-of-the-art generation quality for the first time within the autoregressive paradigm on ShapeNet, surpasses diffusion-based baselines in parameter, training, and inference efficiency, and shows even better performance in dense generation with 8,192 points.
Conclusion: The proposed multi-scale factorization approach successfully aligns autoregressive objectives with the permutation-invariant nature of point sets, demonstrating superior scalability and efficiency while overcoming traditional limitations of autoregressive point cloud generation.
Abstract: Autoregressive point cloud generation has long lagged behind diffusion-based approaches in quality. The performance gap stems from the fact that autoregressive models impose an artificial ordering on inherently unordered point sets, forcing shape generation to proceed as a sequence of local predictions. This sequential bias emphasizes short-range continuity but undermines the model’s capacity to capture long-range dependencies, hindering its ability to enforce global structural properties such as symmetry, consistent topology, and large-scale geometric regularities. Inspired by the level-of-detail (LOD) principle in shape modeling, we propose PointNSP, a coarse-to-fine generative framework that preserves global shape structure at low resolutions and progressively refines fine-grained geometry at higher scales through a next-scale prediction paradigm. This multi-scale factorization aligns the autoregressive objective with the permutation-invariant nature of point sets, enabling rich intra-scale interactions while avoiding brittle fixed orderings. Experiments on ShapeNet show that PointNSP establishes state-of-the-art (SOTA) generation quality for the first time within the autoregressive paradigm. In addition, it surpasses strong diffusion-based baselines in parameter, training, and inference efficiency. Finally, in dense generation with 8,192 points, PointNSP’s advantages become even more pronounced, underscoring its scalability potential.
[415] CLIP-Free, Label-Free, Zero-Shot Concept Bottleneck Models
Fawaz Sammani, Jonas Fischer, Nikos Deligiannis
Main category: cs.CV
TL;DR: A method to convert any frozen visual classifier into a Concept Bottleneck Model without CLIP, manual annotation, or training, achieving SOTA performance.
Details
Motivation: Current CBMs are limited: they either rely on CLIP or require manual annotation, and all need training a linear classifier. The authors want to overcome these three limitations simultaneously.Method: Align the original classifier’s distribution (over class indices) with its vision-language counterpart distribution from textual class names, while preserving classifier performance. No image-concept labels needed, CLIP-free, and zero-shot linear classifier derivation.
Result: Applied to over 40 visual classifiers, the method sets new SOTA, surpassing even supervised CLIP-based CBMs. Also achieves SOTA in zero-shot image captioning, outperforming CLIP-based methods.
Conclusion: The proposed method successfully lifts all three limitations of traditional CBMs, enabling CLIP-free, label-free, zero-shot conversion of any frozen visual classifier into a CBM while preserving performance and achieving state-of-the-art results.
Abstract: Concept Bottleneck Models (CBMs) map dense, high-dimensional feature representations into a set of human-interpretable concepts which are then combined linearly to make a prediction. However, modern CBMs rely on the CLIP model to establish a mapping from dense feature representations to textual concepts, and it remains unclear how to design CBMs for models other than CLIP. Methods that do not use CLIP instead require manual, labor intensive annotation to associate feature representations with concepts. Furthermore, all CBMs necessitate training a linear classifier to map the extracted concepts to class labels. In this work, we lift all three limitations simultaneously by proposing a method that converts any frozen visual classifier into a CBM without requiring image-concept labels (label-free), without relying on the CLIP model (CLIP-free), and by deriving the linear classifier in a zero-shot manner. Our method is formulated by aligning the original classifier’s distribution (over discrete class indices) with its corresponding vision-language counterpart distribution derived from textual class names, while preserving the classifier’s performance. The approach requires no ground-truth image-class annotations, and is highly data-efficient and preserves the classifier’s reasoning process. Applied and tested on over 40 visual classifiers, our resulting CLIP-free, zero-shot CBM sets a new state of the art, surpassing even supervised CLIP-based CBMs. Finally, we also show that our method can be used for zero-shot image captioning, outperforming existing methods based on CLIP, and achieving state of the art results.
[416] InsightDrive: Insight Scene Representation for End-to-End Autonomous Driving
Ruiqi Song, Xianda Guo, Yanlun Peng, Qinggong Wei, Hangbin Wu, Long Chen
Main category: cs.CV
TL;DR: InsightDrive: A lightweight end-to-end autonomous driving framework that combines attention-centric explicit scene representation with reasoning-centric implicit scene representation using LLM knowledge distillation and diffusion-based planning.
Details
Motivation: Conventional autonomous driving methods rely on explicit global scene representations (3D detection, mapping, prediction), but human drivers selectively attend to task-relevant regions and implicitly reason over broader traffic context. The authors aim to align scene understanding more closely with human cognitive patterns.Method: 1) Introduces Insight scene representation combining explicit (attention-centric) and implicit (reasoning-centric) representations. 2) Uses Chain-of-Thought instructions to model human driving cognition. 3) Designs task-level Mixture-of-Experts adapter to inject LLM knowledge at minimal parameter cost. 4) Conditions planner on both representations using diffusion-based generative policy. 5) Establishes knowledge distillation pipeline from human driving to LLMs to onboard models.
Result: Extensive experiments on nuScenes and Navsim benchmarks show significant improvements over conventional scene representation approaches.
Conclusion: InsightDrive successfully bridges the gap between human-like cognitive reasoning and autonomous driving systems through a lightweight framework that combines explicit and implicit scene representations with LLM knowledge distillation, achieving superior performance on standard benchmarks.
Abstract: Conventional end-to-end autonomous driving methods often rely on explicit global scene representations, which typically consist of 3D object detection, online mapping, and motion prediction. In contrast, human drivers selectively attend to task-relevant regions and implicitly reason over the broader traffic context. Motivated by this observation, we introduce a lightweight end-to-end autonomous driving framework, InsightDrive. Unlike approaches that directly embed large language models (LLMs), InsightDrive introduces an Insight scene representation that jointly models attention-centric explicit scene representation and reasoning-centric implicit scene representation, so that scene understanding aligns more closely with human cognitive patterns for trajectory planning. To this end, we employ Chain-of-Thought (CoT) instructions to model human driving cognition and design a task-level Mixture-of-Experts (MoE) adapter that injects this knowledge into the autonomous driving model at negligible parameter cost. We further condition the planner on both explicit and implicit scene representations and employ a diffusion-based generative policy, which produces robust trajectory predictions and decisions. The overall framework establishes a knowledge distillation pipeline that transfers human driving knowledge to LLMs and subsequently to onboard models. Extensive experiments on the nuScenes and Navsim benchmarks demonstrate that InsightDrive achieves significant improvements over conventional scene representation approaches.
[417] Interpreting ResNet-based CLIP via Neuron-Attention Decomposition
Edmund Bu, Yossi Gandelsman
Main category: cs.CV
TL;DR: Neuron-head pairs in CLIP-ResNet can be approximated by single directions in embedding space, enabling text interpretation and applications in semantic segmentation and dataset monitoring.
Details
Motivation: To develop interpretable units in neural networks by analyzing individual computation paths, specifically neuron-head pairs in CLIP-ResNet, to understand their contributions and enable practical applications.Method: Analyze pairwise combinations of neurons and attention heads in CLIP’s attention-pooling layer, approximate neuron-head pairs as single directions in CLIP-ResNet’s image-text embedding space, and interpret them by associating with text.
Result: Found that neuron-head pairs can be approximated by single embedding directions, only a sparse set significantly contribute to output, some polysemantic pairs represent sub-concepts, and these insights enable training-free semantic segmentation and dataset distribution shift monitoring.
Conclusion: Examining individual computation paths reveals interpretable units in neural networks, and these units can be effectively utilized for downstream tasks like semantic segmentation and dataset monitoring.
Abstract: We present a novel technique for interpreting the neurons in CLIP-ResNet by decomposing their contributions to the output into individual computation paths. More specifically, we analyze all pairwise combinations of neurons and the following attention heads of CLIP’s attention-pooling layer. We find that these neuron-head pairs can be approximated by a single direction in CLIP-ResNet’s image-text embedding space. Leveraging this insight, we interpret each neuron-head pair by associating it with text. Additionally, we find that only a sparse set of the neuron-head pairs have a significant contribution to the output value, and that some neuron-head pairs, while polysemantic, represent sub-concepts of their corresponding neurons. We use these observations for two applications. First, we employ the pairs for training-free semantic segmentation, outperforming previous methods for CLIP-ResNet. Second, we utilize the contributions of neuron-head pairs to monitor dataset distribution shifts. Our results demonstrate that examining individual computation paths in neural networks uncovers interpretable units, and that such units can be utilized for downstream tasks.
[418] OmniSVG: A Unified Scalable Vector Graphics Generation Model
Yiying Yang, Wei Cheng, Sijin Chen, Xianfang Zeng, Fukun Yin, Jiaxu Zhang, Liao Wang, Gang Yu, Xingjun Ma, Yu-Gang Jiang
Main category: cs.CV
TL;DR: OmniSVG is a unified framework for generating high-quality, complex SVG images using pre-trained Vision-Language Models, addressing limitations of existing methods that produce unstructured outputs or only simple monochrome icons.
Details
Motivation: SVG is important for graphic design due to resolution independence and editability, but existing SVG generation methods are limited - they either produce unstructured outputs with high computational cost or can only generate simple monochrome icons with oversimplified structures.Method: OmniSVG parameterizes SVG commands and coordinates into discrete tokens, decoupling structural logic from low-level geometry for efficient training while maintaining expressiveness. It leverages pre-trained Vision-Language Models for end-to-end multimodal SVG generation.
Result: Extensive experiments show OmniSVG outperforms existing methods. The authors also introduce MMSVG-2M, a multimodal dataset with 2 million richly annotated SVG assets, and a standardized evaluation protocol for conditional SVG generation tasks.
Conclusion: OmniSVG demonstrates potential for integration into professional SVG design workflows by generating high-quality, complex SVG images efficiently while maintaining structural expressiveness.
Abstract: Scalable Vector Graphics (SVG) is an important image format widely adopted in graphic design because of their resolution independence and editability. The study of generating high-quality SVG has continuously drawn attention from both designers and researchers in the AIGC community. However, existing methods either produces unstructured outputs with huge computational cost or is limited to generating monochrome icons of over-simplified structures. To produce high-quality and complex SVG, we propose OmniSVG, a unified framework that leverages pre-trained Vision-Language Models (VLMs) for end-to-end multimodal SVG generation. By parameterizing SVG commands and coordinates into discrete tokens, OmniSVG decouples structural logic from low-level geometry for efficient training while maintaining the expressiveness of complex SVG structure. To further advance the development of SVG synthesis, we introduce MMSVG-2M, a multimodal dataset with two million richly annotated SVG assets, along with a standardized evaluation protocol for conditional SVG generation tasks. Extensive experiments show that OmniSVG outperforms existing methods and demonstrates its potential for integration into professional SVG design workflows.
[419] M3DLayout: A Multi-Source Dataset of 3D Indoor Layouts and Structured Descriptions for 3D Generation
Yiheng Zhang, Zhuojiang Cai, Mingdao Wang, Meitong Guo, Tianxiao Li, Li Lin, Yuwang Wang
Main category: cs.CV
TL;DR: M3DLayout is a large-scale multi-source dataset for 3D indoor layout generation, combining real scans, CAD designs, and procedural scenes to overcome limitations of existing datasets.
Details
Motivation: Current 3D indoor layout generation models are limited by small, low-diversity datasets with poor annotations. Object layout is crucial for bridging language instructions to 3D scenes, enabling physical plausibility, semantic control, and interactive editing.Method: Created M3DLayout dataset with 21,367 layouts and 433k+ object instances from three sources: real-world scans, professional CAD designs, and procedurally generated scenes. Each layout has structured text descriptions at global, relational, and fine-grained levels.
Result: Benchmarking with text-conditioned diffusion and autoregressive models shows the dataset provides solid foundation for training. Multi-source composition enhances diversity, especially Inf3DLayout subset enabling generation of more complex, detailed scenes with rich small-object information.
Conclusion: M3DLayout serves as valuable resource for advancing text-driven 3D scene synthesis research, addressing dataset limitations and enabling more sophisticated layout generation models.
Abstract: In text-driven 3D scene generation, object layout serves as a crucial intermediate representation that bridges high-level language instructions with detailed geometric output. It not only provides a structural blueprint for ensuring physical plausibility but also supports semantic controllability and interactive editing. However, the learning capabilities of current 3D indoor layout generation models are constrained by the limited scale, diversity, and annotation quality of existing datasets. To address this, we introduce M3DLayout, a large-scale, multi-source dataset for 3D indoor layout generation. M3DLayout comprises 21,367 layouts and over 433k object instances, integrating three distinct sources: real-world scans, professional CAD designs, and procedurally generated scenes. Each layout is paired with detailed structured text describing global scene summaries, relational placements of large furniture, and fine-grained arrangements of smaller items. This diverse and richly annotated resource enables models to learn complex spatial and semantic patterns across a wide variety of indoor environments. To assess the potential of M3DLayout, we establish a benchmark using both a text-conditioned diffusion model and a text-conditioned autoregressive model. Experimental results demonstrate that our dataset provides a solid foundation for training layout generation models. Its multi-source composition enhances diversity, notably through the Inf3DLayout subset which provides rich small-object information, enabling the generation of more complex and detailed scenes. We hope that M3DLayout can serve as a valuable resource for advancing research in text-driven 3D scene synthesis. All dataset and code will be made public upon acceptance.
[420] Dynamic Attention Analysis for Backdoor Detection in Text-to-Image Diffusion Models
Zhongqi Wang, Jie Zhang, Shiguang Shan, Xilin Chen
Main category: cs.CV
TL;DR: DAA: Dynamic Attention Analysis for detecting backdoors in text-to-image diffusion models by analyzing evolution patterns in cross-attention maps, achieving superior performance over static methods.
Details
Motivation: Previous backdoor detection methods focus on static features, but diffusion models have inherent dynamism that can serve as better indicators for detecting backdoor attacks with textual triggers.Method: Two approaches: DAA-I treats attention maps as spatially independent and uses Frobenius norm; DAA-S uses graph-based state equations to model spatial correlations between attention maps and analyzes global asymptotic stability.
Result: Significantly outperforms existing methods across six backdoor attack scenarios, achieving average F1 Score of 79.27% and AUC of 86.27%.
Conclusion: Dynamic characteristics of diffusion models are effective indicators for backdoor detection, and the proposed DAA framework provides a novel perspective that leverages attention map evolution patterns for improved security.
Abstract: Recent studies have revealed that text-to-image diffusion models are vulnerable to backdoor attacks, where attackers implant stealthy textual triggers to manipulate model outputs. Previous backdoor detection methods primarily focus on the static features of backdoor samples. However, a vital property of diffusion models is their inherent dynamism. This study introduces a novel backdoor detection perspective named Dynamic Attention Analysis (DAA), showing that these dynamic characteristics serve as better indicators for backdoor detection. Specifically, by examining the dynamic evolution of cross-attention maps, we observe that backdoor samples exhibit distinct feature evolution patterns at the $<$EOS$>$ token compared to benign samples. To quantify these dynamic anomalies, we first introduce DAA-I, which treats the tokens’ attention maps as spatially independent and measures dynamic feature using the Frobenius norm. Furthermore, to better capture the interactions between attention maps and refine the feature, we propose a dynamical system-based approach, referred to as DAA-S. This model formulates the spatial correlations among attention maps using a graph-based state equation and we theoretically analyze the global asymptotic stability of this method. Extensive experiments across six representative backdoor attack scenarios demonstrate that our approach significantly surpasses existing detection methods, achieving an average F1 Score of 79.27% and an AUC of 86.27%. The code is available at https://github.com/Robin-WZQ/DAA.
[421] SPIRAL: Semantic-Aware Progressive LiDAR Scene Generation and Understanding
Dekai Zhu, Yixuan Hu, Youquan Liu, Dongyue Lu, Lingdong Kong, Slobodan Ilic
Main category: cs.CV
TL;DR: Spiral is a novel range-view LiDAR diffusion model that simultaneously generates depth, reflectance images, and semantic maps, outperforming existing methods while being computationally efficient.
Details
Motivation: Existing range-view LiDAR generation methods only produce unlabeled scenes, and using pretrained segmentation models for labeling results in poor cross-modal consistency. There's a need for a method that generates both geometry and semantics together while maintaining the efficiency advantages of range-view representations.Method: Spiral is a range-view LiDAR diffusion model that generates depth, reflectance images, and semantic maps simultaneously. It uses a novel architecture that preserves the computational efficiency of range-view representations while enabling joint generation of geometry and semantics.
Result: Spiral achieves state-of-the-art performance on SemanticKITTI and nuScenes datasets with the smallest parameter size, outperforming two-step methods that combine generative and segmentation models. The generated range images effectively serve as synthetic data augmentation for downstream segmentation tasks, significantly reducing labeling effort.
Conclusion: Spiral successfully addresses the limitation of existing range-view methods by enabling simultaneous generation of geometric and semantic information while maintaining computational efficiency. The approach demonstrates superior performance and practical utility for LiDAR data augmentation in autonomous driving applications.
Abstract: Leveraging recent diffusion models, LiDAR-based large-scale 3D scene generation has achieved great success. While recent voxel-based approaches can generate both geometric structures and semantic labels, existing range-view methods are limited to producing unlabeled LiDAR scenes. Relying on pretrained segmentation models to predict the semantic maps often results in suboptimal cross-modal consistency. To address this limitation while preserving the advantages of range-view representations, such as computational efficiency and simplified network design, we propose Spiral, a novel range-view LiDAR diffusion model that simultaneously generates depth, reflectance images, and semantic maps. Furthermore, we introduce novel semantic-aware metrics to evaluate the quality of the generated labeled range-view data. Experiments on the SemanticKITTI and nuScenes datasets demonstrate that Spiral achieves state-of-the-art performance with the smallest parameter size, outperforming two-step methods that combine the generative and segmentation models. Additionally, we validate that range images generated by Spiral can be effectively used for synthetic data augmentation in the downstream segmentation training, significantly reducing the labeling effort on LiDAR data.
[422] Uni-X: Mitigating Modality Conflict with a Two-End-Separated Architecture for Unified Multimodal Models
Jitai Hao, Hao Liu, Xinyan Xiao, Qiang Huang, Jun Yu
Main category: cs.CV
TL;DR: Uni-X proposes an X-shaped architecture for unified multimodal models that separates modality-specific processing at the ends while sharing middle layers, solving gradient conflicts between vision and text in autoregressive transformers.
Details
Motivation: Current unified multimodal models using shared autoregressive transformers suffer from severe gradient conflicts between vision and text modalities, especially in shallow and deep layers, due to fundamentally different low-level statistical properties of images and text.Method: Uni-X uses a two-end-separated, middle-shared architecture: initial and final layers are dedicated to modality-specific processing, while middle layers share parameters for high-level semantic fusion. This X-shaped design eliminates gradient conflicts at both ends and reduces conflicts in shared layers.
Result: Uni-X achieves superior training efficiency and parameter efficiency. At 3B parameters, it matches or surpasses 7B AR-based unified multimodal models, achieving GenEval score of 82 for image generation while maintaining strong performance in text and vision understanding tasks.
Conclusion: Uni-X establishes a parameter-efficient and scalable foundation for future unified multimodal modeling by solving gradient conflicts through its X-shaped architecture, enabling better training efficiency and performance with fewer parameters.
Abstract: Unified Multimodal Models (UMMs) built on shared autoregressive (AR) transformers are attractive for their architectural simplicity. However, we identify a critical limitation: when trained on multimodal inputs, modality-shared transformers suffer from severe gradient conflicts between vision and text, particularly in shallow and deep layers. We trace this issue to the fundamentally different low-level statistical properties of images and text, while noting that conflicts diminish in middle layers where representations become more abstract and semantically aligned. To overcome this challenge, we propose Uni-X, a two-end-separated, middle-shared architecture. Uni-X dedicates its initial and final layers to modality-specific processing, while maintaining shared parameters in the middle layers for high-level semantic fusion. This X-shaped design not only eliminates gradient conflicts at both ends but also further alleviates residual conflicts in the shared layers. Extensive experiments validate the effectiveness of Uni-X. Under identical training conditions, Uni-X achieves superior training efficiency compared to strong baselines. When scaled to 3B parameters with larger training data, Uni-X matches or surpasses 7B AR-based UMMs, achieving a GenEval score of 82 for image generation alongside strong performance in text and vision understanding tasks. These results establish Uni-X as a parameter-efficient and scalable foundation for future unified multimodal modeling. Our code is available at https://github.com/CURRENTF/Uni-X
[423] Hallo4: High-Fidelity Dynamic Portrait Animation via Direct Preference Optimization
Jiahao Cui, Yan Chen, Mingwang Xu, Hanlin Shang, Yuxuan Chen, Yun Zhan, Zilong Dong, Yao Yao, Jingdong Wang, Siyu Zhu
Main category: cs.CV
TL;DR: A human-preference-aligned diffusion framework for generating photorealistic portrait animations with improved lip sync, facial expressions, and body motion dynamics through direct preference optimization and temporal motion modulation.
Details
Motivation: Generating highly dynamic and photorealistic portrait animations driven by audio and skeletal motion is challenging due to the need for precise lip synchronization, natural facial expressions, and high-fidelity body motion dynamics.Method: Two key innovations: 1) Direct preference optimization tailored for human-centric animation using curated human preference data, 2) Temporal motion modulation that resolves spatiotemporal resolution mismatches by reshaping motion conditions into dimensionally aligned latent features through temporal channel redistribution and proportional feature expansion.
Result: Experiments demonstrate obvious improvements in lip-audio synchronization, expression vividness, body motion coherence over baseline methods, alongside notable gains in human preference metrics.
Conclusion: The proposed human-preference-aligned diffusion framework effectively addresses challenges in portrait animation generation and is complementary to existing UNet and DiT-based approaches.
Abstract: Generating highly dynamic and photorealistic portrait animations driven by audio and skeletal motion remains challenging due to the need for precise lip synchronization, natural facial expressions, and high-fidelity body motion dynamics. We propose a human-preference-aligned diffusion framework that addresses these challenges through two key innovations. First, we introduce direct preference optimization tailored for human-centric animation, leveraging a curated dataset of human preferences to align generated outputs with perceptual metrics for portrait motion-video alignment and naturalness of expression. Second, the proposed temporal motion modulation resolves spatiotemporal resolution mismatches by reshaping motion conditions into dimensionally aligned latent features through temporal channel redistribution and proportional feature expansion, preserving the fidelity of high-frequency motion details in diffusion-based synthesis. The proposed mechanism is complementary to existing UNet and DiT-based portrait diffusion approaches, and experiments demonstrate obvious improvements in lip-audio synchronization, expression vividness, body motion coherence over baseline methods, alongside notable gains in human preference metrics. Our model and source code can be found at: https://github.com/fudan-generative-vision/hallo4.
[424] Generating Fit Check Videos with a Handheld Camera
Bowei Chen, Brian Curless, Ira Kemelmacher-Shlizerman, Steven M. Seitz
Main category: cs.CV
TL;DR: A method for generating full-body videos from just two mirror selfies and phone motion data, using video diffusion models with multi-reference attention and image-based fine-tuning.
Details
Motivation: Current full-body video capture requires complex setups with mounted cameras, careful framing, and repeated practice. The authors aim to create a more convenient solution using only handheld mobile devices.Method: Uses two static photos (front and back) taken in a mirror, along with IMU motion reference from holding a mobile phone. Proposes a video diffusion-based model with parameter-free frame generation strategy and multi-reference attention mechanism to integrate appearance from both selfies. Also introduces image-based fine-tuning to enhance frame sharpness and improve shadows/reflections for realistic human-scene composition.
Result: Synthesizes realistic videos of users performing target motions, enabling rendering into new scenes with consistent illumination and shadows.
Conclusion: The approach provides a convenient solution for full-body video capture using only mobile devices, overcoming limitations of traditional camera setups through novel video diffusion techniques.
Abstract: Self-captured full-body videos are popular, but most deployments require mounted cameras, carefully-framed shots, and repeated practice. We propose a more convenient solution that enables full-body video capture using handheld mobile devices. Our approach takes as input two static photos (front and back) of you in a mirror, along with an IMU motion reference that you perform while holding your mobile phone, and synthesizes a realistic video of you performing a similar target motion. We enable rendering into a new scene, with consistent illumination and shadows. We propose a novel video diffusion-based model to achieve this. Specifically, we propose a parameter-free frame generation strategy and a multi-reference attention mechanism to effectively integrate appearance information from both the front and back selfies into the video diffusion model. Further, we introduce an image-based fine-tuning strategy to enhance frame sharpness and improve shadows and reflections generation for more realistic human-scene composition.
[425] Learning to Generate Rigid Body Interactions with Video Diffusion Models
David Romero, Ariana Bermudez, Hao Li, Fabio Pizzati, Ivan Laptev
Main category: cs.CV
TL;DR: KineMask is a video generation method that enables realistic rigid body control and object interactions by using a single image and object velocity input, with a two-stage training strategy that removes future motion supervision via object masks.
Details
Motivation: Current video generation models struggle with physically plausible object interactions and lack object-level control mechanisms, limiting their potential as world simulators for robotics and embodied decision making.Method: Two-stage training strategy that gradually removes future motion supervision via object masks, training video diffusion models on synthetic scenes of simple interactions, and integrating low-level motion control with high-level textual conditioning via predicted scene descriptions.
Result: KineMask achieves strong improvements over recent models of comparable size in generating realistic object interactions and demonstrates complementary roles of low- and high-level conditioning in video diffusion models.
Conclusion: KineMask addresses limitations in current video generation approaches by enabling realistic rigid body control and object interactions, with potential applications in robotics and embodied AI through its integration of low-level motion control with high-level conditioning.
Abstract: Recent video generation models have achieved remarkable progress and are now deployed in film, social media production, and advertising. Beyond their creative potential, such models also hold promise as world simulators for robotics and embodied decision making. Despite strong advances, however, current approaches still struggle to generate physically plausible object interactions and lack object-level control mechanisms. To address these limitations, we introduce KineMask, an approach for video generation that enables realistic rigid body control, interactions, and effects. Given a single image and a specified object velocity, our method generates videos with inferred motions and future object interactions. We propose a two-stage training strategy that gradually removes future motion supervision via object masks. Using this strategy we train video diffusion models (VDMs) on synthetic scenes of simple interactions and demonstrate significant improvements of object interactions in real scenes. Furthermore, KineMask integrates low-level motion control with high-level textual conditioning via predicted scene descriptions, leading to support for synthesis of complex dynamical phenomena. Our experiments show that KineMask achieves strong improvements over recent models of comparable size. Ablation studies further highlight the complementary roles of low- and high-level conditioning in VDMs. Our code, model, and data will be made publicly available. Project Page: https://daromog.github.io/KineMask/
[426] Video Anomaly Detection with Semantics-Aware Information Bottleneck
Juntong Li, Lingwei Dang, Qingxin Xiao, Shishuo Shang, Jiajia Cheng, Haomin Wu, Yun Hao, Qingyao Wu
Main category: cs.CV
TL;DR: SIB-VAD: A semi-supervised video anomaly detection framework using adaptive information bottleneck filtering and semantic-aware enhancement to address generalization issues and improve semantic anomaly perception.
Details
Motivation: Address two critical challenges in semi-supervised video anomaly detection: (1) Strong generalization blurs normal-abnormal boundaries, with existing memory modules having rigid prototype-matching limitations; (2) Reliance on low-level appearance/motion cues fails to detect high-level semantic anomalies in complex scenes.Method: Proposes SIB-VAD with two key components: (1) Sparse Feature Filtering Module (SFFM) that compresses normal features into low-dimensional manifolds using information bottleneck principle with adaptive routing to select suitable normal bottleneck subspaces; (2) Multimodal prediction framework jointly modeling appearance, motion, and semantics with consistency constraints and joint error computation.
Result: Experimental results validate the effectiveness of the feature filtering paradigm based on semantics-aware information bottleneck. The framework demonstrates improved VAD performance compared to existing approaches.
Conclusion: SIB-VAD successfully addresses limitations of existing methods by replacing rigid memory modules with adaptive information bottleneck filtering and enhancing semantic awareness through multimodal modeling, achieving more robust video anomaly detection performance.
Abstract: Semi-supervised video anomaly detection methods face two critical challenges: (1) Strong generalization blurs the boundary between normal and abnormal patterns. Although existing approaches attempt to alleviate this issue using memory modules, their rigid prototype-matching process limits adaptability to diverse scenarios; (2) Relying solely on low-level appearance and motion cues makes it difficult to perceive high-level semantic anomalies in complex scenes. To address these limitations, we propose SIB-VAD, a novel framework based on adaptive information bottleneck filtering and semantic-aware enhancement. We propose the Sparse Feature Filtering Module (SFFM) to replace traditional memory modules. It compresses normal features directly into a low-dimensional manifold based on the information bottleneck principle and uses an adaptive routing mechanism to dynamically select the most suitable normal bottleneck subspace. Trained only on normal data, SFFMs only learn normal low-dimensional manifolds, while abnormal features deviate and are effectively filtered. Unlike memory modules, SFFM directly removes abnormal information and adaptively handles scene variations. To improve semantic awareness, we further design a multimodal prediction framework that jointly models appearance, motion, and semantics. Through multimodal consistency constraints and joint error computation, it achieves more robust VAD performance. Experimental results validate the effectiveness of our feature filtering paradigm based on semantics-aware information bottleneck. Project page at https://qzfm.github.io/sib_vad_project_page/
[427] 4DGT: Learning a 4D Gaussian Transformer Using Real-World Monocular Videos
Zhen Xu, Zhengqin Li, Zhao Dong, Xiaowei Zhou, Richard Newcombe, Zhaoyang Lv
Main category: cs.CV
TL;DR: 4DGT is a 4D Gaussian-based Transformer model for dynamic scene reconstruction from monocular posed videos, using a rolling-window approach to process 64 consecutive frames with feed-forward inference, reducing reconstruction time from hours to seconds.
Details
Motivation: To enable efficient dynamic scene reconstruction from monocular posed videos by overcoming limitations of optimization-based methods (slow reconstruction time) and handling complex, time-varying environments with varying object lifespans.Method: Uses 4D Gaussian as inductive bias to unify static and dynamic components; processes 64 consecutive posed frames in rolling-window fashion; employs novel density control strategy for handling longer space-time input; performs purely feed-forward inference with Transformer architecture.
Result: Significantly outperforms prior Gaussian-based networks on real-world videos; achieves on-par accuracy with optimization-based methods on cross-domain videos; reduces reconstruction time from hours to seconds; scales effectively to long video sequences.
Conclusion: 4DGT demonstrates that a feed-forward Transformer model with 4D Gaussian representation can achieve efficient and accurate dynamic scene reconstruction from monocular posed videos, bridging the gap between optimization-based and learning-based approaches.
Abstract: We propose 4DGT, a 4D Gaussian-based Transformer model for dynamic scene reconstruction, trained entirely on real-world monocular posed videos. Using 4D Gaussian as an inductive bias, 4DGT unifies static and dynamic components, enabling the modeling of complex, time-varying environments with varying object lifespans. We proposed a novel density control strategy in training, which enables our 4DGT to handle longer space-time input and remain efficient rendering at runtime. Our model processes 64 consecutive posed frames in a rolling-window fashion, predicting consistent 4D Gaussians in the scene. Unlike optimization-based methods, 4DGT performs purely feed-forward inference, reducing reconstruction time from hours to seconds and scaling effectively to long video sequences. Trained only on large-scale monocular posed video datasets, 4DGT can outperform prior Gaussian-based networks significantly in real-world videos and achieve on-par accuracy with optimization-based methods on cross-domain videos. Project page: https://4dgt.github.io
[428] MAMMA: Markerless & Automatic Multi-Person Motion Action Capture
Hanz Cuevas-Velasquez, Anastasios Yiannakidis, Soyong Shin, Giorgio Becherini, Markus Höschle, Joachim Tesch, Taylor Obersat, Tsvetelina Alexiadis, Eni Halilaj, Michael J. Black
Main category: cs.CV
TL;DR: MAMMA is a markerless motion-capture pipeline that accurately recovers SMPL-X parameters from multi-view video of two-person interactions using dense 2D contact-aware surface landmarks and a novel query-based architecture.
Details
Motivation: Traditional marker-based mocap systems are accurate but require specialized hardware, manual marker placement, and extensive post-processing, making them costly and time-consuming. Existing learning-based methods struggle with occlusions, physical interactions, and are mostly designed for single-person capture.Method: The method predicts dense 2D contact-aware surface landmarks conditioned on segmentation masks, enabling person-specific correspondence estimation under occlusion. It uses a novel architecture with learnable queries for each landmark, trained on a large synthetic multi-view dataset combining diverse human motions including extreme poses, hand motions, and close interactions.
Result: The approach handles complex person-person interactions with greater accuracy than existing methods, offers competitive reconstruction quality compared to commercial marker-based solutions without manual cleanup, and introduces evaluation benchmarks for dense-landmark prediction and markerless motion capture.
Conclusion: MAMMA provides an effective markerless motion-capture solution for two-person interactions that overcomes limitations of traditional systems and existing learning-based methods, with plans to release the dataset, benchmark, method, and code for research purposes.
Abstract: We present MAMMA, a markerless motion-capture pipeline that accurately recovers SMPL-X parameters from multi-view video of two-person interaction sequences. Traditional motion-capture systems rely on physical markers. Although they offer high accuracy, their requirements of specialized hardware, manual marker placement, and extensive post-processing make them costly and time-consuming. Recent learning-based methods attempt to overcome these limitations, but most are designed for single-person capture, rely on sparse keypoints, or struggle with occlusions and physical interactions. In this work, we introduce a method that predicts dense 2D contact-aware surface landmarks conditioned on segmentation masks, enabling person-specific correspondence estimation even under heavy occlusion. We employ a novel architecture that exploits learnable queries for each landmark. We demonstrate that our approach can handle complex person–person interaction and offers greater accuracy than existing methods. To train our network, we construct a large, synthetic multi-view dataset combining human motions from diverse sources, including extreme poses, hand motions, and close interactions. Our dataset yields high-variability synthetic sequences with rich body contact and occlusion, and includes SMPL-X ground-truth annotations with dense 2D landmarks. The result is a system capable of capturing human motion without the need for markers. Our approach offers competitive reconstruction quality compared to commercial marker-based motion-capture solutions, without the extensive manual cleanup. Finally, we address the absence of common benchmarks for dense-landmark prediction and markerless motion capture by introducing two evaluation settings built from real multi-view sequences. We will release our dataset, benchmark, method, training code, and pre-trained model weights for research purposes.
[429] Dynamic Multimodal Prototype Learning in Vision-Language Models
Xingyu Zhu, Shuo Wang, Beier Zhu, Miaoge Li, Yunfan Li, Junfeng Fang, Zhicai Wang, Dongsheng Wang, Hanwang Zhang
Main category: cs.CV
TL;DR: ProtoMM is a training-free framework that constructs multimodal prototypes (combining textual descriptions and visual particles) for test-time adaptation of vision-language models, achieving state-of-the-art zero-shot performance.
Details
Motivation: Previous test-time adaptation methods for vision-language models focus only on textual prototypes, which are insufficient due to ambiguous semantics in class names. This leads to limited performance as textual prototypes fail to capture comprehensive visual concepts.Method: ProtoMM constructs multimodal prototypes as discrete distributions over textual descriptions and visual particles. Visual particles are dynamically updated as testing data flows, enabling continual learning. The framework formulates semantic distance between prototypes and test images as an optimal transport problem to quantify their importance.
Result: Extensive experiments on 15 zero-shot benchmarks show ProtoMM achieves 1.03% average accuracy improvement over state-of-the-art methods on ImageNet and its variant datasets.
Conclusion: ProtoMM effectively addresses the limitations of text-only prototypes by leveraging multimodal features for comprehensive prototype learning, enabling better adaptation of vision-language models during test time without requiring training.
Abstract: With the increasing attention to pre-trained vision-language models (VLMs), \eg, CLIP, substantial efforts have been devoted to many downstream tasks, especially in test-time adaptation (TTA). However, previous works focus on learning prototypes only in the textual modality while overlooking the ambiguous semantics in class names. These ambiguities lead to textual prototypes that are insufficient to capture visual concepts, resulting in limited performance. To address this issue, we introduce \textbf{ProtoMM}, a training-free framework that constructs multimodal prototypes to adapt VLMs during the test time. By viewing the prototype as a discrete distribution over the textual descriptions and visual particles, ProtoMM has the ability to combine the multimodal features for comprehensive prototype learning. More importantly, the visual particles are dynamically updated as the testing stream flows. This allows our multimodal prototypes to continually learn from the data, enhancing their generalizability in unseen scenarios. In addition, we quantify the importance of the prototypes and test images by formulating their semantic distance as an optimal transport problem. Extensive experiments on 15 zero-shot benchmarks demonstrate the effectiveness of our method, achieving a 1.03% average accuracy improvement over state-of-the-art methods on ImageNet and its variant datasets.
[430] Robust Phase-Shifting Profilometry for Arbitrary Motion
Geyou Zhang, Kai Liu, Ao Li, Ce Zhu
Main category: cs.CV
TL;DR: RPSP-AM: A robust phase-shifting profilometry method that eliminates motion errors (ghosting artifacts and ripple distortions) for arbitrary 6-DoF motion, non-rigid deformations, and multi-target movements using optical flow alignment and binomial self-compensation.
Details
Motivation: Traditional phase-shifting profilometry (PSP) is highly accurate but extremely sensitive to object motion, with existing compensation methods still leaving residual errors in complex motion scenarios like 6-DoF motion, non-rigid deformations, and multi-target movements.Method: Two-stage approach: 1) Pixel-wise image alignment using dense optical flow tracking to eliminate ghosting artifacts from image misalignment. 2) Image-sequential binomial self-compensation (I-BSC) method that performs weighted summation of homogeneous fringe images using binomial coefficients to correct ripple-like distortions from phase deviation, with low computational complexity comparable to traditional four-step PSP.
Result: Extensive experiments show RPSP-AM outperforms state-of-the-art methods in compensating for both ghosting artifacts and ripple-like distortions under challenging conditions including 6-DoF motion, non-rigid deformations, and multi-target movements.
Conclusion: The proposed method extends PSP applicability to arbitrary motion scenarios, enabling high-fidelity motion-error-free 3D reconstruction with potential for widespread adoption in robotics, industrial inspection, and medical reconstruction.
Abstract: Phase-shifting profilometry (PSP) enables high-accuracy 3D reconstruction but remains highly susceptible to object motion. Although numerous studies have explored compensation for motion-induced errors, residual inaccuracies still persist, particularly in complex motion scenarios. In this paper, we propose a robust phase-shifting profilometry for arbitrary motion (RPSP-AM), including six-degrees-of-freedom (6-DoF) motion (translation and rotation in any direction), non-rigid deformations, and multi-target movements, achieving high-fidelity motion-error-free 3D reconstruction. We categorize motion errors into two components: 1) ghosting artifacts induced by image misalignment, and 2) ripple-like distortions induced by phase deviation. To eliminate the ghosting artifacts, we perform pixel-wise image alignment based on dense optical flow tracking. To correct ripple-like distortions, we propose a high-accuracy, low-complexity image-sequential binomial self-compensation (I-BSC) method, which performs a summation of the homogeneous fringe images weighted by binomial coefficients, exponentially reducing the ripple-like distortions with a competitive computational speed compared with the traditional four-step phase-shifting method. Extensive experimental results demonstrate that, under challenging conditions such as 6-DoF motion, non-rigid deformations, and multi-target movements, the proposed RPSP-AM outperforms state-of-the-art (SoTA) methods in compensating for both ghosting artifacts and ripple-like distortions. Our approach extends the applicability of PSP to arbitrary motion scenarios, endowing it with potential for widespread adoption in fields such as robotics, industrial inspection, and medical reconstruction.
[431] Benchmarking pig detection and tracking under diverse and challenging conditions
Jonathan Henrich, Christian Post, Maximilian Zilke, Parth Shiroya, Emma Chanut, Amir Mollazadeh Yamchi, Ramin Yahyapour, Thomas Kneib, Imke Traulsen
Main category: cs.CV
TL;DR: This paper addresses the lack of systematic benchmarking for pig behavior monitoring by creating two datasets (PigDetect and PigTrack) and evaluating object detection and multi-object tracking methods under realistic barn conditions.
Details
Motivation: There's a need for automated monitoring of individual pig behavior for animal welfare and farm management, but despite extensive research on object detection and multi-object tracking in pig farming, no systematic benchmarking study has been conducted.Method: Created two datasets: PigDetect for object detection and PigTrack for multi-object tracking, using diverse image/video material from realistic barn conditions with challenging scenarios (occlusions, bad visibility). Evaluated state-of-the-art detection models and compared SORT-based vs end-to-end trainable tracking methods.
Result: Challenging training images improve detection performance; state-of-the-art models outperform real-time alternatives. For tracking, SORT-based methods have better detection performance while end-to-end models show better association performance. Models generalize well to unseen pens, highlighting the importance of high-quality training data.
Conclusion: The study provides systematic benchmarking for pig monitoring, showing that challenging training data improves performance and that while SORT-based tracking currently performs better, end-to-end models show promise for future improvements. The publicly available datasets facilitate reproducibility and further development.
Abstract: To ensure animal welfare and effective management in pig farming, monitoring individual behavior is a crucial prerequisite. While monitoring tasks have traditionally been carried out manually, advances in machine learning have made it possible to collect individualized information in an increasingly automated way. Central to these methods is the localization of animals across space (object detection) and time (multi-object tracking). Despite extensive research of these two tasks in pig farming, a systematic benchmarking study has not yet been conducted. In this work, we address this gap by curating two datasets: PigDetect for object detection and PigTrack for multi-object tracking. The datasets are based on diverse image and video material from realistic barn conditions, and include challenging scenarios such as occlusions or bad visibility. For object detection, we show that challenging training images improve detection performance beyond what is achievable with randomly sampled images alone. Comparing different approaches, we found that state-of-the-art models offer substantial improvements in detection quality over real-time alternatives. For multi-object tracking, we observed that SORT-based methods achieve superior detection performance compared to end-to-end trainable models. However, end-to-end models show better association performance, suggesting they could become strong alternatives in the future. We also investigate characteristic failure cases of end-to-end models, providing guidance for future improvements. The detection and tracking models trained on our datasets perform well in unseen pens, suggesting good generalization capabilities. This highlights the importance of high-quality training data. The datasets and research code are made publicly available to facilitate reproducibility, re-use and further development.
[432] PIF-Net: Ill-Posed Prior Guided Multispectral and Hyperspectral Image Fusion via Invertible Mamba and Fusion-Aware LoRA
Baisong Li, Xingwang Wang, Haixiao Xu
Main category: cs.CV
TL;DR: PIF-Net is a novel fusion framework that incorporates ill-posed priors to effectively fuse multispectral and hyperspectral images, using invertible Mamba architecture for global spectral modeling and a lightweight Fusion-Aware Low-Rank Adaptation module for dynamic feature calibration.
Details
Motivation: Multispectral and hyperspectral image fusion (MHIF) aims to produce high-quality images with both rich spectral information and fine spatial details, but faces fundamental ill-posedness due to spectral-spatial trade-offs and data misalignment issues that previous methods haven't effectively addressed.Method: Proposes PIF-Net with: 1) Explicit incorporation of ill-posed priors to handle data misalignment; 2) Invertible Mamba architecture for global spectral modeling while maintaining information consistency and computational efficiency; 3) Fusion-Aware Low-Rank Adaptation module for dynamic calibration of spectral and spatial features while keeping the model lightweight.
Result: Extensive experiments on multiple benchmark datasets show PIF-Net achieves significantly better image restoration performance than current state-of-the-art methods while maintaining model efficiency.
Conclusion: PIF-Net effectively addresses the ill-posed nature of MHIF through explicit ill-posed priors, invertible architecture for information consistency, and lightweight dynamic feature calibration, achieving superior performance with maintained efficiency.
Abstract: The goal of multispectral and hyperspectral image fusion (MHIF) is to generate high-quality images that simultaneously possess rich spectral information and fine spatial details. However, due to the inherent trade-off between spectral and spatial information and the limited availability of observations, this task is fundamentally ill-posed. Previous studies have not effectively addressed the ill-posed nature caused by data misalignment. To tackle this challenge, we propose a fusion framework named PIF-Net, which explicitly incorporates ill-posed priors to effectively fuse multispectral images and hyperspectral images. To balance global spectral modeling with computational efficiency, we design a method based on an invertible Mamba architecture that maintains information consistency during feature transformation and fusion, ensuring stable gradient flow and process reversibility. Furthermore, we introduce a novel fusion module called the Fusion-Aware Low-Rank Adaptation module, which dynamically calibrates spectral and spatial features while keeping the model lightweight. Extensive experiments on multiple benchmark datasets demonstrate that PIF-Net achieves significantly better image restoration performance than current state-of-the-art methods while maintaining model efficiency.
[433] DynaStride: Dynamic Stride Windowing with MMCoT for Instructional Multi-Scene Captioning
Eddison Pham, Prisha Priyadarshini, Adrian Maliackel, Kanishk Bandi, Cristian Meo, Kevin Zhu
Main category: cs.CV
TL;DR: DynaStride is a pipeline for generating coherent scene-level captions in instructional videos without manual segmentation, using adaptive frame sampling, multimodal windowing, and dynamic stride selection to improve temporal coherence and educational value.
Details
Motivation: Current scene-level captioning in instructional videos often fails to capture temporal structure and visual-textual alignment, leading to incoherent captions that undermine educational intent and procedural learning.Method: DynaStride uses adaptive frame sampling and multimodal windowing on YouCookII dataset, employs multimodal chain-of-thought for action-object pairs, and refines them with dynamic stride window selection algorithm to balance temporal context and redundancy.
Result: Outperforms strong baselines (VLLaMA3, GPT-4o) on both N-gram metrics (BLEU, METEOR) and semantic similarity measures (BERTScore, CLIPScore), with qualitative analysis showing better temporal coherence and informativeness.
Conclusion: DynaStride effectively generates more coherent and informative scene-level captions for instructional videos, offering a promising approach for AI-powered educational content generation without requiring manual segmentation.
Abstract: Scene-level captioning in instructional videos can enhance learning by requiring an understanding of both visual cues and temporal structure. By aligning visual cues with textual guidance, this understanding supports procedural learning and multimodal reasoning, providing a richer context for skill acquisition. However, captions that fail to capture this structure may lack coherence and quality, which can create confusion and undermine the video’s educational intent. To address this gap, we introduce DynaStride, a pipeline to generate coherent, scene-level captions without requiring manual scene segmentation. Using the YouCookII dataset’s scene annotations, DynaStride performs adaptive frame sampling and multimodal windowing to capture key transitions within each scene. It then employs a multimodal chain-of-thought process to produce multiple action-object pairs, which are refined and fused using a dynamic stride window selection algorithm that adaptively balances temporal context and redundancy. The final scene-level caption integrates visual semantics and temporal reasoning in a single instructional caption. Empirical evaluations against strong baselines, including VLLaMA3 and GPT-4o, demonstrate consistent gains on both N-gram-based metrics (BLEU, METEOR) and semantic similarity measures (BERTScore, CLIPScore). Qualitative analyses further show that DynaStride produces captions that are more temporally coherent and informative, suggesting a promising direction for improving AI-powered instructional content generation.
[434] SplatSSC: Decoupled Depth-Guided Gaussian Splatting for Semantic Scene Completion
Rui Qian, Haozhi Cao, Tianchen Deng, Shenghai Yuan, Lihua Xie
Main category: cs.CV
TL;DR: SplatSSC: A monocular 3D semantic scene completion framework using depth-guided Gaussian initialization and decoupled aggregation to improve efficiency and accuracy.
Details
Motivation: Existing object-centric SSC methods using 3D Gaussian primitives suffer from inefficient random initialization and outlier primitives that cause erroneous artifacts, limiting their performance and efficiency.Method: 1) Depth-guided initialization using Group-wise Multi-scale Fusion (GMF) module to generate sparse representative Gaussian primitives from multi-scale image and depth features. 2) Decoupled Gaussian Aggregator (DGA) that separates geometric and semantic predictions during Gaussian-to-voxel splatting to reduce outlier noise. 3) Probability Scale Loss for optimization.
Result: State-of-the-art performance on Occ-ScanNet dataset: +6.3% IoU and +4.1% mIoU improvement over prior methods, while reducing latency and memory cost by >9.3%.
Conclusion: SplatSSC effectively addresses limitations of random Gaussian initialization and outlier artifacts through depth-guided initialization and decoupled aggregation, achieving superior efficiency and accuracy in monocular 3D semantic scene completion.
Abstract: Monocular 3D Semantic Scene Completion (SSC) is a challenging yet promising task that aims to infer dense geometric and semantic descriptions of a scene from a single image. While recent object-centric paradigms significantly improve efficiency by leveraging flexible 3D Gaussian primitives, they still rely heavily on a large number of randomly initialized primitives, which inevitably leads to 1) inefficient primitive initialization and 2) outlier primitives that introduce erroneous artifacts. In this paper, we propose SplatSSC, a novel framework that resolves these limitations with a depth-guided initialization strategy and a principled Gaussian aggregator. Instead of random initialization, SplatSSC utilizes a dedicated depth branch composed of a Group-wise Multi-scale Fusion (GMF) module, which integrates multi-scale image and depth features to generate a sparse yet representative set of initial Gaussian primitives. To mitigate noise from outlier primitives, we develop the Decoupled Gaussian Aggregator (DGA), which enhances robustness by decomposing geometric and semantic predictions during the Gaussian-to-voxel splatting process. Complemented with a specialized Probability Scale Loss, our method achieves state-of-the-art performance on the Occ-ScanNet dataset, outperforming prior approaches by over 6.3% in IoU and 4.1% in mIoU, while reducing both latency and memory cost by more than 9.3%.
[435] Modeling Rapid Contextual Learning in the Visual Cortex with Fast-Weight Deep Autoencoder Networks
Yue Li, Weifan Wang, Tai Sing Lee
Main category: cs.CV
TL;DR: ViT-based autoencoder with LoRA fast weights shows that familiarity training enables early layers to learn global context, similar to biological rapid learning in visual cortex.
Details
Motivation: To understand how familiarity training induces global context sensitivity in early layers of deep networks, inspired by neurophysiological findings of rapid learning in visual cortex through recurrent interactions.Method: Used Vision Transformer (ViT)-based autoencoder with Low-Rank Adaptation (LoRA) to implement fast weights as short-term memory traces, analyzing how familiarity training affects representation alignment and attention scope.
Result: (1) ViT self-attention performs manifold transform similar to neural circuit models; (2) Early layer representations align with top-layer global context; (3) Attention scope broadens within remembered context; (4) LoRA fast weights amplify these effects.
Conclusion: Familiarity training introduces global sensitivity to earlier hierarchical layers, and hybrid fast-slow weight architectures provide viable computational models for studying rapid global context learning in biological systems.
Abstract: Recent neurophysiological studies have revealed that the early visual cortex can rapidly learn global image context, as evidenced by a sparsification of population responses and a reduction in mean activity when exposed to familiar versus novel image contexts. This phenomenon has been attributed primarily to local recurrent interactions, rather than changes in feedforward or feedback pathways, supported by both empirical findings and circuit-level modeling. Recurrent neural circuits capable of simulating these effects have been shown to reshape the geometry of neural manifolds, enhancing robustness and invariance to irrelevant variations. In this study, we employ a Vision Transformer (ViT)-based autoencoder to investigate, from a functional perspective, how familiarity training can induce sensitivity to global context in the early layers of a deep neural network. We hypothesize that rapid learning operates via fast weights, which encode transient or short-term memory traces, and we explore the use of Low-Rank Adaptation (LoRA) to implement such fast weights within each Transformer layer. Our results show that (1) The proposed ViT-based autoencoder’s self-attention circuit performs a manifold transform similar to a neural circuit model of the familiarity effect. (2) Familiarity training aligns latent representations in early layers with those in the top layer that contains global context information. (3) Familiarity training broadens the self-attention scope within the remembered image context. (4) These effects are significantly amplified by LoRA-based fast weights. Together, these findings suggest that familiarity training introduces global sensitivity to earlier layers in a hierarchical network, and that a hybrid fast-and-slow weight architecture may provide a viable computational model for studying rapid global context learning in the brain.
[436] PETAR: Localized Findings Generation with Mask-Aware Vision-Language Modeling for PET Automated Reporting
Danyal Maqbool, Changhee Lee, Zachary Huemann, Samuel D. Church, Matthew E. Larson, Scott B. Perlman, Tomas A. Romero, Joshua D. Warner, Meghan Lubner, Xin Tie, Jameson Merkow, Junjie Hu, Steve Y. Cho, Tyler J. Bradshaw
Main category: cs.CV
TL;DR: First large-scale PET/CT dataset with lesion-level correspondence between 3D volumes and free-text findings, plus a 3D vision-language model for mask-aware PET reporting that outperforms baselines.
Details
Motivation: Automating PET report generation is challenging due to complexity of 3D volumes, wide range of clinical findings, and limited annotated datasets. Current approaches lack lesion-level correspondence between imaging and text.Method: 1) Created PETARSeg-11K dataset with 11,356 lesion descriptions paired with 3D segmentations. 2) Developed PETAR-4B model that jointly encodes PET, CT, and 3D lesion segmentation masks using 3D focal prompts to capture fine-grained lesion details (less than 0.1% of volume).
Result: PETAR-4B substantially outperforms all 2D and 3D baselines on automated metrics. Human study with five physicians confirms clinical utility and establishes correlations between automated metrics and expert judgment.
Conclusion: This work provides foundational dataset and novel architecture for 3D medical vision-language understanding in PET, advancing automated PET reporting with clinically validated performance.
Abstract: Generating automated reports for 3D positron emission tomography (PET) is an important and challenging task in medical imaging. PET plays a vital role in oncology, but automating report generation is difficult due to the complexity of whole-body 3D volumes, the wide range of potential clinical findings, and the limited availability of annotated datasets. To address these challenges, we introduce PETARSeg-11K, the first large-scale, publicly available dataset that provides lesion-level correspondence between 3D PET/CT volumes and free-text radiological findings. It comprises 11,356 lesion descriptions paired with 3D segmentations. Second, we propose PETAR-4B, a 3D vision-language model designed for mask-aware, spatially grounded PET/CT reporting. PETAR-4B jointly encodes PET, CT, and 3D lesion segmentation masks, using a 3D focal prompt to capture fine-grained details of lesions that normally comprise less than 0.1% of the volume. Evaluations using automated metrics show PETAR-4B substantially outperforming all 2D and 3D baselines. A human study involving five physicians – the first of its kind for automated PET reporting – confirms the model’s clinical utility and establishes correlations between automated metrics and expert judgment. This work provides a foundational dataset and a novel architecture, advancing 3D medical vision-language understanding in PET.
[437] MBMamba: When Memory Buffer Meets Mamba for Structure-Aware Image Deblurring
Hu Gao, Xiaoning Lei, Xichen Xu, Depeng Dang, Lizhuang Ma
Main category: cs.CV
TL;DR: MBMamba: A structure-aware image deblurring network using Mamba architecture with memory buffer mechanism and Ising-inspired regularization loss to address local pixel forgetting and channel redundancy issues.
Details
Motivation: Mamba architecture shows promise for image deblurring but suffers from local pixel forgetting and channel redundancy due to its flatten-and-scan strategy. Existing solutions increase computational complexity and hinder real-time performance.Method: Proposes MBMamba without changing original Mamba architecture: 1) Memory buffer mechanism to preserve historical information for later fusion, enabling reliable modeling of relevance between adjacent features; 2) Ising-inspired regularization loss that simulates energy minimization of physical system’s “mutual attraction” between pixels to maintain image structure and coherence.
Result: Outperforms state-of-the-art approaches on widely used benchmarks.
Conclusion: MBMamba effectively addresses Mamba’s limitations for image deblurring while maintaining computational efficiency, achieving superior performance through memory buffer mechanism and physics-inspired regularization.
Abstract: The Mamba architecture has emerged as a promising alternative to CNNs and Transformers for image deblurring. However, its flatten-and-scan strategy often results in local pixel forgetting and channel redundancy, limiting its ability to effectively aggregate 2D spatial information. Although existing methods mitigate this by modifying the scan strategy or incorporating local feature modules, it increase computational complexity and hinder real-time performance. In this paper, we propose a structure-aware image deblurring network without changing the original Mamba architecture. Specifically, we design a memory buffer mechanism to preserve historical information for later fusion, enabling reliable modeling of relevance between adjacent features. Additionally, we introduce an Ising-inspired regularization loss that simulates the energy minimization of the physical system’s “mutual attraction” between pixels, helping to maintain image structure and coherence. Building on this, we develop MBMamba. Experimental results show that our method outperforms state-of-the-art approaches on widely used benchmarks.
[438] Hybrid Swin Attention Networks for Simultaneously Low-Dose PET and CT Denoising
Yichao Liu, Hengzhi Xue, YueYang Teng, Junwen Guo
Main category: cs.CV
TL;DR: HSANet: A lightweight hybrid Swin attention network with efficient global attention and hybrid upsampling modules for superior LDCT/PET denoising while maintaining practical GPU deployment.
Details
Motivation: LDCT and PET reduce radiation exposure but increase noise/artifacts that compromise diagnostic accuracy, creating a need for effective denoising methods that maintain radiation safety while improving image quality.Method: Proposed Hybrid Swin Attention Network (HSANet) with Efficient Global Attention (EGA) modules for enhanced spatial/channel interaction and a hybrid upsampling module to prevent overfitting to noise.
Result: HSANet achieves superior denoising performance compared to existing methods on public LDCT/PET dataset while maintaining lightweight model size suitable for standard GPU memory configurations.
Conclusion: HSANet provides a highly practical solution for real-world clinical applications by delivering excellent denoising performance with efficient computational requirements.
Abstract: Low-dose computed tomography (LDCT) and positron emission tomography (PET) have emerged as safer alternatives to conventional imaging modalities by significantly reducing radiation exposure. However, this reduction often results in increased noise and artifacts, which can compromise diagnostic accuracy. Consequently, denoising for LDCT/PET has become a vital area of research aimed at enhancing image quality while maintaining radiation safety. In this study, we introduce a novel Hybrid Swin Attention Network (HSANet), which incorporates Efficient Global Attention (EGA) modules and a hybrid upsampling module. The EGA modules enhance both spatial and channel-wise interaction, improving the network’s capacity to capture relevant features, while the hybrid upsampling module mitigates the risk of overfitting to noise. We validate the proposed approach using a publicly available LDCT/PET dataset. Experimental results demonstrate that HSANet achieves superior denoising performance compared to existing methods, while maintaining a lightweight model size suitable for deployment on GPUs with standard memory configurations. This makes our approach highly practical for real-world clinical applications.
[439] A Cross-Hierarchical Difference Feature Fusion Network Based on Multiscale Encoder-Decoder for Hyperspectral Change Detection
Mingshuai Sheng, Bhatti Uzair Aslam, Junfeng Zhang, Siling Feng, Yonis Gulzar
Main category: cs.CV
TL;DR: Proposes CHDFFN network for hyperspectral change detection using multiscale encoder-decoder with attention mechanisms and adaptive feature fusion to improve spatial-spectral feature capture and change pattern characterization.
Details
Motivation: Existing hyperspectral change detection methods suffer from incomplete capture of multiscale spatial-spectral features and insufficient fusion of differential feature information, limiting performance in environmental monitoring and disaster assessment applications.Method: Cross-Hierarchical Differential Feature Fusion Network (CHDFFN) with: 1) Multiscale feature extraction subnetwork using custom encoder-decoder backbone with residual connections and dual-core channel-spatial attention; 2) Spatial-spectral change feature learning module for hierarchical change representations; 3) Adaptive high-level feature fusion module with dynamic weight assignment.
Result: Experiments on four public hyperspectral datasets show average maximum improvements of 4.61% OA, 19.79% KC, and 18.90% F1 compared to state-of-the-art methods.
Conclusion: CHDFFN effectively addresses limitations of existing methods by better capturing multiscale spatial-spectral features and fusing differential information, demonstrating superior performance in hyperspectral change detection tasks.
Abstract: Hyperspectral change detection (HCD) is one of the core applications of remote sensing images, holding significant research value in fields like environmental monitoring and disaster assessment. However, existing methods often suffer from incomplete capture of multiscale spatial-spectral features and insufficient fusion of differential feature information. To address these challenges, this paper proposes a Cross-Hierarchical Differential Feature Fusion Network (CHDFFN) based on a multiscale encoder-decoder. Firstly, a multiscale feature extraction subnetwork is designed, taking the customized encoder-decoder as the backbone, combined with residual connections and the proposed dual-core channel-spatial attention module to achieve multi-level extraction and initial integration of spatial-spectral features. The encoder embeds convolutional blocks with different receptive field sizes to capture multiscale representations from shallow details to deep semantics. The decoder fuses the encoder’s output via skip connections to gradually restore spatial resolution while suppressing background noise and redundancy. To enhance the model’s ability to capture differential features between bi-temporal hyperspectral images, a spatial-spectral change feature learning module is designed to learn hierarchical change representations. Additionally, an adaptive high-level feature fusion module is proposed, dynamically balancing the contribution of hierarchical differential features by adaptively assigning weights, which effectively strengthens the model’s capability to characterize complex change patterns. Finally, experiments on four public hyperspectral datasets show that compared with some state-of-the-art methods, the average maximum improvements of OA, KC, and F1 are 4.61%, 19.79%, and 18.90% respectively, verifying the model’s effectiveness.
[440] HyCoVAD: A Hybrid SSL-LLM Model for Complex Video Anomaly Detection
Mohammad Mahdi Hemmatyar, Mahdi Jafari, Mohammad Amin Yousefi, Mohammad Reza Nemati, Mobin Azadani, Hamid Reza Rastad, Amirmohammad Akbari
Main category: cs.CV
TL;DR: HyCoVAD is a hybrid SSL-LLM model for complex video anomaly detection that combines self-supervised learning for temporal pattern analysis with LLM-based semantic validation, achieving 72.5% AUC on ComplexVAD dataset while reducing computational costs.
Details
Motivation: Current VAD methods struggle with complex anomalies involving intricate relationships and temporal dependencies among multiple entities. SSL methods lack semantic understanding, while LLMs are computationally expensive and lack spatial localization.Method: HyCoVAD combines a multi-task SSL temporal analyzer (built on nnFormer backbone) with an LLM validator. SSL module identifies suspected anomaly frames through multiple proxy tasks, then LLM applies structured, rule-based reasoning to validate anomalies with semantic context.
Result: Achieves 72.5% frame-level AUC on ComplexVAD dataset, outperforming existing baselines by 12.5% while reducing LLM computation. The approach effectively handles complex anomalies with better semantic understanding.
Conclusion: HyCoVAD successfully bridges the gap between SSL’s spatiotemporal pattern modeling and LLM’s semantic reasoning, providing an efficient hybrid solution for complex video anomaly detection with released taxonomy, protocol, and code for future research.
Abstract: Video anomaly detection (VAD) is crucial for intelligent surveillance, but a significant challenge lies in identifying complex anomalies, which are events defined by intricate relationships and temporal dependencies among multiple entities rather than by isolated actions. While self-supervised learning (SSL) methods effectively model low-level spatiotemporal patterns, they often struggle to grasp the semantic meaning of these interactions. Conversely, large language models (LLMs) offer powerful contextual reasoning but are computationally expensive for frame-by-frame analysis and lack fine-grained spatial localization. We introduce HyCoVAD, Hybrid Complex Video Anomaly Detection, a hybrid SSL-LLM model that combines a multi-task SSL temporal analyzer with LLM validator. The SSL module is built upon an nnFormer backbone which is a transformer-based model for image segmentation. It is trained with multiple proxy tasks, learns from video frames to identify those suspected of anomaly. The selected frames are then forwarded to the LLM, which enriches the analysis with semantic context by applying structured, rule-based reasoning to validate the presence of anomalies. Experiments on the challenging ComplexVAD dataset show that HyCoVAD achieves a 72.5% frame-level AUC, outperforming existing baselines by 12.5% while reducing LLM computation. We release our interaction anomaly taxonomy, adaptive thresholding protocol, and code to facilitate future research in complex VAD scenarios.
[441] Dynamic Prompt Generation for Interactive 3D Medical Image Segmentation Training
Tidiane Camaret Ndir, Alexander Pfefferle, Robin Tibor Schirrmeister
Main category: cs.CV
TL;DR: Training strategy combining dynamic volumetric prompts and adaptive cropping for efficient interactive 3D biomedical image segmentation, achieving competitive performance on benchmark tasks.
Details
Motivation: Current foundation models lack either volumetric awareness or interactive capabilities for 3D biomedical image segmentation, requiring efficient models that can iteratively refine predictions based on user prompts.Method: Combines dynamic volumetric prompt generation with content-aware adaptive cropping to optimize image encoder usage. Simulates realistic user interaction patterns during training while addressing computational challenges of learning from sequential refinement feedback on a single GPU. Initializes network using publicly available nnInteractive segmentation model weights.
Result: Achieved strong performance on Foundation Models for Interactive 3D Biomedical Image Segmentation competition: average final Dice score of 0.6385, normalized surface distance of 0.6614, and area-under-the-curve metrics of 2.4799 (Dice) and 2.5671 (NSD).
Conclusion: The proposed training strategy effectively addresses computational and interactive challenges in 3D biomedical image segmentation, demonstrating competitive performance through efficient use of resources and realistic interaction simulation.
Abstract: Interactive 3D biomedical image segmentation requires efficient models that can iteratively refine predictions based on user prompts. Current foundation models either lack volumetric awareness or suffer from limited interactive capabilities. We propose a training strategy that combines dynamic volumetric prompt generation with content-aware adaptive cropping to optimize the use of the image encoder. Our method simulates realistic user interaction patterns during training while addressing the computational challenges of learning from sequential refinement feedback on a single GPU. For efficient training, we initialize our network using the publicly available weights from the nnInteractive segmentation model. Evaluation on the \textbf{Foundation Models for Interactive 3D Biomedical Image Segmentation} competition demonstrates strong performance with an average final Dice score of 0.6385, normalized surface distance of 0.6614, and area-under-the-curve metrics of 2.4799 (Dice) and 2.5671 (NSD).
[442] Global-to-local image quality assessment in optical microscopy via fast and robust deep learning predictions
Elena Corbetta, Thomas Bocklitz
Main category: cs.CV
TL;DR: μDeepIQA applies deep convolutional neural networks for fast, stable image quality assessment in optical microscopy, providing patch-wise quality visualization and generalizing beyond ideal image conditions.
Details
Motivation: Optical microscopy requires reliable image quality assessment (IQA) for proper data analysis, but traditional IQA methods are computationally expensive for large datasets and unstable for images outside ideal domains. Deep learning approaches offer superior performance, generalizability, and faster predictions.Method: μDeepIQA adapts a deep convolutional neural network architecture originally designed for natural image IQA to optical microscopy data. The model is retrained to predict both individual quality metrics and global quality scores, and provides patch-wise quality predictions for spatial visualization.
Result: The method provides fast and stable quality predictions that generalize even outside the ideal range of standard methods. It enables patch-wise quality assessment and visualization of spatially varying quality within single images.
Conclusion: Optical microscopy studies benefit from deep learning-based IQA due to stable performance with outliers, ability to assess small image patches, rapid predictions, and superior generalizability compared to traditional methods.
Abstract: Optical microscopy is one of the most widely used techniques in research studies for life sciences and biomedicine. These applications require reliable experimental pipelines to extract valuable knowledge from the measured samples and must be supported by image quality assessment (IQA) to ensure correct processing and analysis of the image data. IQA methods are implemented with variable complexity. However, while most quality metrics have a straightforward implementation, they might be time consuming and computationally expensive when evaluating a large dataset. In addition, quality metrics are often designed for well-defined image features and may be unstable for images out of the ideal domain. To overcome these limitations, recent works have proposed deep learning-based IQA methods, which can provide superior performance, increased generalizability and fast prediction. Our method, named $\mathrmμ$DeepIQA, is inspired by previous studies and applies a deep convolutional neural network designed for IQA on natural images to optical microscopy measurements. We retrained the same architecture to predict individual quality metrics and global quality scores for optical microscopy data. The resulting models provide fast and stable predictions of image quality by generalizing quality estimation even outside the ideal range of standard methods. In addition, $\mathrmμ$DeepIQA provides patch-wise prediction of image quality and can be used to visualize spatially varying quality in a single image. Our study demonstrates that optical microscopy-based studies can benefit from the generalizability of deep learning models due to their stable performance in the presence of outliers, the ability to assess small image patches, and rapid predictions.
[443] Have We Scene It All? Scene Graph-Aware Deep Point Cloud Compression
Nikolaos Stathoulopoulos, Christoforos Kanellakis, George Nikolakopoulos
Main category: cs.CV
TL;DR: A semantic scene graph-based deep compression framework for 3D point clouds achieves 98% data reduction while preserving structural and semantic fidelity, supporting downstream robotic applications.
Details
Motivation: Efficient transmission of 3D point cloud data is critical for multi-agent robotic systems under bandwidth constraints and intermittent connectivity, but large point clouds degrade system performance.Method: Decomposes point clouds into semantically coherent patches, encodes them into compact latent representations using semantic-aware encoders conditioned by FiLM, and uses a folding-based decoder guided by latent features and graph node attributes for reconstruction.
Result: Achieves state-of-the-art compression rates (up to 98% data reduction) on SemanticKITTI and nuScenes datasets while preserving structural and semantic fidelity. Supports downstream applications like multi-robot pose graph optimization and map merging with accuracy comparable to raw LiDAR scans.
Conclusion: The semantic scene graph-based deep compression framework enables efficient point cloud transmission for multi-agent robotic systems while maintaining data quality for critical perception tasks.
Abstract: Efficient transmission of 3D point cloud data is critical for advanced perception in centralized and decentralized multi-agent robotic systems, especially nowadays with the growing reliance on edge and cloud-based processing. However, the large and complex nature of point clouds creates challenges under bandwidth constraints and intermittent connectivity, often degrading system performance. We propose a deep compression framework based on semantic scene graphs. The method decomposes point clouds into semantically coherent patches and encodes them into compact latent representations with semantic-aware encoders conditioned by Feature-wise Linear Modulation (FiLM). A folding-based decoder, guided by latent features and graph node attributes, enables structurally accurate reconstruction. Experiments on the SemanticKITTI and nuScenes datasets show that the framework achieves state-of-the-art compression rates, reducing data size by up to 98% while preserving both structural and semantic fidelity. In addition, it supports downstream applications such as multi-robot pose graph optimization and map merging, achieving trajectory accuracy and map alignment comparable to those obtained with raw LiDAR scans.
[444] B2N3D: Progressive Learning from Binary to N-ary Relationships for 3D Object Grounding
Feng Xiao, Hongbin Xu, Hai Ci, Wenxiong Kang
Main category: cs.CV
TL;DR: A progressive relational learning framework for 3D object grounding that extends from binary to n-ary relationships to better match referential descriptions globally.
Details
Motivation: Current 3D object localization methods only model pairwise relationships, ignoring the global perceptual significance of n-ary combinations in multi-modal relational understanding, which is essential for distinguishing similar objects using natural language descriptions.Method: Proposes a progressive relational learning framework that extends relational learning from binary to n-ary relationships, uses grouped supervision loss for n-ary relational learning (since specific annotations for referred objects are absent), and employs a multi-modal network with hybrid attention mechanisms within scene graphs created with n-ary relationships.
Result: The method outperforms state-of-the-art approaches on ReferIt3D and ScanRefer benchmarks, demonstrating the advantages of n-ary relational perception in 3D localization.
Conclusion: N-ary relational learning is crucial for effective 3D object grounding with natural language, as it captures global perceptual significance that pairwise relationships miss, leading to superior localization performance.
Abstract: Localizing 3D objects using natural language is essential for robotic scene understanding. The descriptions often involve multiple spatial relationships to distinguish similar objects, making 3D-language alignment difficult. Current methods only model relationships for pairwise objects, ignoring the global perceptual significance of n-ary combinations in multi-modal relational understanding. To address this, we propose a novel progressive relational learning framework for 3D object grounding. We extend relational learning from binary to n-ary to identify visual relations that match the referential description globally. Given the absence of specific annotations for referred objects in the training data, we design a grouped supervision loss to facilitate n-ary relational learning. In the scene graph created with n-ary relationships, we use a multi-modal network with hybrid attention mechanisms to further localize the target within the n-ary combinations. Experiments and ablation studies on the ReferIt3D and ScanRefer benchmarks demonstrate that our method outperforms the state-of-the-art, and proves the advantages of the n-ary relational perception in 3D localization.
[445] Multivariate Gaussian Representation Learning for Medical Action Evaluation
Luming Yang, Haoxian Liu, Siqing Li, Alper Yilmaz
Main category: cs.CV
TL;DR: The paper introduces CPREval-6k dataset and GaussMedAct framework for fine-grained medical action evaluation, achieving 92.1% Top-1 accuracy with real-time inference.
Details
Motivation: Medical vision faces challenges in fine-grained action evaluation due to lack of comprehensive datasets, high precision requirements, and insufficient modeling of rapid spatiotemporal dynamics in medical procedures.Method: GaussMedAct framework uses Multivariate Gaussian Representation to project joint motions to temporally scaled space and decompose actions into adaptive 3D Gaussian tokens. Hybrid Spatial Encoding employs Cartesian and Vector dual-stream strategy to utilize skeletal joint and bone features.
Result: Achieves 92.1% Top-1 accuracy on CPREval-6k benchmark, outperforming baselines by +5.9% accuracy with only 10% FLOPs. Cross-dataset experiments confirm robustness superiority.
Conclusion: The proposed GaussMedAct framework with CPREval-6k dataset advances medical motion analysis through adaptive spatiotemporal representation learning, enabling accurate real-time evaluation of fine-grained medical actions.
Abstract: Fine-grained action evaluation in medical vision faces unique challenges due to the unavailability of comprehensive datasets, stringent precision requirements, and insufficient spatiotemporal dynamic modeling of very rapid actions. To support development and evaluation, we introduce CPREval-6k, a multi-view, multi-label medical action benchmark containing 6,372 expert-annotated videos with 22 clinical labels. Using this dataset, we present GaussMedAct, a multivariate Gaussian encoding framework, to advance medical motion analysis through adaptive spatiotemporal representation learning. Multivariate Gaussian Representation projects the joint motions to a temporally scaled multi-dimensional space, and decomposes actions into adaptive 3D Gaussians that serve as tokens. These tokens preserve motion semantics through anisotropic covariance modeling while maintaining robustness to spatiotemporal noise. Hybrid Spatial Encoding, employing a Cartesian and Vector dual-stream strategy, effectively utilizes skeletal information in the form of joint and bone features. The proposed method achieves 92.1% Top-1 accuracy with real-time inference on the benchmark, outperforming baseline by +5.9% accuracy with only 10% FLOPs. Cross-dataset experiments confirm the superiority of our method in robustness.
[446] Towards Fast and Scalable Normal Integration using Continuous Components
Francesco Milano, Jen Jen Chung, Lionel Ott, Roland Siegwart
Main category: cs.CV
TL;DR: A novel normal integration method that treats the problem as estimating relative scales of continuous components rather than pixel-level depth optimization, achieving significant speed improvements.
Details
Motivation: Existing normal integration methods require iterative global optimization at the pixel level, which scales poorly to larger normal maps and becomes computationally expensive.Method: Recasts normal integration as estimating relative scales of continuous components, constraining pixels in the same component to vary jointly. Includes component estimation heuristics, optimization term rebalancing, and iterative component merging to reduce problem size.
Result: Achieves state-of-the-art results on standard benchmarks in seconds, with one-order-of-magnitude speedup over pixel-level approaches on large-resolution normal maps.
Conclusion: The component-based approach provides an efficient and scalable solution to normal integration, dramatically reducing computational complexity while maintaining high accuracy.
Abstract: Surface normal integration is a fundamental problem in computer vision, dealing with the objective of reconstructing a surface from its corresponding normal map. Existing approaches require an iterative global optimization to jointly estimate the depth of each pixel, which scales poorly to larger normal maps. In this paper, we address this problem by recasting normal integration as the estimation of relative scales of continuous components. By constraining pixels belonging to the same component to jointly vary their scale, we drastically reduce the number of optimization variables. Our framework includes a heuristic to accurately estimate continuous components from the start, a strategy to rebalance optimization terms, and a technique to iteratively merge components to further reduce the size of the problem. Our method achieves state-of-the-art results on the standard normal integration benchmark in as little as a few seconds and achieves one-order-of-magnitude speedup over pixel-level approaches on large-resolution normal maps.
[447] Rethinking Multimodal Point Cloud Completion: A Completion-by-Correction Perspective
Wang Luo, Di Wu, Hengyuan Na, Yinlin Zhu, Miao Hu, Guocong Quan
Main category: cs.CV
TL;DR: PGNet introduces a new Completion-by-Correction paradigm for point cloud completion, shifting from unconstrained synthesis to guided refinement using shape priors, achieving state-of-the-art results.
Details
Motivation: Current point cloud completion methods using Completion-by-Inpainting paradigm often result in structural inconsistencies and topological artifacts due to limited geometric and semantic constraints when synthesizing missing structures from fused latent features.Method: Proposes Completion-by-Correction paradigm: start with topologically complete shape prior from pretrained image-to-3D model, then perform feature-space correction to align with partial observation. PGNet implements this with multi-stage framework: dual-feature encoding to ground generative prior, synthesizing coarse structurally aligned scaffold, and progressive refinement via hierarchical correction.
Result: Experiments on ShapeNetViPC dataset show PGNet outperforms state-of-the-art baselines with 23.5% reduction in average Chamfer Distance and 7.1% improvement in F-score.
Conclusion: The Completion-by-Correction paradigm provides a more robust approach to point cloud completion by shifting from unconstrained synthesis to guided refinement, enabling structurally consistent and observation-aligned reconstruction.
Abstract: Point cloud completion aims to reconstruct complete 3D shapes from partial observations, which is a challenging problem due to severe occlusions and missing geometry. Despite recent advances in multimodal techniques that leverage complementary RGB images to compensate for missing geometry, most methods still follow a Completion-by-Inpainting paradigm, synthesizing missing structures from fused latent features. We empirically show that this paradigm often results in structural inconsistencies and topological artifacts due to limited geometric and semantic constraints. To address this, we rethink the task and propose a more robust paradigm, termed Completion-by-Correction, which begins with a topologically complete shape prior generated by a pretrained image-to-3D model and performs feature-space correction to align it with the partial observation. This paradigm shifts completion from unconstrained synthesis to guided refinement, enabling structurally consistent and observation-aligned reconstruction. Building upon this paradigm, we introduce PGNet, a multi-stage framework that conducts dual-feature encoding to ground the generative prior, synthesizes a coarse yet structurally aligned scaffold, and progressively refines geometric details via hierarchical correction. Experiments on the ShapeNetViPC dataset demonstrate the superiority of PGNet over state-of-the-art baselines in terms of average Chamfer Distance (-23.5%) and F-score (+7.1%).
[448] Capturing Context-Aware Route Choice Semantics for Trajectory Representation Learning
Ji Cao, Yu Wang, Tongya Zheng, Jie Song, Qinghong Guo, Zujie Ren, Canghong Jin, Gang Chen, Mingli Song
Main category: cs.CV
TL;DR: CORE is a trajectory representation learning framework that integrates context-aware route choice semantics into embeddings by using LLMs for environmental perception and MoE architecture for route choice encoding.
Details
Motivation: Most existing trajectory representation learning methods treat trajectories as static spatiotemporal sequences, ignoring the underlying decision-making process of route choices, which limits the semantic richness of learned representations.Method: CORE has three main components: 1) Multi-granular Environment Perception Module using LLMs to distill environmental semantics from POI distributions, 2) Route Choice Encoder with MoE architecture to capture route choice patterns using context-enriched road network and navigational factors, and 3) Transformer encoder to aggregate representations into global trajectory embeddings.
Result: Extensive experiments on 4 real-world datasets across 6 downstream tasks show CORE consistently outperforms 12 state-of-the-art TRL methods, achieving an average improvement of 9.79% over the best-performing baseline.
Conclusion: CORE successfully bridges the gap by integrating context-aware route choice semantics into trajectory embeddings, demonstrating superior performance across multiple downstream tasks and providing a more semantically rich representation of trajectories.
Abstract: Trajectory representation learning (TRL) aims to encode raw trajectory data into low-dimensional embeddings for downstream tasks such as travel time estimation, mobility prediction, and trajectory similarity analysis. From a behavioral perspective, a trajectory reflects a sequence of route choices within an urban environment. However, most existing TRL methods ignore this underlying decision-making process and instead treat trajectories as static, passive spatiotemporal sequences, thereby limiting the semantic richness of the learned representations. To bridge this gap, we propose CORE, a TRL framework that integrates context-aware route choice semantics into trajectory embeddings. CORE first incorporates a multi-granular Environment Perception Module, which leverages large language models (LLMs) to distill environmental semantics from point of interest (POI) distributions, thereby constructing a context-enriched road network. Building upon this backbone, CORE employs a Route Choice Encoder with a mixture-of-experts (MoE) architecture, which captures route choice patterns by jointly leveraging the context-enriched road network and navigational factors. Finally, a Transformer encoder aggregates the route-choice-aware representations into a global trajectory embedding. Extensive experiments on 4 real-world datasets across 6 downstream tasks demonstrate that CORE consistently outperforms 12 state-of-the-art TRL methods, achieving an average improvement of 9.79% over the best-performing baseline. Our code is available at https://github.com/caoji2001/CORE.
[449] A Comprehensive Survey on World Models for Embodied AI
Xinqing Li, Xin He, Le Zhang, Min Wu, Xiaoli Li, Yun Liu
Main category: cs.CV
TL;DR: Survey paper proposing unified framework for world models in embodied AI with three-axis taxonomy, systematic evaluation metrics, and analysis of current challenges.
Details
Motivation: Embodied AI requires agents that can perceive, act, and anticipate how actions reshape future world states. World models serve as internal simulators for environment dynamics, but there's a need for a unified framework to systematize this rapidly evolving field.Method: Proposes a three-axis taxonomy: (1) Functionality (Decision-Coupled vs. General-Purpose), (2) Temporal Modeling (Sequential Simulation vs. Global Difference Prediction), (3) Spatial Representation (Global Latent Vector, Token Feature Sequence, Spatial Latent Grid, Decomposed Rendering). Systematizes data resources and metrics across robotics, autonomous driving, and general video settings.
Result: Provides quantitative comparison of state-of-the-art models and identifies key open challenges: scarcity of unified datasets, need for evaluation metrics assessing physical consistency over pixel fidelity, trade-off between model performance and computational efficiency, and difficulty achieving long-horizon temporal consistency while mitigating error accumulation.
Conclusion: World models are crucial for embodied AI as internal simulators. The survey offers a unified framework and taxonomy to systematize the field, identifies current limitations, and maintains a curated bibliography to support future research in developing more effective world models for real-world embodied AI applications.
Abstract: Embodied AI requires agents that perceive, act, and anticipate how actions reshape future world states. World models serve as internal simulators that capture environment dynamics, enabling forward and counterfactual rollouts to support perception, prediction, and decision making. This survey presents a unified framework for world models in embodied AI. Specifically, we formalize the problem setting and learning objectives, and propose a three-axis taxonomy encompassing: (1) Functionality, Decision-Coupled vs. General-Purpose; (2) Temporal Modeling, Sequential Simulation and Inference vs. Global Difference Prediction; (3) Spatial Representation, Global Latent Vector, Token Feature Sequence, Spatial Latent Grid, and Decomposed Rendering Representation. We systematize data resources and metrics across robotics, autonomous driving, and general video settings, covering pixel prediction quality, state-level understanding, and task performance. Furthermore, we offer a quantitative comparison of state-of-the-art models and distill key open challenges, including the scarcity of unified datasets and the need for evaluation metrics that assess physical consistency over pixel fidelity, the trade-off between model performance and the computational efficiency required for real-time control, and the core modeling difficulty of achieving long-horizon temporal consistency while mitigating error accumulation. Finally, we maintain a curated bibliography at https://github.com/Li-Zn-H/AwesomeWorldModels.
[450] CrossVid: A Comprehensive Benchmark for Evaluating Cross-Video Reasoning in Multimodal Large Language Models
Jingyao Li, Jingyun Wang, Molin Tan, Haochen Wang, Cilin Yan, Likun Shi, Jiayin Cai, Xiaolong Jiang, Yao Hu
Main category: cs.CV
TL;DR: CrossVid is the first benchmark for evaluating multimodal LLMs’ cross-video reasoning abilities, featuring 5,331 videos and 9,015 QA pairs across 10 tasks, showing current models struggle with integrating evidence across multiple videos.
Details
Motivation: Existing video understanding benchmarks focus on single-video analysis and fail to assess MLLMs' ability to reason across multiple videos simultaneously. Recent multi-view video benchmarks have limited tasks that don't reflect diverse real-world cross-video reasoning scenarios.Method: Created CrossVid benchmark with hierarchical tasks across four high-level dimensions and ten specific tasks. Includes 5,331 videos and 9,015 challenging QA pairs in single-choice, multiple-choice, and open-ended formats to comprehensively evaluate spatial-temporal reasoning in cross-video contexts.
Result: Gemini-2.5-Pro performed best with 50.4% average accuracy. Most current MLLMs struggle with cross-video reasoning tasks due to inability to integrate or compare evidence distributed across multiple videos for reasoning.
Conclusion: CrossVid reveals significant limitations in current MLLMs’ cross-video reasoning capabilities and provides a comprehensive benchmark to guide future advancements in enhancing MLLMs’ ability to reason across multiple videos.
Abstract: Cross-Video Reasoning (CVR) presents a significant challenge in video understanding, which requires simultaneous understanding of multiple videos to aggregate and compare information across groups of videos. Most existing video understanding benchmarks focus on single-video analysis, failing to assess the ability of multimodal large language models (MLLMs) to simultaneously reason over various videos. Recent benchmarks evaluate MLLMs’ capabilities on multi-view videos that capture different perspectives of the same scene. However, their limited tasks hinder a thorough assessment of MLLMs in diverse real-world CVR scenarios. To this end, we introduce CrossVid, the first benchmark designed to comprehensively evaluate MLLMs’ spatial-temporal reasoning ability in cross-video contexts. Firstly, CrossVid encompasses a wide spectrum of hierarchical tasks, comprising four high-level dimensions and ten specific tasks, thereby closely reflecting the complex and varied nature of real-world video understanding. Secondly, CrossVid provides 5,331 videos, along with 9,015 challenging question-answering pairs, spanning single-choice, multiple-choice, and open-ended question formats. Through extensive experiments on various open-source and closed-source MLLMs, we observe that Gemini-2.5-Pro performs best on CrossVid, achieving an average accuracy of 50.4%. Notably, our in-depth case study demonstrates that most current MLLMs struggle with CVR tasks, primarily due to their inability to integrate or compare evidence distributed across multiple videos for reasoning. These insights highlight the potential of CrossVid to guide future advancements in enhancing MLLMs’ CVR capabilities.
[451] Augmenting Moment Retrieval: Zero-Dependency Two-Stage Learning
Zhengxuan Wei, Jiajin Tang, Sibei Yang
Main category: cs.CV
TL;DR: AMR is a zero-external-dependency Augmented Moment Retrieval framework that addresses data scarcity, boundary ambiguity, and semantic discrimination issues in video moment retrieval without requiring additional labeled data.
Details
Motivation: Existing moment retrieval methods suffer from three critical bottlenecks: (1) data scarcity leading to shallow keyword-feature associations, (2) boundary ambiguity in transition regions between events, and (3) insufficient discrimination of fine-grained semantics. These limitations cause models to fall into local optima due to insufficient annotations and lack robust boundary/semantic discrimination capabilities.Method: AMR uses a two-stage training framework: 1) Cold-start stage employs curriculum learning on augmented data to build foundational boundary/semantic awareness. 2) Distillation stage introduces dual query sets: Original Queries maintain DETR-based localization using frozen Base Queries, while Active Queries dynamically adapt to real-data distributions. A cross-stage distillation loss enforces consistency between Original and Base Queries to prevent knowledge forgetting while enabling real-world generalization.
Result: Experiments on multiple benchmarks show that AMR achieves improved performance over prior state-of-the-art approaches, demonstrating effectiveness in overcoming the identified bottlenecks.
Conclusion: AMR successfully addresses key limitations in moment retrieval by resolving ambiguous boundary information and semantic confusion without additional labeled data, while preserving enhanced boundary and semantic discriminative capabilities that generalize to real-world scenarios.
Abstract: Existing Moment Retrieval methods face three critical bottlenecks: (1) data scarcity forces models into shallow keyword-feature associations; (2) boundary ambiguity in transition regions between adjacent events; (3) insufficient discrimination of fine-grained semantics (e.g., distinguishing kicking" vs. throwing" a ball). In this paper, we propose a zero-external-dependency Augmented Moment Retrieval framework, AMR, designed to overcome local optima caused by insufficient data annotations and the lack of robust boundary and semantic discrimination capabilities. AMR is built upon two key insights: (1) it resolves ambiguous boundary information and semantic confusion in existing annotations without additional data (avoiding costly manual labeling), and (2) it preserves boundary and semantic discriminative capabilities enhanced by training while generalizing to real-world scenarios, significantly improving performance. Furthermore, we propose a two-stage training framework with cold-start and distillation adaptation. The cold-start stage employs curriculum learning on augmented data to build foundational boundary/semantic awareness. The distillation stage introduces dual query sets: Original Queries maintain DETR-based localization using frozen Base Queries from the cold-start model, while Active Queries dynamically adapt to real-data distributions. A cross-stage distillation loss enforces consistency between Original and Base Queries, preventing knowledge forgetting while enabling real-world generalization. Experiments on multiple benchmarks show that AMR achieves improved performance over prior state-of-the-art approaches.
[452] BrainPuzzle: Hybrid Physics and Data-Driven Reconstruction for Transcranial Ultrasound Tomography
Shengyu Chen, Shihang Feng, Yi Luo, Xiaowei Jia, Youzuo Lin
Main category: cs.CV
TL;DR: BrainPuzzle: A hybrid two-stage framework combining physics-based reverse time migration with transformer-based machine learning for accurate speed-of-sound mapping in transcranial ultrasound imaging.
Details
Motivation: Transcranial ultrasound imaging faces challenges due to skull-induced signal attenuation, mode conversion, phase aberration, and practical limitations of full-aperture arrays. Traditional physics-based methods struggle with weak signals, while purely data-driven approaches produce anatomically plausible but quantitatively biased results under low SNR and sparse-aperture conditions.Method: Two-stage hybrid framework: 1) Physics-based reverse time migration (time-reversal acoustics) on multi-angle acquisitions to produce migration fragments preserving structural details under low SNR; 2) Transformer-based super-resolution encoder-decoder with graph-based attention unit (GAU) to fuse fragments into coherent, accurate SoS images. Uses partial-array acquisition with movable low-count transducers for improved feasibility.
Result: Experiments on two synthetic datasets show BrainPuzzle achieves superior SoS reconstruction accuracy and image completeness compared to existing methods, demonstrating potential for advancing quantitative ultrasound brain imaging.
Conclusion: BrainPuzzle successfully addresses limitations of both physics-based and purely data-driven approaches by combining their strengths, enabling accurate quantitative transcranial ultrasound imaging even under challenging conditions of low SNR and sparse aperture coverage.
Abstract: Ultrasound brain imaging remains challenging due to the large difference in sound speed between the skull and brain tissues and the difficulty of coupling large probes to the skull. This work aims to achieve quantitative transcranial ultrasound by reconstructing an accurate speed-of-sound (SoS) map of the brain. Traditional physics-based full-waveform inversion (FWI) is limited by weak signals caused by skull-induced attenuation, mode conversion, and phase aberration, as well as incomplete spatial coverage since full-aperture arrays are clinically impractical. In contrast, purely data-driven methods that learn directly from raw ultrasound data often fail to model the complex nonlinear and nonlocal wave propagation through bone, leading to anatomically plausible but quantitatively biased SoS maps under low signal-to-noise and sparse-aperture conditions. To address these issues, we propose BrainPuzzle, a hybrid two-stage framework that combines physical modeling with machine learning. In the first stage, reverse time migration (time-reversal acoustics) is applied to multi-angle acquisitions to produce migration fragments that preserve structural details even under low SNR. In the second stage, a transformer-based super-resolution encoder-decoder with a graph-based attention unit (GAU) fuses these fragments into a coherent and quantitatively accurate SoS image. A partial-array acquisition strategy using a movable low-count transducer set improves feasibility and coupling, while the hybrid algorithm compensates for the missing aperture. Experiments on two synthetic datasets show that BrainPuzzle achieves superior SoS reconstruction accuracy and image completeness, demonstrating its potential for advancing quantitative ultrasound brain imaging.
[453] HARMONY: Hidden Activation Representations and Model Output-Aware Uncertainty Estimation for Vision-Language Models
Erum Mushtaq, Zalan Fabian, Yavuz Faruk Bakman, Anil Ramakrishna, Mahdi Soltanolkotabi, Salman Avestimehr
Main category: cs.CV
TL;DR: HARMONY is a novel uncertainty estimation framework for vision-language models that integrates text tokens, model output probabilities, and hidden representations to better capture inter-token dependencies and vision-text alignment, outperforming existing methods.
Details
Motivation: Existing uncertainty estimation methods have limitations: probability-based approaches use heuristic aggregation that fails to capture complex token relationships and is influenced by language priors, while representation-based methods lose inter-token dependencies. The paper aims to better leverage multimodal alignment signals in hidden representations for improved uncertainty estimation.Method: HARMONY integrates three components at token level: generated text tokens, model’s uncertainty score (MaxProb), and hidden representations capturing visual understanding. It uses appropriate input mapping design and suitable architecture to process these signals effectively, addressing inter-token dependencies and vision-text alignment.
Result: Experiments across two open-ended VQA benchmarks (A-OKVQA and VizWiz) and four VLMs (LLaVA-7B, LLaVA-13B, InstructBLIP, and Qwen-VL) show HARMONY consistently matches or surpasses existing approaches, achieving up to 5% improvement in AUROC and 9% in PRR.
Conclusion: HARMONY demonstrates that effectively processing multimodal alignment signals in hidden representations significantly improves uncertainty estimation performance for vision-language models, providing a more reliable framework for quantifying model output reliability.
Abstract: Uncertainty Estimation (UE) plays a central role in quantifying the reliability of model outputs and reducing unsafe generations via selective prediction. In this regard, most existing probability-based UE approaches rely on predefined functions, aggregating token probabilities into a single UE score using heuristics such as length-normalization. However, these methods often fail to capture the complex relationships between generated tokens and struggle to identify biased probabilities often influenced by \textbf{language priors}. Another line of research uses hidden representations of the model and trains simple MLP architectures to predict uncertainty. However, such functions often lose the intricate \textbf{ inter-token dependencies}. While prior works show that hidden representations encode multimodal alignment signals, our work demonstrates that how these signals are processed has a significant impact on the UE performance. To effectively leverage these signals to identify inter-token dependencies, and vision-text alignment, we propose \textbf{HARMONY} (Hidden Activation Representations and Model Output-Aware Uncertainty Estimation for Vision-Language Models), a novel UE framework that integrates generated tokens (’text’), model’s uncertainty score at the output (‘MaxProb’), and its internal belief on the visual understanding of the image and the generated token (captured by ‘hidden representations’) at token level via appropriate input mapping design and suitable architecture choice. Our experimental experiments across two open-ended VQA benchmarks (A-OKVQA, and VizWiz) and four state-of-the-art VLMs (LLaVA-7B, LLaVA-13B, InstructBLIP, and Qwen-VL) show that HARMONY consistently matches or surpasses existing approaches, achieving up to 5% improvement in AUROC and 9% in PRR.
[454] DynamicTree: Interactive Real Tree Animation via Sparse Voxel Spectrum
Yaokun Li, Lihe Ding, Xiao Chen, Guang Tan, Tianfan Xue
Main category: cs.CV
TL;DR: DynamicTree: A framework for generating realistic, interactive 3D motion for 3DGS-reconstructed trees using sparse voxel spectrum representation and real-time modal analysis.
Details
Motivation: Existing methods struggle to generate structurally consistent and realistic 4D motion for complex real trees in applications like virtual reality, games, and world simulation.Method: Uses compact sparse voxel spectrum to represent tree movement; generates mesh motion from 3D Gaussian Splatting reconstructions, then binds Gaussians to deform mesh; enables fast modal analysis for real-time interactive responses to external forces.
Result: Achieves realistic and responsive tree animations, significantly outperforming existing approaches in both visual quality and computational efficiency.
Conclusion: DynamicTree is the first framework for generating long-term, interactive 3D motion for real tree reconstructions, enabled by sparse voxel spectrum representation and supported by the new 4DTree dataset.
Abstract: Generating dynamic and interactive 3D trees has wide applications in virtual reality, games, and world simulation. However, existing methods still face various challenges in generating structurally consistent and realistic 4D motion for complex real trees. In this paper, we propose DynamicTree, the first framework that can generate long-term, interactive 3D motion for 3DGS reconstructions of real trees. Unlike prior optimization-based methods, our approach generates dynamics in a fast feed-forward manner. The key success of our approach is the use of a compact sparse voxel spectrum to represent the tree movement. Given a 3D tree from Gaussian Splatting reconstruction, our pipeline first generates mesh motion using the sparse voxel spectrum and then binds Gaussians to deform the mesh. Additionally, the proposed sparse voxel spectrum can also serve as a basis for fast modal analysis under external forces, allowing real-time interactive responses. To train our model, we also introduce 4DTree, the first large-scale synthetic 4D tree dataset containing 8,786 animated tree meshes with 100-frame motion sequences. Extensive experiments demonstrate that our method achieves realistic and responsive tree animations, significantly outperforming existing approaches in both visual quality and computational efficiency.
[455] AttnRegDeepLab: A Two-Stage Decoupled Framework for Interpretable Embryo Fragmentation Grading
Ming-Jhe Lee
Main category: cs.CV
TL;DR: AttnRegDeepLab: A dual-branch multi-task learning framework for embryo fragmentation grading that integrates attention gates and multi-scale regression to improve clinical explainability while maintaining segmentation accuracy.
Details
Motivation: Manual embryo fragmentation grading in IVF is subjective and inefficient. Existing deep learning solutions lack clinical explainability and suffer from accumulated errors in segmentation area estimation.Method: Proposes AttnRegDeepLab with dual-branch MTL: 1) Modified DeepLabV3+ decoder with Attention Gates in skip connections to suppress cytoplasmic noise, 2) Multi-Scale Regression Head with Feature Injection to propagate grading priors, 3) 2-stage decoupled training to address gradient conflict, 4) Range-based loss for weakly labeled data.
Result: Achieves robust grading precision while maintaining excellent segmentation accuracy (Dice coefficient = 0.729), outperforming end-to-end counterparts that sacrifice contour integrity for grading error minimization.
Conclusion: Provides a clinically interpretable solution that balances visual fidelity and quantitative precision for embryo fragmentation assessment in IVF.
Abstract: Embryo fragmentation is a morphological indicator critical for evaluating developmental potential in In Vitro Fertilization (IVF). However, manual grading is subjective and inefficient, while existing deep learning solutions often lack clinical explainability or suffer from accumulated errors in segmentation area estimation. To address these issues, this study proposes AttnRegDeepLab (Attention-Guided Regression DeepLab), a framework characterized by dual-branch Multi-Task Learning (MTL). A vanilla DeepLabV3+ decoder is modified by integrating Attention Gates into its skip connections, explicitly suppressing cytoplasmic noise to preserve contour details. Furthermore, a Multi-Scale Regression Head is introduced with a Feature Injection mechanism to propagate global grading priors into the segmentation task, rectifying systematic quantification errors. A 2-stage decoupled training strategy is proposed to address the gradient conflict in MTL. Also, a range-based loss is designed to leverage weakly labeled data. Our method achieves robust grading precision while maintaining excellent segmentation accuracy (Dice coefficient =0.729), in contrast to the end-to-end counterpart that might minimize grading error at the expense of contour integrity. This work provides a clinically interpretable solution that balances visual fidelity and quantitative precision.
[456] PRISM-Bench: A Benchmark of Puzzle-Based Visual Tasks with CoT Error Detection
Yusu Qian, Cheng Wan, Chao Jia, Yinfei Yang, Qingyu Zhao, Zhe Gan
Main category: cs.CV
TL;DR: PRISM-Bench is a benchmark for evaluating multimodal LLMs’ reasoning processes using puzzle-based visual challenges where models must identify the first incorrect step in a chain-of-thought containing exactly one error.
Details
Motivation: Current MLLMs show remarkable progress on vision-language tasks but their reasoning processes remain unreliable. Existing evaluations only measure final-answer accuracy, lacking diagnostic assessment of how reasoning unfolds and whether models can detect logical errors in their own or others' reasoning chains.Method: PRISM-Bench introduces visual puzzles requiring multi-step symbolic, geometric, and analogical reasoning. The diagnostic task presents a puzzle with a step-by-step chain-of-thought containing exactly one error, and models must identify the first incorrect step. This approach resists shortcuts based on superficial pattern matching.
Result: Evaluations across state-of-the-art MLLMs reveal a persistent gap between fluent generation and faithful reasoning. Models that produce plausible chain-of-thoughts often fail to locate simple logical faults, highlighting the disconnect between answer generation and reasoning verification.
Conclusion: PRISM-Bench offers a sharper lens on multimodal reasoning competence by disentangling answer generation from reasoning verification. The benchmark underscores the need for diagnostic evaluation protocols in developing trustworthy MLLMs that can reliably reason and detect errors.
Abstract: Multimodal large language models (MLLMs) have achieved remarkable progress on vision-language tasks, yet their reasoning processes remain sometimes unreliable. We introduce PRISM-Bench, a benchmark of puzzle-based visual challenges designed to evaluate not only whether models can solve problems, but how their reasoning unfolds. Unlike prior evaluations that measure only final-answer accuracy, PRISM-Bench introduces a diagnostic task: given a visual puzzle and a step-by-step chain-of-thought (CoT) containing exactly one error, models must identify the first incorrect step. This setting enables fine-grained assessment of logical consistency, error detection, and visual reasoning. The puzzles in PRISM-Bench require multi-step symbolic, geometric, and analogical reasoning, resisting shortcuts based on superficial pattern matching. Evaluations across state-of-the-art MLLMs reveal a persistent gap between fluent generation and faithful reasoning: models that produce plausible CoTs often fail to locate simple logical faults. By disentangling answer generation from reasoning verification, PRISM-Bench offers a sharper lens on multimodal reasoning competence and underscores the need for diagnostic evaluation protocols in the development of trustworthy MLLMs.
[457] 3EED: Ground Everything Everywhere in 3D
Rong Li, Yuhao Dong, Tianshuai Hu, Ao Liang, Youquan Liu, Dongyue Lu, Liang Pan, Lingdong Kong, Junwei Liang, Ziwei Liu
Main category: cs.CV
TL;DR: 3EED is a large-scale multi-platform 3D visual grounding benchmark with RGB/LiDAR data from vehicles, drones, and quadrupeds, featuring 128K+ objects and 22K+ referring expressions across diverse outdoor scenes.
Details
Motivation: Existing 3D visual grounding benchmarks are limited to indoor settings, single-platform constraints, and small scale, creating a need for comprehensive outdoor benchmarks that support embodied agents in open-world environments.Method: Developed a scalable annotation pipeline combining vision-language model prompting with human verification, plus platform-aware normalization and cross-modal alignment techniques for cross-platform learning.
Result: Created a dataset 10x larger than existing ones, with significant performance gaps revealed in evaluations, highlighting challenges in generalizable 3D grounding across platforms.
Conclusion: 3EED advances language-driven 3D embodied perception research by providing a comprehensive benchmark for cross-platform visual grounding in outdoor environments.
Abstract: Visual grounding in 3D is the key for embodied agents to localize language-referred objects in open-world environments. However, existing benchmarks are limited to indoor focus, single-platform constraints, and small scale. We introduce 3EED, a multi-platform, multi-modal 3D grounding benchmark featuring RGB and LiDAR data from vehicle, drone, and quadruped platforms. We provide over 128,000 objects and 22,000 validated referring expressions across diverse outdoor scenes – 10x larger than existing datasets. We develop a scalable annotation pipeline combining vision-language model prompting with human verification to ensure high-quality spatial grounding. To support cross-platform learning, we propose platform-aware normalization and cross-modal alignment techniques, and establish benchmark protocols for in-domain and cross-platform evaluations. Our findings reveal significant performance gaps, highlighting the challenges and opportunities of generalizable 3D grounding. The 3EED dataset and benchmark toolkit are released to advance future research in language-driven 3D embodied perception.
[458] MoEGCL: Mixture of Ego-Graphs Contrastive Representation Learning for Multi-View Clustering
Jian Zhu, Xin Zou, Jun Sun, Cheng Luo, Lei Liu, Lingfang Zeng, Ning Zhang, Bian Wu, Chang Tang, Lirong Dai
Main category: cs.CV
TL;DR: MoEGCL introduces fine-grained ego-graph fusion at sample level using Mixture-of-Experts, outperforming coarse view-level fusion methods in multi-view clustering.
Details
Motivation: Existing GNN-based multi-view clustering methods suffer from coarse-grained graph fusion, where separate graph structures for each view are fused at view level, which is too rough and limits performance.Method: Two modules: 1) Mixture of Ego-Graphs Fusion (MoEGF) constructs ego graphs and uses Mixture-of-Experts network for fine-grained fusion at sample level; 2) Ego Graph Contrastive Learning (EGCL) aligns fused representation with view-specific representation, enhancing similarity of samples from same cluster.
Result: Extensive experiments show MoEGCL achieves state-of-the-art results in deep multi-view clustering tasks.
Conclusion: MoEGCL addresses coarse-grained fusion limitation through sample-level ego-graph fusion and contrastive learning, significantly improving multi-view clustering performance.
Abstract: In recent years, the advancement of Graph Neural Networks (GNNs) has significantly propelled progress in Multi-View Clustering (MVC). However, existing methods face the problem of coarse-grained graph fusion. Specifically, current approaches typically generate a separate graph structure for each view and then perform weighted fusion of graph structures at the view level, which is a relatively rough strategy. To address this limitation, we present a novel Mixture of Ego-Graphs Contrastive Representation Learning (MoEGCL). It mainly consists of two modules. In particular, we propose an innovative Mixture of Ego-Graphs Fusion (MoEGF), which constructs ego graphs and utilizes a Mixture-of-Experts network to implement fine-grained fusion of ego graphs at the sample level, rather than the conventional view-level fusion. Additionally, we present the Ego Graph Contrastive Learning (EGCL) module to align the fused representation with the view-specific representation. The EGCL module enhances the representation similarity of samples from the same cluster, not merely from the same sample, further boosting fine-grained graph representation. Extensive experiments demonstrate that MoEGCL achieves state-of-the-art results in deep multi-view clustering tasks. The source code is publicly available at https://github.com/HackerHyper/MoEGCL.
[459] Chain-of-Visual-Thought: Teaching VLMs to See and Think Better with Continuous Visual Tokens
Yiming Qin, Bomin Wei, Jiaxin Ge, Konstantinos Kallidromitis, Stephanie Fu, Trevor Darrell, XuDong Wang
Main category: cs.CV
TL;DR: COVT enables VLMs to reason through continuous visual tokens, improving perceptual understanding by 3-16% across diverse benchmarks while maintaining efficiency.
Details
Motivation: Current VLMs excel at linguistic reasoning but struggle with dense visual perception tasks like spatial reasoning and geometric awareness due to limited mechanisms for capturing dense visual information across spatial dimensions.Method: Introduces Chain-of-Visual-Thought (COVT) framework that enables VLMs to reason through continuous visual tokens (compact latent representations encoding rich perceptual cues). Distills knowledge from lightweight vision experts to capture 2D appearance, 3D geometry, spatial layout, and edge structure within ~20 tokens. During training, VLMs autoregressively predict visual tokens to reconstruct dense supervision signals (depth, segmentation, edges, DINO features).
Result: Integration into strong VLMs (Qwen2.5-VL and LLaVA) consistently improves performance by 3% to 16% across more than ten diverse perception benchmarks including CV-Bench, MMVP, RealWorldQA, MMStar, WorldMedQA, and HRBench.
Conclusion: Compact continuous visual thinking enables more precise, grounded, and interpretable multimodal intelligence while preserving efficiency, with optional dense prediction decoding for interpretability.
Abstract: Vision-Language Models (VLMs) excel at reasoning in linguistic space but struggle with perceptual understanding that requires dense visual perception, e.g., spatial reasoning and geometric awareness. This limitation stems from the fact that current VLMs have limited mechanisms to capture dense visual information across spatial dimensions. We introduce Chain-of-Visual-Thought (COVT), a framework that enables VLMs to reason not only in words but also through continuous visual tokens-compact latent representations that encode rich perceptual cues. Within a small budget of roughly 20 tokens, COVT distills knowledge from lightweight vision experts, capturing complementary properties such as 2D appearance, 3D geometry, spatial layout, and edge structure. During training, the VLM with COVT autoregressively predicts these visual tokens to reconstruct dense supervision signals (e.g., depth, segmentation, edges, and DINO features). At inference, the model reasons directly in the continuous visual token space, preserving efficiency while optionally decoding dense predictions for interpretability. Evaluated across more than ten diverse perception benchmarks, including CV-Bench, MMVP, RealWorldQA, MMStar, WorldMedQA, and HRBench, integrating COVT into strong VLMs such as Qwen2.5-VL and LLaVA consistently improves performance by 3% to 16% and demonstrates that compact continuous visual thinking enables more precise, grounded, and interpretable multimodal intelligence.
[460] Physics-Informed Image Restoration via Progressive PDE Integration
Shamika Likhite, Santiago López-Tapia, Aggelos K. Katsaggelos
Main category: cs.CV
TL;DR: Physics-informed PDE dynamics integrated into deep learning architectures for motion deblurring, achieving superior restoration with minimal computational overhead.
Details
Motivation: Existing deep learning motion deblurring methods struggle with capturing long-range spatial dependencies in blur patterns. Convolutional methods have limited receptive fields and require extremely deep networks for global modeling, motivating the need for approaches that incorporate physical priors to guide feature evolution.Method: Progressive training framework that integrates physics-informed PDE dynamics into state-of-the-art restoration architectures. Uses advection-diffusion equations to model feature evolution, capturing directional flow characteristics of motion blur while enabling principled global spatial modeling.
Result: PDE-enhanced deblurring models achieve superior restoration quality with minimal overhead (only ~1% increase in inference GMACs). Consistent improvements in perceptual quality across multiple architectures (FFTformer, NAFNet, Restormer, Stripformer). Significant PSNR and SSIM improvements on standard motion deblurring benchmarks.
Conclusion: Incorporating mathematical physics principles through PDE-based global layers enhances deep learning-based image restoration. Establishes a promising direction for physics-informed neural network design in computer vision applications.
Abstract: Motion blur, caused by relative movement between camera and scene during exposure, significantly degrades image quality and impairs downstream computer vision tasks such as object detection, tracking, and recognition in dynamic environments. While deep learning-based motion deblurring methods have achieved remarkable progress, existing approaches face fundamental challenges in capturing the long-range spatial dependencies inherent in motion blur patterns. Traditional convolutional methods rely on limited receptive fields and require extremely deep networks to model global spatial relationships. These limitations motivate the need for alternative approaches that incorporate physical priors to guide feature evolution during restoration. In this paper, we propose a progressive training framework that integrates physics-informed PDE dynamics into state-of-the-art restoration architectures. By leveraging advection-diffusion equations to model feature evolution, our approach naturally captures the directional flow characteristics of motion blur while enabling principled global spatial modeling. Our PDE-enhanced deblurring models achieve superior restoration quality with minimal overhead, adding only approximately 1% to inference GMACs while providing consistent improvements in perceptual quality across multiple state-of-the-art architectures. Comprehensive experiments on standard motion deblurring benchmarks demonstrate that our physics-informed approach improves PSNR and SSIM significantly across four diverse architectures, including FFTformer, NAFNet, Restormer, and Stripformer. These results validate that incorporating mathematical physics principles through PDE-based global layers can enhance deep learning-based image restoration, establishing a promising direction for physics-informed neural network design in computer vision applications.
[461] 3D Motion Perception of Binocular Vision Target with PID-CNN
Jiazhao Shi, Pan Pan, Haotian Shi
Main category: cs.CV
TL;DR: A neural network for 3D motion perception of binocular vision targets that provides real-time 3D coordinates, velocity, and acceleration with spatiotemporal perception capabilities.
Details
Motivation: To develop a neural network capable of perceiving 3D motion information from binocular vision, providing real-time tracking of position, velocity, and acceleration, while exploring neural network design principles from a PID control perspective.Method: Designed a small PID convolutional neural network (17 layers, 413K parameters) with feature reuse via concatenation and pooling. Analyzed neural networks as second-order difference equations with nonlinearities. Trained and tested on simulated randomly moving ball datasets.
Result: Prediction accuracy approached the upper limit that input image resolution can represent. Demonstrated practical feature reuse and analyzed experimental errors, shortcomings, and improvement directions.
Conclusion: High-dimensional convolution improves computational efficiency and feature space utilization. PID information has potential advantages for implementing memory and attention mechanisms in neural networks.
Abstract: This article trained a network for perceiving three-dimensional motion information of binocular vision target, which can provide real-time three-dimensional coordinate, velocity, and acceleration, and has a basic spatiotemporal perception capability. Understood the ability of neural networks to fit nonlinear problems from the perspective of PID. Considered a single-layer neural network as using a second-order difference equation and a nonlinearity to describe a local problem. Multilayer networks gradually transform the raw representation to the desired representation through multiple such combinations. Analysed some reference principles for designing neural networks. Designed a relatively small PID convolutional neural network, with a total of 17 layers and 413 thousand parameters. Implemented a simple but practical feature reuse method by concatenation and pooling. The network was trained and tested using the simulated randomly moving ball datasets, and the experimental results showed that the prediction accuracy was close to the upper limit that the input image resolution can represent. Analysed the experimental results and errors, as well as the existing shortcomings and possible directions for improvement. Finally, discussed the advantages of high-dimensional convolution in improving computational efficiency and feature space utilization. As well as the potential advantages of using PID information to implement memory and attention mechanisms.
[462] Beyond Randomness: Understand the Order of the Noise in Diffusion
Song Yan, Min Li, Bi Xinliang, Jian Yang, Yusen Zhang, Guanye Xiong, Yunwei Lan, Tao Zhang, Wei Zhai, Zheng-Jun Zha
Main category: cs.CV
TL;DR: The paper reveals that initial noise in text-to-content diffusion models contains analyzable semantic patterns, proposes a training-free “Semantic Erasure-Injection” method to modulate noise for better generation control.
Details
Motivation: Challenge the conventional view that initial noise in T2C diffusion models is purely random; discover that noise actually contains rich semantic information that can be analyzed and manipulated for better generation control.Method: Two-step “Semantic Erasure-Injection” process: 1) Erase unwanted semantics from noise using information theory principles, 2) Inject desired semantics into cleaned noise by leveraging the equivalence between diffusion generation process and semantic injection.
Result: Method is consistently effective across various T2C models (both DiT and UNet architectures), training-free, and provides a universal tool for consistent generation optimization.
Conclusion: Initial noise in diffusion models contains analyzable semantic patterns; the proposed noise modulation method offers a novel perspective for optimizing diffusion model generation and enables universal consistent generation control.
Abstract: In text-driven content generation (T2C) diffusion model, semantic of generated content is mostly attributed to the process of text embedding and attention mechanism interaction. The initial noise of the generation process is typically characterized as a random element that contributes to the diversity of the generated content. Contrary to this view, this paper reveals that beneath the random surface of noise lies strong analyzable patterns. Specifically, this paper first conducts a comprehensive analysis of the impact of random noise on the model’s generation. We found that noise not only contains rich semantic information, but also allows for the erasure of unwanted semantics from it in an extremely simple way based on information theory, and using the equivalence between the generation process of diffusion model and semantic injection to inject semantics into the cleaned noise. Then, we mathematically decipher these observations and propose a simple but efficient training-free and universal two-step “Semantic Erasure-Injection” process to modulate the initial noise in T2C diffusion model. Experimental results demonstrate that our method is consistently effective across various T2C models based on both DiT and UNet architectures and presents a novel perspective for optimizing the generation of diffusion model, providing a universal tool for consistent generation.
[463] When Robots Obey the Patch: Universal Transferable Patch Attacks on Vision-Language-Action Models
Hui Lu, Yi Yu, Yiming Yang, Chenyu Yi, Qixin Zhang, Bingquan Shen, Alex C. Kot, Xudong Jiang
Main category: cs.CV
TL;DR: UPA-RFAS is a universal adversarial patch attack framework that transfers across different VLA models, tasks, and viewpoints by learning in a shared feature space with robust optimization and VLA-specific attention/semantic attacks.
Details
Motivation: Current adversarial attacks on Vision-Language-Action (VLA) models are limited - they overfit to single models and fail in black-box settings. There's a need for universal, transferable attacks that work across unknown architectures, finetuned variants, and sim-to-real shifts.Method: UPA-RFAS combines: 1) Feature-space objective with ℓ₁ deviation prior and repulsive InfoNCE loss for transferable representation shifts; 2) Robustness-augmented two-phase min-max procedure with inner loop for invisible perturbations and outer loop for universal patch optimization; 3) Two VLA-specific losses: Patch Attention Dominance to hijack text→vision attention and Patch Semantic Misalignment to induce image-text mismatch.
Result: Experiments show UPA-RFAS consistently transfers across diverse VLA models, manipulation suites, and physical executions. It works across models, tasks, and viewpoints, exposing a practical patch-based attack surface.
Conclusion: UPA-RFAS establishes a strong baseline for universal adversarial attacks on VLA-driven robots, demonstrating transferability across unknown architectures and sim-to-real shifts, highlighting security vulnerabilities in VLA systems.
Abstract: Vision-Language-Action (VLA) models are vulnerable to adversarial attacks, yet universal and transferable attacks remain underexplored, as most existing patches overfit to a single model and fail in black-box settings. To address this gap, we present a systematic study of universal, transferable adversarial patches against VLA-driven robots under unknown architectures, finetuned variants, and sim-to-real shifts. We introduce UPA-RFAS (Universal Patch Attack via Robust Feature, Attention, and Semantics), a unified framework that learns a single physical patch in a shared feature space while promoting cross-model transfer. UPA-RFAS combines (i) a feature-space objective with an $\ell_1$ deviation prior and repulsive InfoNCE loss to induce transferable representation shifts, (ii) a robustness-augmented two-phase min-max procedure where an inner loop learns invisible sample-wise perturbations and an outer loop optimizes the universal patch against this hardened neighborhood, and (iii) two VLA-specific losses: Patch Attention Dominance to hijack text$\to$vision attention and Patch Semantic Misalignment to induce image-text mismatch without labels. Experiments across diverse VLA models, manipulation suites, and physical executions show that UPA-RFAS consistently transfers across models, tasks, and viewpoints, exposing a practical patch-based attack surface and establishing a strong baseline for future defenses.
[464] Harnessing Diffusion-Generated Synthetic Images for Fair Image Classification
Abhipsa Basu, Aviral Gupta, Abhijnya Bhat, R. Venkatesh Babu
Main category: cs.CV
TL;DR: The paper proposes diffusion model finetuning techniques (LoRA and DreamBooth) to generate balanced training data for debiasing image classification, outperforming vanilla Stable Diffusion and matching SOTA debiasing methods.
Details
Motivation: Image classification systems inherit biases from uneven group representation in training data (e.g., blond hair disproportionately associated with females). Existing approaches using Stable Diffusion struggle to preserve original data distributions when generating balanced data.Method: Multiple diffusion-finetuning techniques: LoRA and DreamBooth, with clustering within groups to prevent DreamBooth models from being overwhelmed by intra-group variations. Generate group-balanced data for pretraining, then fine-tune on real data.
Result: Finetuning approaches outperform vanilla Stable Diffusion on average and achieve results comparable to SOTA debiasing techniques like Group-DRO, while surpassing them as dataset bias severity increases.
Conclusion: Diffusion model finetuning techniques effectively generate balanced training data for debiasing, with performance improving relative to existing methods as bias severity increases.
Abstract: Image classification systems often inherit biases from uneven group representation in training data. For example, in face datasets for hair color classification, blond hair may be disproportionately associated with females, reinforcing stereotypes. A recent approach leverages the Stable Diffusion model to generate balanced training data, but these models often struggle to preserve the original data distribution. In this work, we explore multiple diffusion-finetuning techniques, e.g., LoRA and DreamBooth, to generate images that more accurately represent each training group by learning directly from their samples. Additionally, in order to prevent a single DreamBooth model from being overwhelmed by excessive intra-group variations, we explore a technique of clustering images within each group and train a DreamBooth model per cluster. These models are then used to generate group-balanced data for pretraining, followed by fine-tuning on real data. Experiments on multiple benchmarks demonstrate that the studied finetuning approaches outperform vanilla Stable Diffusion on average and achieve results comparable to SOTA debiasing techniques like Group-DRO, while surpassing them as the dataset bias severity increases.
[465] STORM: Segment, Track, and Object Re-Localization from a Single Image
Yu Deng, Teng Cao, Hikaru Shindo, Jiahong Xue, Quentin Delfosse, Kristian Kersting
Main category: cs.CV
TL;DR: STORM is an annotation-free, real-time 6D pose estimation system that combines vision-language understanding with feature matching to track objects without manual segmentation or 3D models, featuring automatic failure recovery.
Details
Motivation: Existing 6D pose estimation methods require pre-defined 3D models and manual segmentation annotations, which are labor-intensive and perform poorly under occlusions or rapid motion. There's a need for annotation-free, robust systems that can handle real-world industrial challenges.Method: Three-stage pipeline: 1) Contextual object descriptions guide localization using vision-language understanding, 2) Self-cross-attention mechanisms identify candidate regions and produce precise masks, 3) Automatic re-registration detects tracking failures through feature similarity monitoring and recovers from severe occlusions/rapid motion.
Result: Achieves state-of-the-art accuracy on challenging industrial datasets with multi-object occlusions, high-speed motion, and varying illumination. Operates at real-time speeds without requiring additional training.
Conclusion: STORM provides a practical, annotation-free solution that significantly reduces deployment overhead for applications like flexible manufacturing and intelligent quality control, overcoming limitations of traditional pose estimation methods.
Abstract: Accurate 6D pose estimation and tracking are fundamental capabilities for physical AI systems such as robots. However, existing approaches typically require a pre-defined 3D model of the target and rely on a manually annotated segmentation mask in the first frame, which is labor-intensive and leads to reduced performance when faced with occlusions or rapid movement. To address these limitations, we propose STORM (Segment, Track, and Object Re-localization from a single iMage), an open-source robust real-time 6D pose estimation system that requires no manual annotation. STORM employs a novel three-stage pipeline combining vision-language understanding with feature matching: contextual object descriptions guide localization, self-cross-attention mechanisms identify candidate regions, and produce precise masks and 3D models for accurate pose estimation. Another key innovation is our automatic re-registration mechanism that detects tracking failures through feature similarity monitoring and recovers from severe occlusions or rapid motion. STORM achieves state-of-the-art accuracy on challenging industrial datasets featuring multi-object occlusions, high-speed motion, and varying illumination, while operating at real-time speeds without additional training. This annotation-free approach significantly reduces deployment overhead, providing a practical solution for modern applications, such as flexible manufacturing and intelligent quality control.
[466] LiNeXt: Revisiting LiDAR Completion with Efficient Non-Diffusion Architectures
Wenzhe He, Xiaojun Chen, Ruiqi Wang, Ruihui Li, Huilong Pi, Jiapeng Zhang, Zhuo Tang, Kenli Li
Main category: cs.CV
TL;DR: LiNeXt: A lightweight non-diffusion network for real-time 3D LiDAR scene completion that achieves 199.8x faster inference than diffusion-based methods while improving accuracy.
Details
Motivation: Previous diffusion-based methods for 3D LiDAR scene completion suffer from significant computational overhead due to multi-step iterative sampling, limiting real-time applicability in autonomous vehicles.Method: Two-stage approach: 1) Noise-to-Coarse (N2C) Module denoises input point clouds in single pass, eliminating iterative sampling; 2) Refine Module uses coarse output and intermediate features for precise refinement. Also introduces Distance-aware Selected Repeat strategy to handle LiDAR’s distance-dependent spatial distribution.
Result: On SemanticKITTI: 199.8x inference speedup, 50.7% reduction in Chamfer Distance, uses only 6.1% of parameters compared to LiDiff diffusion model.
Conclusion: LiNeXt demonstrates superior efficiency and effectiveness for real-time scene completion, making it practical for autonomous vehicle perception systems.
Abstract: 3D LiDAR scene completion from point clouds is a fundamental component of perception systems in autonomous vehicles. Previous methods have predominantly employed diffusion models for high-fidelity reconstruction. However, their multi-step iterative sampling incurs significant computational overhead, limiting its real-time applicability. To address this, we propose LiNeXt-a lightweight, non-diffusion network optimized for rapid and accurate point cloud completion. Specifically, LiNeXt first applies the Noise-to-Coarse (N2C) Module to denoise the input noisy point cloud in a single pass, thereby obviating the multi-step iterative sampling of diffusion-based methods. The Refine Module then takes the coarse point cloud and its intermediate features from the N2C Module to perform more precise refinement, further enhancing structural completeness. Furthermore, we observe that LiDAR point clouds exhibit a distance-dependent spatial distribution, being densely sampled at proximal ranges and sparsely sampled at distal ranges. Accordingly, we propose the Distance-aware Selected Repeat strategy to generate a more uniformly distributed noisy point cloud. On the SemanticKITTI dataset, LiNeXt achieves a 199.8x speedup in inference, reduces Chamfer Distance by 50.7%, and uses only 6.1% of the parameters compared with LiDiff. These results demonstrate the superior efficiency and effectiveness of LiNeXt for real-time scene completion.
[467] GFT: Graph Feature Tuning for Efficient Point Cloud Analysis
Manish Dhakal, Venkat R. Dasari, Rajshekhar Sunderraman, Yi Ding
Main category: cs.CV
TL;DR: GFT is a parameter-efficient fine-tuning method for point cloud transformers that uses dynamic graph features with skip connections and cross-attention, achieving competitive performance with fewer trainable parameters.
Details
Motivation: General PEFT methods are suboptimal for point cloud data, and there's a need to further reduce trainable parameters while maintaining performance for point cloud tasks.Method: GFT learns a dynamic graph from initial tokenized inputs using lightweight graph convolution, then passes graph features to deeper layers via skip connections and efficient cross-attention modules.
Result: Extensive experiments on object classification and segmentation show GFT rivals existing methods while reducing trainable parameters.
Conclusion: GFT provides an effective point-cloud-specific PEFT approach that significantly reduces parameter count while maintaining competitive performance on 3D vision tasks.
Abstract: Parameter-efficient fine-tuning (PEFT) significantly reduces computational and memory costs by updating only a small subset of the model’s parameters, enabling faster adaptation to new tasks with minimal loss in performance. Previous studies have introduced PEFTs tailored for point cloud data, as general approaches are suboptimal. To further reduce the number of trainable parameters, we propose a point-cloud-specific PEFT, termed Graph Features Tuning (GFT), which learns a dynamic graph from initial tokenized inputs of the transformer using a lightweight graph convolution network and passes these graph features to deeper layers via skip connections and efficient cross-attention modules. Extensive experiments on object classification and segmentation tasks show that GFT operates in the same domain, rivalling existing methods, while reducing the trainable parameters. Code is available at https://github.com/manishdhakal/GFT.
[468] Explainable Deep Convolutional Multi-Type Anomaly Detection
Alex George, Lyudmila Mihaylova, Sean Anderson
Main category: cs.CV
TL;DR: MultiTypeFCDD is a lightweight convolutional framework for explainable multi-type anomaly detection that uses image-level labels to produce multi-channel heatmaps for different anomaly types, achieving competitive performance at just 1% the size of VLM models.
Details
Motivation: Current explainable anomaly detection methods can localize anomalies but cannot differentiate between anomaly types (e.g., "Crack" vs. "Scratch"), which is crucial for accurate diagnosis and operational decisions. Existing methods also require separate models for each object category, and recent VLMs are too computationally intensive for real-time or embedded systems.Method: MultiTypeFCDD uses a simple convolutional framework that learns from image-level labels to produce multi-channel heatmaps, where each channel corresponds to a specific anomaly type. It functions as a single unified framework capable of differentiating anomaly types across multiple object categories without needing separate models for each category.
Result: The method achieves 96.4% I-AUROC on the Real-IAD dataset while being only about 1% the size of state-of-the-art VLM models used for similar tasks, making it highly practical for resource-constrained applications.
Conclusion: MultiTypeFCDD provides a practical, lightweight alternative for explainable multi-type anomaly detection that addresses the limitations of existing methods by offering type differentiation, unified multi-category capability, and computational efficiency suitable for real-world applications.
Abstract: Explainable anomaly detection methods often have the capability to identify and spatially localise anomalies within an image but lack the capability to differentiate the type of anomaly. Furthermore, they often require the costly training and maintenance of separate models for each object category. The lack of specificity is a significant research gap because identifying the type of anomaly (e.g., “Crack” vs. “Scratch”) is crucial for accurate diagnosis that facilitates cost-saving operational decisions across diverse application domains. While some recent large-scale Vision-Language Models (VLMs) have begun to address this, they are computationally intensive and memory-heavy, restricting their use in real-time or embedded systems. We propose MultiTypeFCDD, a simple and lightweight convolutional framework designed as a practical alternative for explainable multi-type anomaly detection. MultiTypeFCDD uses only image-level labels to learn and produce multi-channel heatmaps, where each channel is trained to correspond to a specific anomaly type. The model functions as a single, unified framework capable of differentiating anomaly types across multiple object categories, eliminating the need to train and manage separate models for each object category. We evaluated our proposed method on the Real-IAD dataset and it delivers competitive results (96.4% I-AUROC) at just over 1% the size of state-of-the-art VLM models used for similar tasks. This makes it a highly practical and viable solution for real-world applications where computational resources are tightly constrained.
[469] CountSteer: Steering Attention for Object Counting in Diffusion Models
Hyemin Boo, Hyoryung Kim, Myungjin Lee, Seunghyeon Lee, Jiyoung Lee, Jang-Hwan Choi, Hyunsoo Cho
Main category: cs.CV
TL;DR: CountSteer is a training-free method that improves text-to-image models’ ability to follow numerical instructions by steering cross-attention hidden states based on the model’s internal awareness of counting accuracy.
Details
Motivation: Text-to-image diffusion models often fail to follow numerical instructions in text prompts, revealing a gap between language and visual representation. However, the authors discovered that these models are implicitly aware of their own counting accuracy through internal signal shifts.Method: CountSteer is a training-free method that improves object count accuracy by steering the model’s cross-attention hidden states during inference. It leverages the model’s latent notion of numerical correctness by observing how internal signals shift depending on whether outputs meet specified counts.
Result: CountSteer improved object-count accuracy by about 4% without compromising visual quality. The method demonstrates that models already encode latent numerical correctness that can be harnessed for better control.
Conclusion: CountSteer represents a simple yet effective step toward more controllable and semantically reliable text-to-image generation by exploiting the model’s existing awareness of numerical correctness through cross-attention steering.
Abstract: Text-to-image diffusion models generate realistic and coherent images but often fail to follow numerical instructions in text, revealing a gap between language and visual representation. Interestingly, we found that these models are not entirely blind to numbers-they are implicitly aware of their own counting accuracy, as their internal signals shift in consistent ways depending on whether the output meets the specified count. This observation suggests that the model already encodes a latent notion of numerical correctness, which can be harnessed to guide generation more precisely. Building on this intuition, we introduce CountSteer, a training-free method that improves generation of specified object counts by steering the model’s cross-attention hidden states during inference. In our experiments, CountSteer improved object-count accuracy by about 4% without compromising visual quality, demonstrating a simple yet effective step toward more controllable and semantically reliable text-to-image generation.
[470] Rethinking Efficient Mixture-of-Experts for Remote Sensing Modality-Missing Classification
Qinghao Gao, Jiahui Qu, Yunsong Li, Wenqian Dong
Main category: cs.CV
TL;DR: MaMOL is a parameter-efficient framework for incomplete multimodal learning that treats modality missing as multi-task learning, using dual-routing to adapt to different missing patterns while maintaining cross-modal knowledge sharing.
Details
Motivation: Multimodal classification in remote sensing suffers from missing modalities due to environmental interference, sensor failures, or atmospheric effects, which degrades performance. Existing methods are computationally expensive and assume complete multimodal data during training, limiting real-world applicability.Method: Proposes Missing-aware Mixture-of-Loras (MaMOL) framework with dual-routing mechanism: 1) task-oriented dynamic router that adaptively activates experts for different missing patterns, and 2) modality-specific-shared static router that maintains stable cross-modal knowledge sharing. Uses lightweight expert updates and shared expert reuse for parameter-efficient adaptation.
Result: Experiments on multiple remote sensing benchmarks show superior robustness and generalization under varying missing rates with minimal computational overhead. Transfer experiments on natural image datasets validate scalability and cross-domain applicability.
Conclusion: MaMOL provides a general and efficient solution for incomplete multimodal learning that outperforms prior methods while being computationally efficient and transferable across domains.
Abstract: Multimodal classification in remote sensing often suffers from missing modalities caused by environmental interference, sensor failures, or atmospheric effects, which severely degrade classification performance. Existing two-stage adaptation methods are computationally expensive and assume complete multimodal data during training, limiting their generalization to real-world incompleteness. To overcome these issues, we propose a Missing-aware Mixture-of-Loras (MaMOL) framework that reformulates modality missing as a multi-task learning problem. MaMOL introduces a dual-routing mechanism: a task-oriented dynamic router that adaptively activates experts for different missing patterns, and a modality-specific-shared static router that maintains stable cross-modal knowledge sharing. Unlike prior methods that train separate networks for each missing configuration, MaMOL achieves parameter-efficient adaptation via lightweight expert updates and shared expert reuse. Experiments on multiple remote sensing benchmarks demonstrate superior robustness and generalization under varying missing rates, with minimal computational overhead. Moreover, transfer experiments on natural image datasets validate its scalability and cross-domain applicability, highlighting MaMOL as a general and efficient solution for incomplete multimodal learning.
[471] Learning to Hear by Seeing: It’s Time for Vision Language Models to Understand Artistic Emotion from Sight and Sound
Dengming Zhang, Weitao You, Jingxiong Li, Weishen Lin, Wenda Shi, Xue Zhao, Heda Zuo, Junxian Wu, Lingyun Sun
Main category: cs.CV
TL;DR: VAEmotionLLM is a two-stage framework that enables Visual Language Models to understand emotion across audio-visual modalities with limited audio pretraining, achieving SOTA on art-centric emotion understanding.
Details
Motivation: Current AVLMs require large-scale audio pretraining, limiting scalability. Most prior work overlooks emotion intentionally expressed in artworks, which convey emotion through joint visual-auditory design.Method: Two-stage framework: 1) VG-Align distills visual pathway into audio pathway by aligning next-token distributions on synchronized audio-video clips. 2) EmoAdapter injects emotion-sensitive residuals and applies emotion supervision for cross-modal emotion understanding.
Result: Achieves state-of-the-art results on ArtEmoBenchmark, outperforming audio-only, visual-only, and audio-visual baselines. Ablations show proposed components are complementary.
Conclusion: VAEmotionLLM enables efficient audio-visual emotion understanding with limited audio pretraining, addressing scalability limitations while enhancing emotion comprehension in multimodal LLMs.
Abstract: Emotion understanding is critical for making Large Language Models (LLMs) more general, reliable, and aligned with humans. Art conveys emotion through the joint design of visual and auditory elements, yet most prior work is human-centered or single-modality, overlooking the emotion intentionally expressed by the artwork. Meanwhile, current Audio-Visual Language Models (AVLMs) typically require large-scale audio pretraining to endow Visual Language Models (VLMs) with hearing, which limits scalability. We present Vision Anchored Audio-Visual Emotion LLM (VAEmotionLLM), a two-stage framework that teaches a VLM to hear by seeing with limited audio pretraining and to understand emotion across modalities. In Stage 1, Vision-Guided Audio Alignment (VG-Align) distills the frozen visual pathway into a new audio pathway by aligning next-token distributions of the shared LLM on synchronized audio-video clips, enabling hearing without a large audio dataset. In Stage 2, a lightweight Cross-Modal Emotion Adapter (EmoAdapter), composed of the Emotion Enhancer and the Emotion Supervisor, injects emotion-sensitive residuals and applies emotion supervision to enhance cross-modal emotion understanding. We also construct ArtEmoBenchmark, an art-centric emotion benchmark that evaluates content and emotion understanding under audio-only, visual-only, and audio-visual inputs. VAEmotionLLM achieves state-of-the-art results on ArtEmoBenchmark, outperforming audio-only, visual-only, and audio-visual baselines. Ablations show that the proposed components are complementary.
[472] Bridging Granularity Gaps: Hierarchical Semantic Learning for Cross-domain Few-shot Segmentation
Sujun Sun, Haowen Gu, Cheng Xie, Yanxu Ren, Mingwu Ren, Haofeng Zhang
Main category: cs.CV
TL;DR: A Hierarchical Semantic Learning framework for Cross-domain Few-shot Segmentation that addresses segmentation granularity gaps through style randomization and multi-scale semantic mining.
Details
Motivation: Existing CD-FSS methods focus only on style gaps between source and target domains while ignoring segmentation granularity gaps, leading to insufficient semantic discriminability for novel classes in target domains.Method: Proposes HSL framework with three modules: 1) Dual Style Randomization (DSR) simulates target domain data with diverse style variations; 2) Hierarchical Semantic Mining (HSM) uses multi-scale superpixels to mine intra-class consistency and inter-class distinction at different granularities; 3) Prototype Confidence-modulated Thresholding (PCMT) mitigates segmentation ambiguity when foreground and background are similar.
Result: Extensive experiments on four popular target domain datasets demonstrate state-of-the-art performance.
Conclusion: The proposed HSL framework effectively addresses both style and granularity gaps in CD-FSS, enhancing semantic discriminability for novel classes in target domains through hierarchical semantic learning.
Abstract: Cross-domain Few-shot Segmentation (CD-FSS) aims to segment novel classes from target domains that are not involved in training and have significantly different data distributions from the source domain, using only a few annotated samples, and recent years have witnessed significant progress on this task. However, existing CD-FSS methods primarily focus on style gaps between source and target domains while ignoring segmentation granularity gaps, resulting in insufficient semantic discriminability for novel classes in target domains. Therefore, we propose a Hierarchical Semantic Learning (HSL) framework to tackle this problem. Specifically, we introduce a Dual Style Randomization (DSR) module and a Hierarchical Semantic Mining (HSM) module to learn hierarchical semantic features, thereby enhancing the model’s ability to recognize semantics at varying granularities. DSR simulates target domain data with diverse foreground-background style differences and overall style variations through foreground and global style randomization respectively, while HSM leverages multi-scale superpixels to guide the model to mine intra-class consistency and inter-class distinction at different granularities. Additionally, we also propose a Prototype Confidence-modulated Thresholding (PCMT) module to mitigate segmentation ambiguity when foreground and background are excessively similar. Extensive experiments are conducted on four popular target domain datasets, and the results demonstrate that our method achieves state-of-the-art performance.
[473] HiGFA: Hierarchical Guidance for Fine-grained Data Augmentation with Diffusion Models
Zhiguang Lu, Qianqian Xu, Peisong Wen, Siran Dai, Qingming Huang
Main category: cs.CV
TL;DR: HiGFA is a diffusion-based data augmentation method that uses hierarchical guidance with confidence-based modulation to generate high-fidelity synthetic images for fine-grained visual classification tasks.
Details
Motivation: Standard diffusion models with text-based guidance (like CFG) lack specificity for fine-grained tasks, often generating misleading examples that degrade classifier performance due to inaccurate capture of subtle, category-defining features.Method: HiGFA leverages diffusion sampling dynamics with hierarchical guidance: early-to-mid stages use strong text and transformed contour guidance to establish scene, style, and structure; final stages activate specialized fine-grained classifier guidance and dynamically modulate all guidance strengths based on prediction confidence.
Result: Experiments on several Fine-Grained Visual Classification (FGVC) datasets demonstrate the effectiveness of HiGFA in generating diverse yet faithful synthetic images for data augmentation.
Conclusion: HiGFA’s hierarchical, confidence-driven orchestration successfully balances global structure formation with precise detail refinement, addressing the challenge of applying diffusion models to fine-grained tasks.
Abstract: Generative diffusion models show promise for data augmentation. However, applying them to fine-grained tasks presents a significant challenge: ensuring synthetic images accurately capture the subtle, category-defining features critical for high fidelity. Standard approaches, such as text-based Classifier-Free Guidance (CFG), often lack the required specificity, potentially generating misleading examples that degrade fine-grained classifier performance. To address this, we propose Hierarchically Guided Fine-grained Augmentation (HiGFA). HiGFA leverages the temporal dynamics of the diffusion sampling process. It employs strong text and transformed contour guidance with fixed strengths in the early-to-mid sampling stages to establish overall scene, style, and structure. In the final sampling stages, HiGFA activates a specialized fine-grained classifier guidance and dynamically modulates the strength of all guidance signals based on prediction confidence. This hierarchical, confidence-driven orchestration enables HiGFA to generate diverse yet faithful synthetic images by intelligently balancing global structure formation with precise detail refinement. Experiments on several FGVC datasets demonstrate the effectiveness of HiGFA.
[474] TempoMaster: Efficient Long Video Generation via Next-Frame-Rate Prediction
Yukuo Ma, Cong Liu, Junke Wang, Junqi Liu, Haibin Huang, Zuxuan Wu, Chi Zhang, Xuelong Li
Main category: cs.CV
TL;DR: TempoMaster is a novel framework for long video generation that treats it as next-frame-rate prediction, starting with low-frame-rate clips as blueprints and progressively increasing frame rates for refinement.
Details
Motivation: The paper addresses the challenge of generating long videos with both temporal coherence and visual quality. Traditional methods struggle with maintaining consistency over long sequences while producing high-quality frames efficiently.Method: TempoMaster formulates video generation as next-frame-rate prediction: 1) Generate low-frame-rate clip as coarse blueprint, 2) Progressively increase frame rate to refine details and motion, 3) Use bidirectional attention within each frame-rate level, 4) Perform autoregression across frame rates.
Result: Extensive experiments show TempoMaster establishes new state-of-the-art in long video generation, excelling in both visual quality and temporal coherence while enabling efficient parallel synthesis.
Conclusion: The TempoMaster framework successfully addresses long video generation challenges through its novel next-frame-rate prediction approach, achieving superior performance in temporal coherence and visual quality with efficient parallel synthesis.
Abstract: We present TempoMaster, a novel framework that formulates long video generation as next-frame-rate prediction. Specifically, we first generate a low-frame-rate clip that serves as a coarse blueprint of the entire video sequence, and then progressively increase the frame rate to refine visual details and motion continuity. During generation, TempoMaster employs bidirectional attention within each frame-rate level while performing autoregression across frame rates, thus achieving long-range temporal coherence while enabling efficient and parallel synthesis. Extensive experiments demonstrate that TempoMaster establishes a new state-of-the-art in long video generation, excelling in both visual and temporal quality.
[475] Can World Simulators Reason? Gen-ViRe: A Generative Visual Reasoning Benchmark
Xinxin Liu, Zhaopan Xu, Ming Li, Kai Wang, Yong Jae Lee, Yuzhang Shang
Main category: cs.CV
TL;DR: Gen-ViRe is a new benchmark framework that quantitatively evaluates video generation models’ Chain-of-Frames reasoning capabilities across six cognitive dimensions, revealing gaps between visual quality and actual reasoning depth.
Details
Motivation: Current benchmarks for video generation models focus on fidelity and alignment but fail to assess Chain-of-Frames (CoF) reasoning - the ability to simulate continuous, physics-governed dynamics through visual sequences. This evaluation gap prevents systematic understanding of model capabilities and principled guidance for improvement.Method: Gen-ViRe decomposes CoF reasoning into six cognitive dimensions (from perceptual logic to abstract planning) and 24 subtasks. It uses multi-source data curation, minimal prompting protocols, and hybrid VLM-assisted evaluation with detailed criteria to provide quantitative assessment.
Result: Experiments on state-of-the-art systems reveal substantial discrepancies between impressive visual quality and actual reasoning depth, establishing baselines and diagnostic tools for advancing genuine world simulators.
Conclusion: Gen-ViRe provides the first quantitative framework to assess video models as reasoners, enabling systematic evaluation of core cognitive abilities in multi-step planning, algorithmic logic, and abstract pattern extrapolation for developing better world simulators.
Abstract: While Chain-of-Thought (CoT) prompting enables sophisticated symbolic reasoning in LLMs, it remains confined to discrete text and cannot simulate the continuous, physics-governed dynamics of the real world. Recent video generation models have emerged as potential world simulators through Chain-of-Frames (CoF) reasoning – materializing thought as frame-by-frame visual sequences, with each frame representing a physically-grounded reasoning step. Despite compelling demonstrations, a challenge persists: existing benchmarks, focusing on fidelity or alignment, do not assess CoF reasoning and thus cannot measure core cognitive abilities in multi-step planning, algorithmic logic, or abstract pattern extrapolation. This evaluation void prevents systematic understanding of model capabilities and principled guidance for improvement. We introduce Gen-ViRe (Generative Visual Reasoning Benchmark), a framework grounded in cognitive science and real-world AI applications, which decomposes CoF reasoning into six cognitive dimensions – from perceptual logic to abstract planning – and 24 subtasks. Through multi-source data curation, minimal prompting protocols, and hybrid VLM-assisted evaluation with detailed criteria, Gen-ViRe delivers the first quantitative assessment of video models as reasoners. Our experiments on SOTA systems reveal substantial discrepancies between impressive visual quality and actual reasoning depth, establishing baselines and diagnostic tools to advance genuine world simulators.
[476] FreeSwim: Revisiting Sliding-Window Attention Mechanisms for Training-Free Ultra-High-Resolution Video Generation
Yunfeng Wu, Jiayi Song, Zhenxiong Tan, Zihao He, Songhua Liu
Main category: cs.CV
TL;DR: Training-free method for ultra-high-resolution video generation using pretrained video Diffusion Transformers with inward sliding window attention and cross-attention override for global coherence.
Details
Motivation: Quadratic time/memory complexity of attention in Transformer-based video generators makes end-to-end training for ultra-high-resolution videos prohibitively expensive, requiring a training-free solution.Method: Uses inward sliding window attention to maintain training-scale receptive field, with dual-path pipeline featuring cross-attention override strategy for global consistency, plus cross-attention caching for efficiency.
Result: Generates ultra-high-resolution videos with fine-grained details and high efficiency in training-free paradigm, achieves superior VBench performance compared to training-based alternatives with competitive efficiency.
Conclusion: Proposes effective training-free approach for high-resolution video generation that preserves visual fidelity while ensuring global coherence through novel attention mechanisms.
Abstract: The quadratic time and memory complexity of the attention mechanism in modern Transformer based video generators makes end-to-end training for ultra high resolution videos prohibitively expensive. Motivated by this limitation, we introduce a training-free approach that leverages video Diffusion Transformers pretrained at their native scale to synthesize higher resolution videos without any additional training or adaptation. At the core of our method lies an inward sliding window attention mechanism, which originates from a key observation: maintaining each query token’s training scale receptive field is crucial for preserving visual fidelity and detail. However, naive local window attention, unfortunately, often leads to repetitive content and exhibits a lack of global coherence in the generated results. To overcome this challenge, we devise a dual-path pipeline that backs up window attention with a novel cross-attention override strategy, enabling the semantic content produced by local attention to be guided by another branch with a full receptive field and, therefore, ensuring holistic consistency. Furthermore, to improve efficiency, we incorporate a cross-attention caching strategy for this branch to avoid the frequent computation of full 3D attention. Extensive experiments demonstrate that our method delivers ultra-high-resolution videos with fine-grained visual details and high efficiency in a training-free paradigm. Meanwhile, it achieves superior performance on VBench, even compared to training-based alternatives, with competitive or improved efficiency. Codes are available at: https://github.com/WillWu111/FreeSwim
[477] Mixture of Ranks with Degradation-Aware Routing for One-Step Real-World Image Super-Resolution
Xiao He, Zhijun Tu, Kun Cheng, Mingrui Zhu, Jie Hu, Nannan Wang, Xinbo Gao
Main category: cs.CV
TL;DR: Proposes MoR (Mixture-of-Ranks) architecture for single-step Real-ISR using sparse MoE with fine-grained expert partitioning of LoRA ranks, degradation-aware routing, and load balancing.
Details
Motivation: Existing dense Real-ISR models using LoRA fine-tuning cannot adaptively capture heterogeneous degradation characteristics or enable efficient knowledge sharing between inputs under computational constraints.Method: 1) MoR architecture treats each LoRA rank as independent expert with fixed-position shared experts; 2) Degradation estimation module using CLIP embeddings and text pairs; 3) Zero-expert slots and degradation-aware load-balancing loss for dynamic expert activation.
Result: Comprehensive experiments validate framework effectiveness and state-of-the-art performance in real-world image super-resolution.
Conclusion: MoR architecture successfully integrates sparse MoE into Real-ISR, enabling adaptive degradation handling, flexible knowledge recombination, and optimal computational resource allocation.
Abstract: The demonstrated success of sparsely-gated Mixture-of-Experts (MoE) architectures, exemplified by models such as DeepSeek and Grok, has motivated researchers to investigate their adaptation to diverse domains. In real-world image super-resolution (Real-ISR), existing approaches mainly rely on fine-tuning pre-trained diffusion models through Low-Rank Adaptation (LoRA) module to reconstruct high-resolution (HR) images. However, these dense Real-ISR models are limited in their ability to adaptively capture the heterogeneous characteristics of complex real-world degraded samples or enable knowledge sharing between inputs under equivalent computational budgets. To address this, we investigate the integration of sparse MoE into Real-ISR and propose a Mixture-of-Ranks (MoR) architecture for single-step image super-resolution. We introduce a fine-grained expert partitioning strategy that treats each rank in LoRA as an independent expert. This design enables flexible knowledge recombination while isolating fixed-position ranks as shared experts to preserve common-sense features and minimize routing redundancy. Furthermore, we develop a degradation estimation module leveraging CLIP embeddings and predefined positive-negative text pairs to compute relative degradation scores, dynamically guiding expert activation. To better accommodate varying sample complexities, we incorporate zero-expert slots and propose a degradation-aware load-balancing loss, which dynamically adjusts the number of active experts based on degradation severity, ensuring optimal computational resource allocation. Comprehensive experiments validate our framework’s effectiveness and state-of-the-art performance.
[478] Off the Planckian Locus: Using 2D Chromaticity to Improve In-Camera Color
SaiKiran Tedla, Joshua E. Little, Hakki Can Karaimer, Michael S. Brown
Main category: cs.CV
TL;DR: Transition from 1D CCT to 2D chromaticity space with MLP improves colorimetric accuracy for non-Planckian LED lighting, reducing error by 22% while maintaining real-time performance.
Details
Motivation: Traditional CCT-based colorimetric mapping fails with modern LED lighting that deviates from Planckian locus, requiring better illumination characterization.Method: Replace CCT interpolation with lightweight MLP using 2D chromaticity features, trained via lightbox calibration with representative LED sources.
Result: 22% average reduction in angular reproduction error for LED-lit scenes, maintains backward compatibility, handles multi-illuminant scenes, and supports real-time deployment.
Conclusion: 2D chromaticity space with MLP provides superior colorimetric accuracy for modern lighting while preserving computational efficiency and compatibility.
Abstract: Traditional in-camera colorimetric mapping relies on correlated color temperature (CCT)-based interpolation between pre-calibrated transforms optimized for Planckian illuminants such as CIE A and D65. However, modern lighting technologies such as LEDs can deviate substantially from the Planckian locus, exposing the limitations of relying on conventional one-dimensional CCT for illumination characterization. This paper demonstrates that transitioning from 1D CCT (on the Planckian locus) to a 2D chromaticity space (off the Planckian locus) improves colorimetric accuracy across various mapping approaches. In addition, we replace conventional CCT interpolation with a lightweight multi-layer perceptron (MLP) that leverages 2D chromaticity features for robust colorimetric mapping under non-Planckian illuminants. A lightbox-based calibration procedure incorporating representative LED sources is used to train our MLP. Validated across diverse LED lighting, our method reduces angular reproduction error by 22% on average in LED-lit scenes, maintains backward compatibility with traditional illuminants, accommodates multi-illuminant scenes, and supports real-time in-camera deployment with negligible additional computational cost.
[479] Hierarchical Semi-Supervised Active Learning for Remote Sensing
Wei Huang, Zhitong Xiong, Chenying Liu, Xiao Xiang Zhu
Main category: cs.CV
TL;DR: Proposes HSSAL framework combining semi-supervised learning and hierarchical active learning for remote sensing scene classification with limited labeled data.
Details
Motivation: Collecting large-scale labeled data for remote sensing is costly and time-consuming, while vast amounts of unlabeled imagery remain underutilized. Need efficient methods to leverage both labeled and unlabeled data.Method: Hierarchical Semi-Supervised Active Learning (HSSAL) integrates SSL and hierarchical active learning in iterative loop. SSL refines model using labeled data (supervised) and unlabeled data (weak-to-strong self-training). Hierarchical AL uses progressive clustering to select most informative samples based on scalability, diversity, and uncertainty.
Result: HSSAL outperforms SSL- or AL-only baselines on three RS datasets (UCM, AID, NWPU-RESISC45). With only 8%, 4%, and 2% labeled data respectively, achieves over 95% of fully-supervised accuracy.
Conclusion: HSSAL framework demonstrates superior label efficiency by effectively exploiting informativeness of unlabeled data through integrated SSL and hierarchical AL approach.
Abstract: The performance of deep learning models in remote sensing (RS) strongly depends on the availability of high-quality labeled data. However, collecting large-scale annotations is costly and time-consuming, while vast amounts of unlabeled imagery remain underutilized. To address this challenge, we propose a Hierarchical Semi-Supervised Active Learning (HSSAL) framework that integrates semi-supervised learning (SSL) and a novel hierarchical active learning (HAL) in a closed iterative loop. In each iteration, SSL refines the model using both labeled data through supervised learning and unlabeled data via weak-to-strong self-training, improving feature representation and uncertainty estimation. Guided by the refined representations and uncertainty cues of unlabeled samples, HAL then conducts sample querying through a progressive clustering strategy, selecting the most informative instances that jointly satisfy the criteria of scalability, diversity, and uncertainty. This hierarchical process ensures both efficiency and representativeness in sample selection. Extensive experiments on three benchmark RS scene classification datasets, including UCM, AID, and NWPU-RESISC45, demonstrate that HSSAL consistently outperforms SSL- or AL-only baselines. Remarkably, with only 8%, 4%, and 2% labeled training data on UCM, AID, and NWPU-RESISC45, respectively, HSSAL achieves over 95% of fully-supervised accuracy, highlighting its superior label efficiency through informativeness exploitation of unlabeled data. Our code will be publicly available.
[480] GigaWorld-0: World Models as Data Engine to Empower Embodied AI
GigaWorld Team, Angen Ye, Boyuan Wang, Chaojun Ni, Guan Huang, Guosheng Zhao, Haoyun Li, Jiagang Zhu, Kerui Li, Mengyuan Xu, Qiuping Deng, Siting Wang, Wenkang Qin, Xinze Chen, Xiaofeng Wang, Yankai Wang, Yu Cao, Yifan Chang, Yuan Xu, Yun Ye, Yang Wang, Yukun Zhou, Zhengyuan Zhang, Zhehao Dong, Zheng Zhu
Main category: cs.CV
TL;DR: GigaWorld-0 is a unified world model framework that generates high-quality embodied interaction data for training Vision-Language-Action models, enabling strong real-world robot performance without real-world training.
Details
Motivation: World models are becoming essential for scalable, data-efficient embodied AI, but generating diverse, physically realistic, and instruction-aligned training data remains challenging. The authors aim to create a unified framework that can produce high-quality embodied interaction data to train VLA models effectively.Method: GigaWorld-0 combines two synergistic components: GigaWorld-0-Video (large-scale video generation with fine-grained control) and GigaWorld-0-3D (3D generative modeling with Gaussian Splatting reconstruction, system identification, and motion planning). The framework uses joint optimization and efficient training via GigaTrain with FP8-precision and sparse attention.
Result: The system generates high-quality, diverse, and controllable embodied interaction data across multiple dimensions. VLA models (like GigaBrain-0) trained on this synthetic data achieve strong real-world performance, significantly improving generalization and task success on physical robots without any real-world interaction during training.
Conclusion: GigaWorld-0 demonstrates that unified world models can serve as effective data engines for VLA learning, enabling scalable synthesis of visually compelling, spatially coherent, physically plausible, and instruction-aligned training data that transfers successfully to real-world robotic tasks.
Abstract: World models are emerging as a foundational paradigm for scalable, data-efficient embodied AI. In this work, we present GigaWorld-0, a unified world model framework designed explicitly as a data engine for Vision-Language-Action (VLA) learning. GigaWorld-0 integrates two synergistic components: GigaWorld-0-Video, which leverages large-scale video generation to produce diverse, texture-rich, and temporally coherent embodied sequences under fine-grained control of appearance, camera viewpoint, and action semantics; and GigaWorld-0-3D, which combines 3D generative modeling, 3D Gaussian Splatting reconstruction, physically differentiable system identification, and executable motion planning to ensure geometric consistency and physical realism. Their joint optimization enables the scalable synthesis of embodied interaction data that is visually compelling, spatially coherent, physically plausible, and instruction-aligned. Training at scale is made feasible through our efficient GigaTrain framework, which exploits FP8-precision and sparse attention to drastically reduce memory and compute requirements. We conduct comprehensive evaluations showing that GigaWorld-0 generates high-quality, diverse, and controllable data across multiple dimensions. Critically, VLA model (e.g., GigaBrain-0) trained on GigaWorld-0-generated data achieve strong real-world performance, significantly improving generalization and task success on physical robots without any real-world interaction during training.
[481] VeriSciQA: An Auto-Verified Dataset for Scientific Visual Question Answering
Yuyi Li, Daoyuan Chen, Zhen Wang, Yutong Lu, Yaliang Li
Main category: cs.CV
TL;DR: Researchers created VeriSciQA, a high-quality scientific visual question answering dataset using a Generate-then-Verify framework to overcome limitations of existing LVLM-generated data, achieving better model performance than existing datasets.
Details
Motivation: Open-source Large Vision-Language Models struggle with Scientific Visual Question Answering due to lack of large-scale, high-quality datasets. Existing LVLM-generated datasets have systematic errors from model limitations and information asymmetry between figures and text.Method: Proposed a verification-centric Generate-then-Verify framework: first generates QA pairs with figure-associated textual context, then applies cross-modal consistency checks against figures along with auxiliary filters to eliminate erroneous pairs.
Result: Created VeriSciQA dataset with 20,351 QA pairs spanning 20 scientific domains and 12 figure types. It poses a challenging benchmark (64% accuracy for open-source vs 82% for proprietary models). Models fine-tuned on VeriSciQA achieve consistent improvements on SVQA benchmarks with scalable performance gains.
Conclusion: The verification-centric framework successfully creates high-quality SVQA data, and continued data expansion through this scalable approach can advance SVQA capability in the open-source community.
Abstract: Large Vision-Language Models (LVLMs) show promise for scientific applications, yet open-source models still struggle with Scientific Visual Question Answering (SVQA), namely answering questions about figures from scientific papers. A key bottleneck lies in the lack of public, large-scale, high-quality SVQA datasets. Although recent work uses LVLMs to synthesize data at scale, we identify systematic errors in their resulting QA pairs, stemming from LVLMs’ inherent limitations and information asymmetry between figures and text. To address these challenges, we propose a verification-centric Generate-then-Verify framework that first generates QA pairs with figure-associated textual context, then applies cross-modal consistency checks against figures along with auxiliary filters to eliminate erroneous pairs. We instantiate this framework to curate VeriSciQA, a dataset of 20,351 QA pairs spanning 20 scientific domains and 12 figure types. VeriSciQA poses a challenging benchmark for open-source models, with a substantial accuracy gap between the leading open-source models (64%) and a proprietary model (82%). Moreover, models fine-tuned on VeriSciQA achieve consistent improvements on SVQA benchmarks, with performance gains that scale with data size and surpass models trained on existing datasets. Human evaluation further validates the superior correctness of VeriSciQA. Together, these evidences demonstrate that continued data expansion by our scalable framework can further advance SVQA capability in the open-source community.
[482] iMontage: Unified, Versatile, Highly Dynamic Many-to-many Image Generation
Zhoujie Fu, Xianfang Zeng, Jinghong Lan, Xinyao Liao, Cheng Chen, Junyi Chen, Jiacheng Wei, Wei Cheng, Shiyu Liu, Yunuo Chen, Gang Yu, Guosheng Lin
Main category: cs.CV
TL;DR: iMontage repurposes pre-trained video models into all-in-one image generators that produce variable-length image sets with natural transitions and expanded dynamic range.
Details
Motivation: Video models have strong temporal coherence priors but limited dynamics due to continuous training data. Injecting diverse image content into this coherent framework can create image sets with both natural transitions and expansive dynamics.Method: Introduces iMontage framework with minimally invasive adaptation strategy, tailored data curation, and training paradigm to repurpose video models for image generation while preserving motion priors.
Result: iMontage excels at many-in-many-out image tasks, maintaining strong cross-image consistency while generating scenes with extraordinary dynamics beyond conventional scopes.
Conclusion: iMontage successfully unifies diverse image generation and editing tasks by leveraging video models’ temporal coherence while injecting image data diversity, creating a versatile framework for variable-length image set generation.
Abstract: Pre-trained video models learn powerful priors for generating high-quality, temporally coherent content. While these models excel at temporal coherence, their dynamics are often constrained by the continuous nature of their training data. We hypothesize that by injecting the rich and unconstrained content diversity from image data into this coherent temporal framework, we can generate image sets that feature both natural transitions and a far more expansive dynamic range. To this end, we introduce iMontage, a unified framework designed to repurpose a powerful video model into an all-in-one image generator. The framework consumes and produces variable-length image sets, unifying a wide array of image generation and editing tasks. To achieve this, we propose an elegant and minimally invasive adaptation strategy, complemented by a tailored data curation process and training paradigm. This approach allows the model to acquire broad image manipulation capabilities without corrupting its invaluable original motion priors. iMontage excels across several mainstream many-in-many-out tasks, not only maintaining strong cross-image contextual consistency but also generating scenes with extraordinary dynamics that surpass conventional scopes. Find our homepage at: https://kr1sjfu.github.io/iMontage-web/.
[483] 3-Tracer: A Tri-level Temporal-Aware Framework for Audio Forgery Detection and Localization
Shuhan Xia, Xuannan Liu, Xing Cui, Peipei Li
Main category: cs.CV
TL;DR: T3-Tracer is a novel framework for detecting partial audio forgeries that analyzes audio at three hierarchical levels (frame, segment, audio) using two complementary modules to capture both transient and sustained anomalies.
Details
Motivation: Partial audio forgery is challenging to detect because attackers modify only semantically critical frames while preserving overall perceptual authenticity. Existing methods focus on single-frame detection without hierarchical structure to capture anomalies across different temporal levels.Method: T3-Tracer jointly analyzes audio at frame, segment, and audio levels. It consists of two core modules: 1) Frame-Audio Feature Aggregation Module (FA-FAM) that detects frame authenticity by combining frame-level and audio-level temporal information, and 2) Segment-level Multi-Scale Discrepancy-Aware Module (SMDAM) that refines detection by modeling frame features and inter-frame differences across multi-scale temporal windows to identify forged boundaries.
Result: Extensive experiments on three challenging datasets demonstrate that T3-Tracer achieves state-of-the-art performance in detecting partial audio forgeries.
Conclusion: The proposed hierarchical three-level analysis framework effectively addresses the limitations of existing methods by capturing both transient and sustained anomalies, making it a powerful solution for detecting sophisticated partial audio forgeries.
Abstract: Recently, partial audio forgery has emerged as a new form of audio manipulation. Attackers selectively modify partial but semantically critical frames while preserving the overall perceptual authenticity, making such forgeries particularly difficult to detect. Existing methods focus on independently detecting whether a single frame is forged, lacking the hierarchical structure to capture both transient and sustained anomalies across different temporal levels. To address these limitations, We identify three key levels relevant to partial audio forgery detection and present T3-Tracer, the first framework that jointly analyzes audio at the frame, segment, and audio levels to comprehensively detect forgery traces. T3-Tracer consists of two complementary core modules: the Frame-Audio Feature Aggregation Module (FA-FAM) and the Segment-level Multi-Scale Discrepancy-Aware Module (SMDAM). FA-FAM is designed to detect the authenticity of each audio frame. It combines both frame-level and audio-level temporal information to detect intra-frame forgery cues and global semantic inconsistencies. To further refine and correct frame detection, we introduce SMDAM to detect forgery boundaries at the segment level. It adopts a dual-branch architecture that jointly models frame features and inter-frame differences across multi-scale temporal windows, effectively identifying abrupt anomalies that appeared on the forged boundaries. Extensive experiments conducted on three challenging datasets demonstrate that our approach achieves state-of-the-art performance.
[484] AVFakeBench: A Comprehensive Audio-Video Forgery Detection Benchmark for AV-LMMs
Shuhan Xia, Peipei Li, Xuannan Liu, Dongsen Zhang, Xinyu Guo, Zekun Li
Main category: cs.CV
TL;DR: AVFakeBench is the first comprehensive audio-video forgery detection benchmark covering diverse forgery types beyond human-centric deepfakes, with 12K curated questions across 7 forgery types and 4 annotation levels, evaluating AV-LMMs’ detection capabilities.
Details
Motivation: Existing benchmarks are limited to DeepFake-based forgeries and single-granularity annotations, failing to capture the diversity and complexity of real-world audio-video forgery scenarios that extend beyond human-centric manipulations.Method: Proposed AVFakeBench with 12K audio-video questions covering 7 forgery types and 4 annotation levels, using a multi-stage hybrid forgery framework that integrates proprietary models for task planning with expert generative models for precise manipulation.
Result: Evaluation of 11 Audio-Video Large Language Models and 2 prevalent detection methods shows AV-LMMs have potential as emerging forgery detectors but reveal notable weaknesses in fine-grained perception and reasoning.
Conclusion: AVFakeBench establishes a comprehensive benchmark for audio-video forgery detection, demonstrating both the promise and limitations of AV-LMMs for this task, highlighting the need for improved fine-grained perception and reasoning capabilities.
Abstract: The threat of Audio-Video (AV) forgery is rapidly evolving beyond human-centric deepfakes to include more diverse manipulations across complex natural scenes. However, existing benchmarks are still confined to DeepFake-based forgeries and single-granularity annotations, thus failing to capture the diversity and complexity of real-world forgery scenarios. To address this, we introduce AVFakeBench, the first comprehensive audio-video forgery detection benchmark that spans rich forgery semantics across both human subject and general subject. AVFakeBench comprises 12K carefully curated audio-video questions, covering seven forgery types and four levels of annotations. To ensure high-quality and diverse forgeries, we propose a multi-stage hybrid forgery framework that integrates proprietary models for task planning with expert generative models for precise manipulation. The benchmark establishes a multi-task evaluation framework covering binary judgment, forgery types classification, forgery detail selection, and explanatory reasoning. We evaluate 11 Audio-Video Large Language Models (AV-LMMs) and 2 prevalent detection methods on AVFakeBench, demonstrating the potential of AV-LMMs as emerging forgery detectors while revealing their notable weaknesses in fine-grained perception and reasoning.
[485] CaliTex: Geometry-Calibrated Attention for View-Coherent 3D Texture Generation
Chenyu Liu, Hongze Chen, Jingzhi Bao, Lingting Zhu, Runze Zhang, Weikai Chen, Zeyu Hu, Yingda Yin, Keyang Luo, Xin Wang
Main category: cs.CV
TL;DR: CaliTex introduces geometry-calibrated attention to solve cross-view inconsistency in 3D texture generation by aligning attention with 3D structure through part-aligned and condition-routed attention modules.
Details
Motivation: Current 3D texture generation systems suffer from cross-view inconsistency where textures appear convincing from one viewpoint but fail to align across others, caused by attention ambiguity in unstructured full attention that leads to geometric confusion and unstable appearance-structure coupling.Method: CaliTex introduces geometry-calibrated attention with two modules: Part-Aligned Attention that enforces spatial alignment across semantically matched parts, and Condition-Routed Attention which routes appearance information through geometry-conditioned pathways. This is coupled with a two-stage diffusion transformer to make geometric coherence an inherent network behavior.
Result: CaliTex produces seamless and view-consistent textures and outperforms both open-source and commercial baselines in 3D texture generation.
Conclusion: By explicitly aligning attention with 3D structure through geometry-calibrated attention, CaliTex addresses the fundamental issue of cross-view inconsistency in 3D texture generation, making geometric coherence an inherent property rather than an optimization byproduct.
Abstract: Despite major advances brought by diffusion-based models, current 3D texture generation systems remain hindered by cross-view inconsistency – textures that appear convincing from one viewpoint often fail to align across others. We find that this issue arises from attention ambiguity, where unstructured full attention is applied indiscriminately across tokens and modalities, causing geometric confusion and unstable appearance-structure coupling. To address this, we introduce CaliTex, a framework of geometry-calibrated attention that explicitly aligns attention with 3D structure. It introduces two modules: Part-Aligned Attention that enforces spatial alignment across semantically matched parts, and Condition-Routed Attention which routes appearance information through geometry-conditioned pathways to maintain spatial fidelity. Coupled with a two-stage diffusion transformer, CaliTex makes geometric coherence an inherent behavior of the network rather than a byproduct of optimization. Empirically, CaliTex produces seamless and view-consistent textures and outperforms both open-source and commercial baselines.
[486] Deep Learning-Based Multiclass Classification of Oral Lesions with Stratified Augmentation
Joy Naoum, Revana Salama, Ali Hamdi
Main category: cs.CV
TL;DR: Deep learning model achieves 83.33% accuracy in classifying 16 oral lesions using data augmentation and oversampling to address imbalanced datasets.
Details
Motivation: Oral cancer is often diagnosed late due to visual similarity between benign, precancerous, and malignant lesions. Early computer-aided diagnosis systems could significantly improve clinical outcomes.Method: Deep learning multiclass classifier for 16 oral lesions using stratified data splitting, advanced data augmentation, and oversampling techniques to handle limited and imbalanced datasets.
Result: Achieved 83.33% accuracy, 89.12% precision, and 77.31% recall, outperforming state-of-the-art methods, with particularly notable performance on minority class classification.
Conclusion: The framework demonstrates the effectiveness of oversampling and augmentation strategies for imbalanced medical datasets and shows promise as a first step toward trustworthy computer-aided diagnostic systems for early oral cancer detection.
Abstract: Oral cancer is highly common across the globe and is mostly diagnosed during the later stages due to the close visual similarity to benign, precancerous, and malignant lesions in the oral cavity. Implementing computer aided diagnosis systems early on has the potential to greatly improve clinical outcomes. This research intends to use deep learning to build a multiclass classifier for sixteen different oral lesions. To overcome the challenges of limited and imbalanced datasets, the proposed technique combines stratified data splitting and advanced data augmentation and oversampling to perform the classification. The experimental results, which achieved 83.33 percent accuracy, 89.12 percent precision, and 77.31 percent recall, demonstrate the superiority of the suggested model over state of the art methods now in use. The suggested model effectively conveys the effectiveness of oversampling and augmentation strategies in situations where the minority class classification performance is noteworthy. As a first step toward trustworthy computer aided diagnostic systems for the early detection of oral cancer in clinical settings, the suggested framework shows promise.
[487] TPCNet: Triple physical constraints for Low-light Image Enhancement
Jing-Yi Shi, Ming-Fei Li, Ling-An Wu
Main category: cs.CV
TL;DR: TPCNet: A Retinex-based low-light enhancement method using Kubelka-Munk theory with triple physical constraints in feature space, outperforming SOTA methods on 10 datasets without adding parameters.
Details
Motivation: Existing Retinex-based methods ignore specular reflection and use image-space constraints, limiting generalization. Need to incorporate specular reflection and reformulate physical constraints for better low-light enhancement.Method: Propose Triple Physical Constraints (TPCs) theory based on Kubelka-Munk theory, preserving specular reflection coefficient. Build TPCNet with physical constraints in feature space rather than image space, maintaining relationships between illumination, reflection, and detection.
Result: TPCNet outperforms state-of-the-art methods on 10 datasets in both quantitative metrics and visual quality. Constraints improve performance without introducing new parameters.
Conclusion: The TPC theory and TPCNet effectively address limitations of previous Retinex-based methods by incorporating specular reflection and feature-space constraints, achieving superior low-light image enhancement.
Abstract: Low-light image enhancement is an essential computer vision task to improve image contrast and to decrease the effects of color bias and noise. Many existing interpretable deep-learning algorithms exploit the Retinex theory as the basis of model design. However, previous Retinex-based algorithms, that consider reflected objects as ideal Lambertian ignore specular reflection in the modeling process and construct the physical constraints in image space, limiting generalization of the model. To address this issue, we preserve the specular reflection coefficient and reformulate the original physical constraints in the imaging process based on the Kubelka-Munk theory, thereby constructing constraint relationship between illumination, reflection, and detection, the so-called triple physical constraints (TPCs)theory. Based on this theory, the physical constraints are constructed in the feature space of the model to obtain the TPC network (TPCNet). Comprehensive quantitative and qualitative benchmark and ablation experiments confirm that these constraints effectively improve the performance metrics and visual quality without introducing new parameters, and demonstrate that our TPCNet outperforms other state-of-the-art methods on 10 datasets.
[488] HybridWorldSim: A Scalable and Controllable High-fidelity Simulator for Autonomous Driving
Qiang Li, Yingwenqi Jiang, Tuoxi Li, Duyu Chen, Xiang Feng, Yucheng Ao, Shangyue Liu, Xingchen Yu, Youcheng Cai, Yumeng Liu, Yuexin Ma, Xin Hu, Li Liu, Yu Zhang, Linkun Xu, Bingtao Gao, Xueyuan Wang, Shuchang Zhou, Xianming Liu, Ligang Liu
Main category: cs.CV
TL;DR: HybridWorldSim is a hybrid simulation framework combining neural reconstruction for static backgrounds with generative modeling for dynamic agents, enabling realistic and controllable autonomous driving simulation with visual and spatial consistency.
Details
Motivation: Existing autonomous driving simulation approaches struggle with novel view synthesis under large viewpoint changes and maintaining geometric consistency, limiting their effectiveness for end-to-end autonomous driving development.Method: HybridWorldSim integrates multi-traversal neural reconstruction for static backgrounds with generative modeling for dynamic agents, creating a unified framework that addresses limitations of previous methods.
Result: The framework surpasses previous state-of-the-art methods and is accompanied by the MIRROR dataset, a new multi-traversal dataset capturing diverse routes and environmental conditions across different cities.
Conclusion: HybridWorldSim provides a practical and scalable solution for high-fidelity autonomous driving simulation, offering a valuable resource for research and development in the field.
Abstract: Realistic and controllable simulation is critical for advancing end-to-end autonomous driving, yet existing approaches often struggle to support novel view synthesis under large viewpoint changes or to ensure geometric consistency. We introduce HybridWorldSim, a hybrid simulation framework that integrates multi-traversal neural reconstruction for static backgrounds with generative modeling for dynamic agents. This unified design addresses key limitations of previous methods, enabling the creation of diverse and high-fidelity driving scenarios with reliable visual and spatial consistency. To facilitate robust benchmarking, we further release a new multi-traversal dataset MIRROR that captures a wide range of routes and environmental conditions across different cities. Extensive experiments demonstrate that HybridWorldSim surpasses previous state-of-the-art methods, providing a practical and scalable solution for high-fidelity simulation and a valuable resource for research and development in autonomous driving.
[489] TTSnap: Test-Time Scaling of Diffusion Models via Noise-Aware Pruning
Qingtao Yu, Changlin Song, Minghao Sun, Zhengyang Yu, Vinay Kumar Verma, Soumya Roy, Sumit Negi, Hongdong Li, Dylan Campbell
Main category: cs.CV
TL;DR: TTSnap improves test-time scaling for diffusion models by pruning low-quality noise seeds early using noise-aware reward models, reducing computational costs while maintaining quality.
Details
Motivation: Current test-time scaling methods require fully denoising multiple noise seeds to compute rewards, which is computationally expensive and limits exploration under fixed budgets.Method: Train noise-aware reward models via self-distillation to align intermediate estimate rewards with final clean image rewards, using curriculum training from clean to noisy images, and introduce reward alignment metric.
Result: Improves performance by over 16% compared to existing methods, enables more efficient test-time scaling, and provides orthogonal gains when combined with other techniques.
Conclusion: TTSnap offers an effective framework for computationally efficient test-time scaling by early pruning of low-quality candidates through noise-aware reward prediction.
Abstract: A prominent approach to test-time scaling for text-to-image diffusion models formulates the problem as a search over multiple noise seeds, selecting the one that maximizes a certain image-reward function. The effectiveness of this strategy heavily depends on the number and diversity of noise seeds explored. However, verifying each candidate is computationally expensive, because each must be fully denoised before a reward can be computed. This severely limits the number of samples that can be explored under a fixed budget. We propose test-time scaling with noise-aware pruning (TTSnap), a framework that prunes low-quality candidates without fully denoising them. The key challenge is that reward models are learned in the clean image domain, and the ranking of rewards predicted for intermediate estimates are often inconsistent with those predicted for clean images. To overcome this, we train noise-aware reward models via self-distillation to align the reward for intermediate estimates with that of the final clean images. To stabilize learning across different noise levels, we adopt a curriculum training strategy that progressively shifts the data domain from clean images to noise images. In addition, we introduce a new metric that measures reward alignment and computational budget utilization. Experiments demonstrate that our approach improves performance by over 16% compared with existing methods, enabling more efficient and effective test-time scaling. It also provides orthogonal gains when combined with post-training techniques and local test-time optimization. Code: https://github.com/TerrysLearning/TTSnap/.
[490] Structure is Supervision: Multiview Masked Autoencoders for Radiology
Sonia Laguna, Andrea Agostini, Alain Ryser, Samuel Ruiperez-Campillo, Irene Cannistraci, Moritz Vandenhirtz, Stephan Mandt, Nicolas Deperrois, Farhad Nooralahzadeh, Michael Krauthammer, Thomas M. Sutter, Julia E. Vogt
Main category: cs.CV
TL;DR: MVMAE is a self-supervised framework that uses multi-view radiology images and text reports to learn robust medical representations, outperforming baselines on disease classification tasks.
Details
Motivation: Medical ML systems need pretraining strategies that exploit clinical data structure. Radiology studies have natural multi-view organization (different imaging projections) that can be leveraged for learning view-invariant and disease-relevant representations.Method: MVMAE combines masked image reconstruction with cross-view alignment, using clinical redundancy across projections as self-supervision. MVMAE-V2T extends this by incorporating radiology reports as auxiliary text-based learning signals while maintaining vision-only inference.
Result: Evaluated on MIMIC-CXR, CheXpert, and PadChest datasets for disease classification, MVMAE consistently outperforms supervised and vision-language baselines. MVMAE-V2T provides additional gains, especially in low-label regimes where text supervision is most beneficial.
Conclusion: Structural (multi-view) and textual supervision are complementary paths toward scalable, clinically grounded medical foundation models, with MVMAE demonstrating the value of exploiting clinical data organization for robust medical ML.
Abstract: Building robust medical machine learning systems requires pretraining strategies that exploit the intrinsic structure present in clinical data. We introduce Multiview Masked Autoencoder (MVMAE), a self-supervised framework that leverages the natural multi-view organization of radiology studies to learn view-invariant and disease-relevant representations. MVMAE combines masked image reconstruction with cross-view alignment, transforming clinical redundancy across projections into a powerful self-supervisory signal. We further extend this approach with MVMAE-V2T, which incorporates radiology reports as an auxiliary text-based learning signal to enhance semantic grounding while preserving fully vision-based inference. Evaluated on a downstream disease classification task on three large-scale public datasets, MIMIC-CXR, CheXpert, and PadChest, MVMAE consistently outperforms supervised and vision-language baselines. Furthermore, MVMAE-V2T provides additional gains, particularly in low-label regimes where structured textual supervision is most beneficial. Together, these results establish the importance of structural and textual supervision as complementary paths toward scalable, clinically grounded medical foundation models.
[491] Benchmarking machine learning models for multi-class state recognition in double quantum dot data
Valeria Díaz Moreno, Ryan P Khalili, Daniel Schug, Patrick J. Walsh, Justyna P. Zwolak
Main category: cs.CV
TL;DR: CNNs with min-max normalization offer the best practical trade-off for quantum dot state recognition in charge-stability diagrams, outperforming more complex models like U-Nets and ViTs on experimental data.
Details
Motivation: Scaling quantum dot arrays requires automated tuning strategies that depend on accurate state identification from charge-stability diagrams, necessitating evaluation of machine learning approaches for this task.Method: Benchmarked four ML architectures (U-Nets, visual transformers, CNNs, MDNs) for multi-class state recognition in double-QD charge-stability diagrams across different data budgets and normalization schemes using synthetic and experimental data.
Result: U-Nets and ViTs achieved highest scores on synthetic data (>0.98 MSE) but failed to generalize to experimental data. CNNs offered best trade-off on experimental CSDs with strong accuracy and two orders of magnitude fewer parameters than U-Nets/ViTs. MDNs were computationally efficient but had lower performance.
Conclusion: CNNs with min-max normalization are the most practical approach for quantum dot charge-stability diagram analysis, balancing accuracy, generalization, and computational efficiency for scalable quantum processor tuning.
Abstract: Semiconductor quantum dots (QDs) are a leading platform for scalable quantum processors. However, scaling to large arrays requires reliable, automated tuning strategies for devices’ bootstrapping, calibration, and operation, with many tuning aspects depending on accurately identifying QD device states from charge-stability diagrams (CSDs). In this work, we present a comprehensive benchmarking study of four modern machine learning (ML) architectures for multi-class state recognition in double-QD CSDs. We evaluate their performance across different data budgets and normalization schemes using both synthetic and experimental data. We find that the more resource-intensive models – U-Nets and visual transformers (ViTs) – achieve the highest MSE score (defined as $1-\mathrm{MSE}$) on synthetic data (over $0.98$) but fail to generalize to experimental data. MDNs are the most computationally efficient and exhibit highly stable training, but with substantially lower peak performance. CNNs offer the most favorable trade-off on experimental CSDs, achieving strong accuracy with two orders of magnitude fewer parameters than the U-Nets and ViTs. Normalization plays a nontrivial role: min-max scaling generally yields higher MSE scores but less stable convergence, whereas z-score normalization produces more predictable training dynamics but at reduced accuracy for most models. Overall, our study shows that CNNs with min-max normalization are a practical approach for QD CSDs.
[492] ReasonEdit: Towards Reasoning-Enhanced Image Editing Models
Fukun Yin, Shiyu Liu, Yucheng Han, Zhibo Wang, Peng Xing, Rui Wang, Wei Cheng, Yingming Wang, Aojie Li, Zixin Yin, Pengtao Chen, Xiangyu Zhang, Daxin Jiang, Xianfang Zeng, Gang Yu
Main category: cs.CV
TL;DR: Proposes a reasoning-enhanced image editing framework that unlocks MLLM capabilities through thinking and reflection mechanisms in a loop, achieving significant performance gains over existing methods.
Details
Motivation: Current image editing models freeze MLLM encoders during training, limiting their reasoning capabilities. The authors aim to unlock MLLM's reasoning power to improve instruction understanding and editing accuracy.Method: Introduces a thinking-editing-reflection loop: thinking mechanism interprets abstract instructions using MLLM’s world knowledge, while reflection reviews results, corrects unintended manipulations, and identifies stopping points. Two versions: ReasonEdit-S (initialized from Step1X-Edit) and ReasonEdit-Q (integrated with Qwen-Image-Edit).
Result: Achieves significant performance improvements: +4.3% on ImgEdit, +4.7% on GEdit, and +8.2% on Kris with ReasonEdit-S. ReasonEdit-Q outperforms previous open-source methods on both GEdit and Kris benchmarks.
Conclusion: Unlocking MLLM reasoning capabilities through thinking and reflection mechanisms significantly enhances image editing performance, demonstrating the value of incorporating reasoning loops into editing frameworks.
Abstract: Recent advances in image editing models have shown remarkable progress. A common architectural design couples a multimodal large language model (MLLM) encoder with a diffusion decoder, as seen in systems such as Step1X-Edit and Qwen-Image-Edit, where the MLLM encodes both the reference image and the instruction but remains frozen during training. In this work, we demonstrate that unlocking the reasoning capabilities of MLLM can further push the boundaries of editing models. Specifically, we explore two reasoning mechanisms, thinking and reflection, which enhance instruction understanding and editing accuracy. Based on that, our proposed framework enables image editing in a thinking-editing-reflection loop: the thinking mechanism leverages the world knowledge of MLLM to interpret abstract instructions, while the reflection reviews editing results, automatically corrects unintended manipulations, and identifies the stopping round. Extensive experiments demonstrate that our reasoning approach achieves significant performance gains, with improvements of ImgEdit (+4.3%), GEdit (+4.7%), and Kris (+8.2%) when initializing our DiT from the Step1X-Edit (ReasonEdit-S), and also outperforms previous open-source methods on both GEdit and Kris when integrated with Qwen-Image-Edit (ReasonEdit-Q).
[493] Fusion or Confusion? Assessing the impact of visible-thermal image fusion for automated wildlife detection
Camille Dionne-Pierre, Samuel Foucher, Jérôme Théau, Jérôme Lemaître, Patrick Charbonneau, Maxime Brousseau, Mathieu Varin
Main category: cs.CV
TL;DR: Combining visible and thermal infrared aerial imagery with YOLO11n deep learning improves automated detection of great blue herons and nests, with late fusion achieving best results (93% F1 score).
Details
Motivation: Need efficient wildlife monitoring methods for biodiversity conservation; combining VIS and TIR imagery could add complementary information and improve automated detection compared to single-source approaches.Method: Used synchronous aerial VIS and TIR imagery of great blue herons; tested two fusion methods: early fusion (PCA-based image fusion) and late fusion (CART-based detection fusion); aligned images using deep learning; trained YOLO11n models.
Result: Both fusion methods improved F1 scores over VIS-only model; late fusion achieved best results: 93.0% F1 score for occupied nests (vs 90.2% VIS-only); identified false positives with 90% recall.
Conclusion: VIS-TIR fusion improves detection accuracy but has limitations (TIR FOV constraints, alignment issues); very high-resolution visible sensors alone may be more practical for operational surveys.
Abstract: Efficient wildlife monitoring methods are necessary for biodiversity conservation and management. The combination of remote sensing, aerial imagery and deep learning offer promising opportunities to renew or improve existing survey methods. The complementary use of visible (VIS) and thermal infrared (TIR) imagery can add information compared to a single-source image and improve results in an automated detection context. However, the alignment and fusion process can be challenging, especially since visible and thermal images usually have different fields of view (FOV) and spatial resolutions. This research presents a case study on the great blue heron (Ardea herodias) to evaluate the performances of synchronous aerial VIS and TIR imagery to automatically detect individuals and nests using a YOLO11n model. Two VIS-TIR fusion methods were tested and compared: an early fusion approach and a late fusion approach, to determine if the addition of the TIR image gives any added value compared to a VIS-only model. VIS and TIR images were automatically aligned using a deep learning model. A principal component analysis fusion method was applied to VIS-TIR image pairs to form the early fusion dataset. A classification and regression tree was used to process the late fusion dataset, based on the detection from the VIS-only and TIR-only trained models. Across all classes, both late and early fusion improved the F1 score compared to the VIS-only model. For the main class, occupied nest, the late fusion improved the F1 score from 90.2 (VIS-only) to 93.0%. This model was also able to identify false positives from both sources with 90% recall. Although fusion methods seem to give better results, this approach comes with a limiting TIR FOV and alignment constraints that eliminate data. Using an aircraft-mounted very high-resolution visible sensor could be an interesting option for operationalizing surveys.
[494] DenoiseGS: Gaussian Reconstruction Model for Burst Denoising
Yongsen Cheng, Yuanhao Cai, Yulun Zhang
Main category: cs.CV
TL;DR: DenoiseGS uses 3D Gaussian Splatting for burst denoising, addressing noise degradation and detail loss through Gaussian self-consistency and log-weighted frequency losses, achieving SOTA results with 250× faster inference.
Details
Motivation: Existing burst denoising methods struggle with large motion or have prohibitive computational costs. There's a need for efficient, high-quality denoising that can handle noisy inputs while preserving fine details.Method: Proposes DenoiseGS framework using 3D Gaussian Splatting for burst denoising. Introduces Gaussian self-consistency (GSC) loss to regularize geometry from noisy inputs using high-quality Gaussian point clouds from clean inputs. Adds log-weighted frequency (LWF) loss for spectral domain supervision to preserve fine details by adaptively weighting frequency discrepancies.
Result: Significantly exceeds state-of-the-art NeRF-based methods on both burst denoising and novel view synthesis under noisy conditions, while achieving 250× faster inference speed.
Conclusion: DenoiseGS demonstrates that 3D Gaussian Splatting can be effectively adapted for burst denoising, overcoming challenges of noise degradation and detail loss through novel consistency and frequency-based losses, achieving superior performance with dramatically improved efficiency.
Abstract: Burst denoising methods are crucial for enhancing images captured on handheld devices, but they often struggle with large motion or suffer from prohibitive computational costs. In this paper, we propose DenoiseGS, the first framework to leverage the efficiency of 3D Gaussian Splatting for burst denoising. Our approach addresses two key challenges when applying feedforward Gaussian reconsturction model to noisy inputs: the degradation of Gaussian point clouds and the loss of fine details. To this end, we propose a Gaussian self-consistency (GSC) loss, which regularizes the geometry predicted from noisy inputs with high-quality Gaussian point clouds. These point clouds are generated from clean inputs by the same model that we are training, thereby alleviating potential bias or domain gaps. Additionally, we introduce a log-weighted frequency (LWF) loss to strengthen supervision within the spectral domain, effectively preserving fine-grained details. The LWF loss adaptively weights frequency discrepancies in a logarithmic manner, emphasizing challenging high-frequency details. Extensive experiments demonstrate that DenoiseGS significantly exceeds the state-of-the-art NeRF-based methods on both burst denoising and novel view synthesis under noisy conditions, while achieving 250$\times$ faster inference speed. Code and models are released at https://github.com/yscheng04/DenoiseGS.
[495] One-to-All Animation: Alignment-Free Character Animation and Image Pose Transfer
Shijun Shi, Jing Xu, Zhihang Li, Chunli Peng, Xiaoda Yang, Lijing Lu, Kai Hu, Jiangning Zhang
Main category: cs.CV
TL;DR: One-to-All Animation is a unified diffusion framework for character animation and pose transfer that handles spatially misaligned references with arbitrary layouts through self-supervised outpainting and comprehensive identity feature extraction.
Details
Motivation: Existing diffusion-based character animation methods are limited to spatially aligned reference-pose pairs with matched skeletal structures, leaving the problem of reference-pose misalignment unsolved.Method: 1) Reformulate training as self-supervised outpainting to transform diverse-layout references into unified occluded-input format; 2) Design reference extractor for comprehensive identity feature extraction; 3) Integrate hybrid reference fusion attention for varying resolutions and dynamic sequence lengths; 4) Introduce identity-robust pose control to decouple appearance from skeletal structure; 5) Use token replace strategy for coherent long-video generation.
Result: Extensive experiments show the method outperforms existing approaches in character animation and image pose transfer for references with arbitrary layouts.
Conclusion: One-to-All Animation provides a unified framework for high-fidelity character animation and pose transfer that successfully handles spatially misaligned references, addressing a key limitation of existing methods.
Abstract: Recent advances in diffusion models have greatly improved pose-driven character animation. However, existing methods are limited to spatially aligned reference-pose pairs with matched skeletal structures. Handling reference-pose misalignment remains unsolved. To address this, we present One-to-All Animation, a unified framework for high-fidelity character animation and image pose transfer for references with arbitrary layouts. First, to handle spatially misaligned reference, we reformulate training as a self-supervised outpainting task that transforms diverse-layout reference into a unified occluded-input format. Second, to process partially visible reference, we design a reference extractor for comprehensive identity feature extraction. Further, we integrate hybrid reference fusion attention to handle varying resolutions and dynamic sequence lengths. Finally, from the perspective of generation quality, we introduce identity-robust pose control that decouples appearance from skeletal structure to mitigate pose overfitting, and a token replace strategy for coherent long-video generation. Extensive experiments show that our method outperforms existing approaches. The code and model are available at https://github.com/ssj9596/One-to-All-Animation.
[496] DualCamCtrl: Dual-Branch Diffusion Model for Geometry-Aware Camera-Controlled Video Generation
Hongfei Zhang, Kanghao Chen, Zixin Zhang, Harold Haodong Chen, Yuanhuiyi Lyu, Yuqi Zhang, Shuai Yang, Kun Zhou, Yingcong Chen
Main category: cs.CV
TL;DR: DualCamCtrl is a novel end-to-end diffusion model for camera-controlled video generation that uses a dual-branch framework to generate camera-consistent RGB and depth sequences, achieving better scene understanding and geometric awareness.
Details
Motivation: Existing camera-controlled video generation methods using ray-based camera pose representations often lack sufficient scene understanding and geometric awareness, limiting their ability to faithfully follow specified camera trajectories.Method: Dual-branch framework that mutually generates camera-consistent RGB and depth sequences, with Semantic Guided Mutual Alignment (SIGMA) mechanism for semantics-guided RGB-depth fusion and mutual reinforcement.
Result: Achieves more consistent camera-controlled video generation with over 40% reduction in camera motion errors compared to prior methods, and reveals complementary roles of early/late denoising stages in structure formation and detail refinement.
Conclusion: DualCamCtrl effectively disentangles appearance and geometry modeling through its dual-branch approach with SIGMA, enabling more faithful adherence to specified camera trajectories in video generation.
Abstract: This paper presents DualCamCtrl, a novel end-to-end diffusion model for camera-controlled video generation. Recent works have advanced this field by representing camera poses as ray-based conditions, yet they often lack sufficient scene understanding and geometric awareness. DualCamCtrl specifically targets this limitation by introducing a dual-branch framework that mutually generates camera-consistent RGB and depth sequences. To harmonize these two modalities, we further propose the Semantic Guided Mutual Alignment (SIGMA) mechanism, which performs RGB-depth fusion in a semantics-guided and mutually reinforced manner. These designs collectively enable DualCamCtrl to better disentangle appearance and geometry modeling, generating videos that more faithfully adhere to the specified camera trajectories. Additionally, we analyze and reveal the distinct influence of depth and camera poses across denoising stages and further demonstrate that early and late stages play complementary roles in forming global structure and refining local details. Extensive experiments demonstrate that DualCamCtrl achieves more consistent camera-controlled video generation, with over 40% reduction in camera motion errors compared with prior methods. Our project page: https://soyouthinkyoucantell.github.io/dualcamctrl-page/
[497] PowerCLIP: Powerset Alignment for Contrastive Pre-Training
Masaki Kawamura, Nakamasa Inoue, Rintaro Yanagi, Hirokatsu Kataoka, Rio Yokota
Main category: cs.CV
TL;DR: PowerCLIP is a contrastive vision-language pre-training framework that uses powerset alignment to capture compositional semantics spanning multiple image regions, with efficient non-linear aggregators to avoid exponential computational costs.
Details
Motivation: Existing CLIP-like frameworks struggle to capture compositional semantics that span multiple image regions, as they typically align individual text tokens with specific image patches. There's a need for better fine-grained compositional understanding in vision-language models.Method: Proposes PowerCLIP with powerset alignment that optimizes region-to-phrase alignments by minimizing loss between powersets of image regions and textual parse trees. Introduces efficient non-linear aggregators (NLAs) to reduce computational complexity from O(2^M) to O(M) while approximating exact loss.
Result: PowerCLIP outperforms state-of-the-art methods in zero-shot classification and retrieval tasks, demonstrating improved compositionality and robustness compared to existing approaches.
Conclusion: The powerset alignment framework with efficient NLAs effectively captures compositional semantics across multiple image regions, leading to superior zero-shot performance while maintaining computational feasibility.
Abstract: Contrastive vision-language pre-training frameworks such as CLIP have demonstrated impressive zero-shot performance across a range of vision-language tasks. Recent studies have shown that aligning individual text tokens with specific image patches or regions enhances fine-grained compositional understanding. However, it remains challenging to capture compositional semantics that span multiple image regions. To address this limitation, we propose PowerCLIP, a novel contrastive pre-training framework enhanced by powerset alignment, which exhaustively optimizes region-to-phrase alignments by minimizing the loss defined between powersets of image regions and textual parse trees. Since the naive powerset construction incurs exponential computational cost due to the combinatorial explosion in the number of region subsets, we introduce efficient non-linear aggregators (NLAs) that reduce complexity from O(2^M) to O(M) with respect to the number of regions M, while approximating the exact loss value with arbitrary precision. Our extensive experiments demonstrate that PowerCLIP outperforms state-of-the-art methods in zero-shot classification and retrieval tasks, underscoring the compositionality and robustness of our approach. Our code will be made publicly available.
[498] Fast Multi-view Consistent 3D Editing with Video Priors
Liyi Chen, Ruihuang Li, Guowen Zhang, Pengfei Wang, Lei Zhang
Main category: cs.CV
TL;DR: ViP3DE uses video generation models’ temporal consistency priors for multi-view 3D editing in a single forward pass, avoiding iterative 2D-3D-2D updates.
Details
Motivation: Existing text-driven 3D editing methods are time-consuming and produce over-smoothed results due to iterative multi-view processing and averaging of inconsistent editing signals.Method: Uses pre-trained video generation models conditioned on a single edited view to generate other consistent edited views. Introduces motion-preserved noise blending for camera pose alignment and geometry-aware denoising for better multi-view consistency.
Result: Achieves high-quality 3D editing results in a single forward pass, significantly outperforming existing methods in both editing quality and speed.
Conclusion: ViP3DE demonstrates that leveraging video generation models’ temporal consistency priors enables efficient and high-quality multi-view consistent 3D editing without iterative processing.
Abstract: Text-driven 3D editing enables user-friendly 3D object or scene editing with text instructions. Due to the lack of multi-view consistency priors, existing methods typically resort to employing 2D generation or editing models to process each view individually, followed by iterative 2D-3D-2D updating. However, these methods are not only time-consuming but also prone to over-smoothed results because the different editing signals gathered from different views are averaged during the iterative process. In this paper, we propose generative Video Prior based 3D Editing (ViP3DE) to employ the temporal consistency priors from pre-trained video generation models for multi-view consistent 3D editing in a single forward pass. Our key insight is to condition the video generation model on a single edited view to generate other consistent edited views for 3D updating directly, thereby bypassing the iterative editing paradigm. Since 3D updating requires edited views to be paired with specific camera poses, we propose motion-preserved noise blending for the video model to generate edited views at predefined camera poses. In addition, we introduce geometry-aware denoising to further enhance multi-view consistency by integrating 3D geometric priors into video models. Extensive experiments demonstrate that our proposed ViP3DE can achieve high-quality 3D editing results even within a single forward pass, significantly outperforming existing methods in both editing quality and speed.
cs.AI
[499] Gold-Medal-Level Olympiad Geometry Solving with Efficient Heuristic Auxiliary Constructions
Boyan Duan, Xiao Liang, Shuai Lu, Yaoxiang Wang, Yelong Shen, Kai-Wei Chang, Ying Nian Wu, Mao Yang, Weizhu Chen, Yeyun Gong
Main category: cs.AI
TL;DR: HAGeo is a heuristic-based method for automated geometry theorem proving that achieves gold-medal IMO performance by adding auxiliary constructions, surpassing neural network approaches like AlphaGeometry.
Details
Motivation: Automated theorem proving in Euclidean geometry, especially for IMO-level problems, is a major AI challenge. Current approaches often rely on neural networks, but there's a need for efficient CPU-based methods that can achieve human-level performance.Method: HAGeo uses a heuristic-based method for adding auxiliary constructions in geometric deduction. It builds on the insight that simple random strategies for adding auxiliary points can achieve silver-medal performance, then develops more sophisticated heuristics to reach gold-medal level.
Result: HAGeo solves 28 of 30 problems on the IMO-30 benchmark, achieving gold-medal level performance and surpassing AlphaGeometry by a notable margin. The authors also created HAGeo-409, a more challenging benchmark with 409 problems and human-assessed difficulty levels.
Conclusion: Heuristic-based methods can achieve state-of-the-art performance in geometry theorem proving without neural networks, and the new HAGeo-409 benchmark provides a more comprehensive evaluation framework that sets higher standards for future research.
Abstract: Automated theorem proving in Euclidean geometry, particularly for International Mathematical Olympiad (IMO) level problems, remains a major challenge and an important research focus in Artificial Intelligence. In this paper, we present a highly efficient method for geometry theorem proving that runs entirely on CPUs without relying on neural network-based inference. Our initial study shows that a simple random strategy for adding auxiliary points can achieve silver-medal level human performance on IMO. Building on this, we propose HAGeo, a Heuristic-based method for adding Auxiliary constructions in Geometric deduction that solves 28 of 30 problems on the IMO-30 benchmark, achieving gold-medal level performance and surpassing AlphaGeometry, a competitive neural network-based approach, by a notable margin. To evaluate our method and existing approaches more comprehensively, we further construct HAGeo-409, a benchmark consisting of 409 geometry problems with human-assessed difficulty levels. Compared with the widely used IMO-30, our benchmark poses greater challenges and provides a more precise evaluation, setting a higher bar for geometry theorem proving.
[500] Chunking Strategies for Multimodal AI Systems
Shashanka B R, Mohith Charan R, Seema Banu F
Main category: cs.AI
TL;DR: A comprehensive survey of multimodal chunking strategies across text, images, audio, video, and cross-modal data, analyzing classical and modern approaches with their methodologies, tools, benefits, and challenges.
Details
Motivation: To consolidate the landscape of multimodal chunking strategies and provide researchers/practitioners with technical foundations for developing more effective and efficient multimodal AI systems that scale with modality complexity and improve processing accuracy.Method: Provides comprehensive taxonomy and technical analysis of chunking strategies for each modality: text (fixed-size token windowing, recursive text splitting), images (object-centric visual chunking), audio (silence-based segmentation), video (scene detection), and cross-modal data. Analyzes each approach’s methodology, supporting tools (LangChain, Detectron2, PySceneDetect), benefits, and challenges.
Result: The survey establishes a technical foundation for multimodal chunking, examines granularity-context trade-offs, multimodal alignment issues, and explores emerging cross-modal strategies that preserve alignment and semantic consistency across data types.
Conclusion: Identifies open problems (asynchronous information density, noisy alignment signals) and opportunities for future research in adaptive, learning-based, and task-specific chunking to enable innovations in robust chunking pipelines for real-world applications.
Abstract: Our goal is to consolidate the landscape of multimodal chunking strategies, providing researchers and practitioners with a technical foundation and design space for developing more effective and efficient multimodal AI systems. This survey paves the way for innovations in robust chunking pipelines that scale with modality complexity, enhance processing accuracy, and improve generative coherence in real-world applications. This survey provides a comprehensive taxonomy and technical analysis of chunking strategies tailored for each modality: text, images, audio, video, and cross-modal data. We examine classical and modern approaches such as fixed-size token windowing, recursive text splitting, object-centric visual chunking, silence-based audio segmentation, and scene detection in videos. Each approach is analyzed in terms of its underlying methodology, supporting tools (e.g., LangChain, Detectron2, PySceneDetect), benefits, and challenges, particularly those related to granularity-context trade-offs and multimodal alignment. Furthermore, we explore emerging cross-modal chunking strategies that aim to preserve alignment and semantic consistency across disparate data types [4]. We also include comparative insights, highlight open problems such as asynchronous information density and noisy alignment signals, and identify opportunities for future research in adaptive, learning-based, and task-specific chunking.
[501] A Rosetta Stone for AI Benchmarks
Anson Ho, Jean-Stanislas Denain, David Atanasov, Samuel Albanie, Rohin Shah
Main category: cs.AI
TL;DR: A statistical framework that stitches AI benchmarks together to create a unified scale for comparing model capabilities across time and different benchmarks, enabling long-term trend analysis without assumptions about capability evolution.
Details
Motivation: Most AI benchmarks saturate quickly (within years or months), making it difficult to study long-run trends in AI capabilities. There's a need for a method to compare models across different benchmarks and time periods.Method: Develop a statistical framework that stitches benchmarks together, putting model capabilities and benchmark difficulties on a single numerical scale. This acts as a “Rosetta Stone” for comparison without assuming how capabilities evolve across time or with training compute.
Result: The framework enables three key applications: 1) Measuring speed of AI progress over time and forecasting future capabilities, 2) Estimating algorithmic efficiency improvements (finding higher but broadly consistent estimates with prior work), 3) Detecting rapid accelerations in AI progress.
Conclusion: The proposed statistical framework provides a robust method for studying long-term AI trends by creating a unified scale for comparing capabilities across different benchmarks and time periods, addressing the problem of benchmark saturation.
Abstract: Most AI benchmarks saturate within years or even months after they are introduced, making it hard to study long-run trends in AI capabilities. To address this challenge, we build a statistical framework that stitches benchmarks together, putting model capabilities and benchmark difficulties on a single numerical scale. This acts as a “Rosetta Stone”, allowing us to compare models across a wide range of abilities and time, even if they are not evaluated on the same benchmarks. Moreover, this works without assuming how capabilities evolve across time or with training compute. We demonstrate three applications of this framework. First, we use it to measure the speed of AI progress over time, and to forecast future AI capabilities. Second, we estimate the rate of improvements in algorithmic efficiency, finding estimates that are higher, but broadly consistent with prior work. Finally, we find that our approach can be used to detect rapid accelerations in AI progress.
[502] Reasoning Under Pressure: How do Training Incentives Influence Chain-of-Thought Monitorability?
Matt MacDermott, Qiyao Wei, Rada Djoneva, Francis Rhys Ward
Main category: cs.AI
TL;DR: Training incentives affect how faithfully chain-of-thought reasoning reflects underlying thought processes, impacting monitorability for AI safety.
Details
Motivation: AI systems that output reasoning in natural language offer safety opportunities through monitoring, but the faithfulness of chain-of-thought to actual reasoning depends on training methods.Method: Introduces methodology to measure monitorability based on whether a monitor can predict key latent variables from reasoning. Tests effects of different training incentives (length penalties, KL regularization, adversarial optimization, direct monitorability optimization) while controlling for accuracy.
Result: No consistent effects from common incentives (length penalties, KL regularization). Adversarial optimization degrades monitor performance, while direct optimization for monitorability doesn’t reliably improve it.
Conclusion: Training incentives influence monitorability, with adversarial optimization harming it, but common incentives show no consistent effects. Direct optimization for monitorability isn’t reliably effective.
Abstract: AI systems that output their reasoning in natural language offer an opportunity for safety – we can \emph{monitor} their chain of thought (CoT) for undesirable reasoning, such as the pursuit of harmful objectives. However, the extent to which CoT faithfully reflects the underlying reasoning process, and hence the extent to which it can be usefully monitored, may be influenced by certain aspects of training. We investigate how different \emph{training incentives}, applied to a reasoning model, affect its monitorability. We introduce a novel methodology for measuring monitorability according to whether a monitor can predict a key latent variable using the model’s reasoning. When controlling for accuracy, we do not find evidence for consistent effects from commonly used incentives (length penalties and KL regularisation), but we find that adversarial optimisation (penalising monitor accuracy) degrades monitor performance, while direct optimisation for monitorability does not reliably lead to improvements. Our code is available at https://github.com/QiyaoWei/reasoning-under-pressure.
[503] Trification: A Comprehensive Tree-based Strategy Planner and Structural Verification for Fact-Checking
Anab Maulana Barik, Shou Ziyi, Yang Kaiwen, Yang Qi, Shen Xin
Main category: cs.AI
TL;DR: Trification: A novel automated fact-checking framework that uses verification action dependency graphs to ensure complete claim coverage and logical reasoning.
Details
Motivation: The rapid spread of false information online necessitates automated fact-checking systems. Previous methods fail to verify all claim components and lack structured frameworks to logically connect sub-task results.Method: Trification framework: 1) Generates comprehensive verification actions for complete claim coverage, 2) Structures actions into dependency graphs to model logical interactions, 3) Allows dynamic graph modification for adaptive verification strategies.
Result: Experimental results on two challenging benchmarks show significant improvement in fact-checking accuracy, advancing state-of-the-art in automated fact-checking systems.
Conclusion: Trification addresses key limitations of previous methods by ensuring complete claim verification and providing structured logical reasoning through dynamic dependency graphs, representing an advancement in automated fact-checking technology.
Abstract: Technological advancement allows information to be shared in just a single click, which has enabled the rapid spread of false information. This makes automated fact-checking system necessary to ensure the safety and integrity of our online media ecosystem. Previous methods have demonstrated the effectiveness of decomposing the claim into simpler sub-tasks and utilizing LLM-based multi agent system to execute them. However, those models faces two limitations: they often fail to verify every component in the claim and lack of structured framework to logically connect the results of sub-tasks for a final prediction. In this work, we propose a novel automated fact-checking framework called Trification. Our framework begins by generating a comprehensive set of verification actions to ensure complete coverage of the claim. It then structured these actions into a dependency graph to model the logical interaction between actions. Furthermore, the graph can be dynamically modified, allowing the system to adapt its verification strategy. Experimental results on two challenging benchmarks demonstrate that our framework significantly enhances fact-checking accuracy, thereby advancing current state-of-the-art in automated fact-checking system.
[504] ChartPoint: Guiding MLLMs with Grounding Reflection for Chart Reasoning
Zhengzhuo Xu, SiNan Du, Yiyan Qi, SiwenLu, Chengjin Xu, Chun Yuan, Jian Guo
Main category: cs.AI
TL;DR: PointCoT addresses MLLMs’ numerical hallucinations in chart comprehension by integrating visual grounding with chain-of-thought reasoning, using bounding boxes and chart re-rendering to connect textual reasoning with visual elements.
Details
Motivation: MLLMs struggle with chart comprehension due to reliance on OCR and sparse textual annotations, leading to numerical hallucinations. Existing methods fail to address the core issue of reasoning with visual perception, particularly weak grounding in chart elements and proportional relationships.Method: Propose PointCoT that integrates reflective interaction into chain-of-thought reasoning. MLLMs generate bounding boxes and re-render charts based on location annotations to connect textual reasoning with visual grounding. Create ChartPoint-SFT-62k dataset with 19.2K chart samples featuring CoT, bounding boxes, and re-rendered visualizations.
Result: Develop instruction-tuned models ChartPointQ2 and ChartPointQ2.5 that outperform state-of-the-art across several chart benchmarks, achieving +5.04% improvement on ChartBench.
Conclusion: PointCoT effectively bridges the gap between textual reasoning and visual grounding in chart comprehension, addressing fundamental limitations of MLLMs and significantly improving performance on chart understanding tasks.
Abstract: Multimodal Large Language Models (MLLMs) have emerged as powerful tools for chart comprehension. However, they heavily rely on extracted content via OCR, which leads to numerical hallucinations when chart textual annotations are sparse. While existing methods focus on scaling instructions, they fail to address the fundamental challenge, i.e., reasoning with visual perception. In this paper, we identify a critical observation: MLLMs exhibit weak grounding in chart elements and proportional relationships, as evidenced by their inability to localize key positions to match their reasoning. To bridge this gap, we propose PointCoT, which integrates reflective interaction into chain-of-thought reasoning in charts. By prompting MLLMs to generate bounding boxes and re-render charts based on location annotations, we establish connections between textual reasoning steps and visual grounding regions. We further introduce an automated pipeline to construct ChartPoint-SFT-62k, a dataset featuring 19.2K high-quality chart samples with step-by-step CoT, bounding box, and re-rendered visualizations. Leveraging this data, we develop two instruction-tuned models, ChartPointQ2 and ChartPointQ2.5, which outperform state-of-the-art across several chart benchmarks, e.g., +5.04% on ChartBench.
[505] RL-Struct: A Lightweight Reinforcement Learning Framework for Reliable Structured Output in LLMs
Ruike Hu, Shulei Wu
Main category: cs.AI
TL;DR: A lightweight RL framework using multi-dimensional rewards and gradient regularization to bridge the “Structure Gap” between LLMs’ probabilistic generation and deterministic structured data requirements, achieving high JSON validity and structural accuracy.
Details
Motivation: LLMs struggle with structured data formats due to the "Structure Gap" - the tension between probabilistic token generation and deterministic format requirements. Traditional SFT fails to enforce strict constraints, leading to hallucinated keys or malformed structures, while constrained decoding adds significant inference latency.Method: Proposes a lightweight RL framework with a novel Multi-dimensional Reward Function that decomposes structured output tasks into hierarchical constraints (structural integrity, format correctness, content accuracy, validity). Uses Gradient Regularized Policy Optimization (GRPO) to internalize constraints without a separate critic network, reducing VRAM usage by 40% compared to PPO.
Result: Achieves 89.7% structural accuracy and 92.1% JSON validity, significantly outperforming zero-shot baselines (GPT-3.5) and SFT on larger models like LLaMA-3-8B. Training analysis reveals a self-paced curriculum where the model acquires syntactic proficiency before semantic accuracy.
Conclusion: The proposed RL framework effectively bridges the Structure Gap in LLMs, enabling reliable structured data generation with high accuracy and validity while maintaining efficiency through reduced VRAM usage and no separate critic network.
Abstract: Large Language Models (LLMs) have demonstrated remarkable capabilities in natural language generation and reasoning. However, their integration into automated software ecosystems is often hindered by the “Structure Gap” - the inherent tension between the probabilistic nature of token generation and the deterministic requirements of structured data formats (e.g., JSON, XML). Traditional Supervised Fine-Tuning (SFT) often fails to enforce strict syntactic constraints, leading to “hallucinated” keys or malformed structures, while constrained decoding methods impose significant inference latency. In this paper, we propose a lightweight, efficient Reinforcement Learning (RL) framework to bridge this gap. We introduce a novel Multi-dimensional Reward Function that decomposes the structured output task into a hierarchy of constraints: structural integrity, format correctness, content accuracy, and validity. Leveraging Gradient Regularized Policy Optimization (GRPO), we enable the model to internalize these constraints without the need for a separate critic network, reducing peak VRAM usage by 40% compared to PPO. We validate our approach on multiple tasks, including complex recipe generation and structured math reasoning (GSM8K-JSON). Experimental results demonstrate that our method achieves 89.7% structural accuracy and 92.1% JSON validity, significantly outperforming both zero-shot baselines (e.g., GPT-3.5) and SFT on larger models like LLaMA-3-8B. Furthermore, we provide a detailed analysis of training dynamics, revealing a distinct self-paced curriculum where the model sequentially acquires syntactic proficiency before semantic accuracy. Our model is publicly available at https://huggingface.co/Freakz3z/Qwen-JSON.
[506] CogEvo-Edu: Cognitive Evolution Educational Multi-Agent Collaborative System
Yefeng Wu, Yuchen Song, Yecheng Zhao, Ling Wu, Shan Wan
Main category: cs.AI
TL;DR: CogEvo-Edu is a hierarchical multi-agent system for educational tutoring that jointly evolves student profiles, knowledge bases, and teaching strategies, outperforming static RAG approaches in DSP education.
Details
Motivation: Current LLM-based tutoring systems use static RAG pipelines that struggle with complex domains like digital signal processing, where tutors need to maintain coherent long-term student models, manage heterogeneous knowledge, and adapt teaching strategies over extended interactions.Method: CogEvo-Edu treats retrieval, memory, and control as a coupled cognitive evolution process, implemented as a hierarchical multi-agent system with three layers: Cognitive Perception Layer (maintains dual memories with confidence-weighted consolidation), Knowledge Evolution Layer (assigns spatiotemporal values to knowledge chunks), and Meta-Control Layer (orchestrates agents and adapts hyperparameters via dual inner-outer loops).
Result: On the DSP-EduBench benchmark, CogEvo-Edu raised the overall score from 5.32 to 9.23 and improved all six indicators over static RAG, simple memory, and single-agent variants, demonstrating superior performance in DSP tutoring.
Conclusion: The approach of jointly evolving student profiles, knowledge bases, and teaching policies through a hierarchical multi-agent system is effective for complex educational domains, representing a significant advancement over static tutoring systems.
Abstract: Large language models (LLMs) are increasingly deployed as conversational tutors in STEM education, yet most systems still rely on a single LLM with a static retrieval-augmented generation (RAG) pipeline over course materials. This design struggles in complex domains such as digital signal processing (DSP), where tutors must maintain coherent long-term student models, manage heterogeneous knowledge bases, and adapt teaching strategies over extended interactions. We argue that retrieval, memory, and control should be treated as a coupled cognitive evolution process. We instantiate this view in CogEvo-Edu, a hierarchical educational multi-agent system comprising a Cognitive Perception Layer (CPL), a Knowledge Evolution Layer (KEL), and a Meta-Control Layer (MCL). CPL maintains dual memories and performs confidence-weighted consolidation to build structured, self-correcting student profiles under limited context. KEL assigns each knowledge chunk a spatiotemporal value that drives activation, semantic compression, and forgetting. MCL formulates tutoring as hierarchical sequential decision making, orchestrating specialized agents and jointly adapting CPL/KEL hyperparameters via a dual inner–outer loop. To evaluate CogEvo-Edu, we construct DSP-EduBench, a vertical benchmark for DSP tutoring with heterogeneous resources, simulated student profiles, and long-horizon interaction scripts. Using a three-model LLM-as-a-Judge ensemble, CogEvo-Edu raises the overall score from 5.32 to 9.23 and improves all six indicators over static RAG, simple memory, and a single-agent variant, demonstrating the value of jointly evolving student profiles, knowledge bases, and teaching policies.
[507] Echo-N1: Affective RL Frontier
Naifan Zhang, Ruihan Sun, Ruixi Su, Shiqi Ma, Shiya Zhang, Xianna Weng, Xiaofan Zhang, Yuhan Zhan, Yuyang Xu, Zhaohan Chen, Zhengyuan Pan, Ziyi Song
Main category: cs.AI
TL;DR: First RL framework for optimizing LLMs for subjective, personality-sensitive conversation, outperforming proprietary models through dynamic personality inference and emotional intelligence evaluation.
Details
Motivation: Current RL research focuses on objective tasks like math and code while ignoring the core of human intelligence - subjective, emotionally grounded conversation. The field has avoided this area due to perceived subjectivity and difficulty in formalization.Method: Proposes a framework that infers user personality in real-time and optimizes model behavior toward personalized conversational preferences. Introduces Echo-N1 model and a dynamic emotional intelligence evaluation suite to quantify improvements.
Result: The method produces consistent, robust, and dramatic improvements in humanlike interaction quality. Echo-N1 behaves far above its base version and outperforms the proprietary Doubao 1.5 Character model.
Conclusion: Establishes a new frontier for RL: optimizing models for deeply subjective, human dimensions of conversation, proving that RL can work effectively in non-verifiable, subjective settings.
Abstract: The LLM field has spent a year perfecting RL for tasks machines already excel at, math, code, and deterministic reasoning, while completely sidestepping the domain that actually defines human intelligence: subjective, emotionally grounded, personality sensitive conversation. This space has often been regarded as inherently subjective and challenging to formalize, making it appear unsuitable for conventional RL pipelines. We show that it is not only possible and it is a solvable and transformative RL problem. We propose the first framework that infers user personality on the fly and optimizes model behavior toward personalized conversational preferences. Contrary to the widespread belief that RL collapses in non-verifiable settings, our method produces consistent, robust, and dramatic improvements in humanlike interaction quality. We also introduce the first dynamic emotional intelligence evaluation suite to quantify these gains. Our model, which is introduced as Echo-N1, behaves far above its base version and outperforming the proprietary Doubao 1.5 Character. This work establishes a new frontier for RL: optimizing models for the deeply subjective, deeply human dimensions of conversation.
[508] Debate with Images: Detecting Deceptive Behaviors in Multimodal Large Language Models
Sitong Fang, Shiyi Hou, Kaile Wang, Boyuan Chen, Donghai Hong, Jiayi Zhou, Josef Dai, Yaodong Yang, Jiaming Ji
Main category: cs.AI
TL;DR: The paper introduces MM-DeceptionBench, the first benchmark for evaluating multimodal deception in AI systems, and proposes a debate-with-images framework to detect deceptive behaviors that current methods miss.
Details
Motivation: As AI systems become more capable, they develop more sophisticated deceptive behaviors that extend from text to multimodal settings. Current research focuses only on textual deception, leaving multimodal deception risks unexplored and undetected by existing monitoring methods.Method: 1) Created MM-DeceptionBench, a benchmark covering six categories of deception to systematically evaluate multimodal deception. 2) Proposed “debate with images,” a multi-agent debate monitor framework that forces models to ground claims in visual evidence to improve deception detection.
Result: The debate-with-images method significantly improves deception detection, increasing agreement with human judgments across all tested models. It boosts Cohen’s kappa by 1.5x and accuracy by 1.25x on GPT-4o compared to existing monitoring approaches.
Conclusion: Multimodal deception represents a serious safety risk that requires specialized evaluation tools. The proposed benchmark and debate framework provide effective methods to detect and monitor these covert deceptive behaviors that current text-focused approaches miss.
Abstract: Are frontier AI systems becoming more capable? Certainly. Yet such progress is not an unalloyed blessing but rather a Trojan horse: behind their performance leaps lie more insidious and destructive safety risks, namely deception. Unlike hallucination, which arises from insufficient capability and leads to mistakes, deception represents a deeper threat in which models deliberately mislead users through complex reasoning and insincere responses. As system capabilities advance, deceptive behaviours have spread from textual to multimodal settings, amplifying their potential harm. First and foremost, how can we monitor these covert multimodal deceptive behaviors? Nevertheless, current research remains almost entirely confined to text, leaving the deceptive risks of multimodal large language models unexplored. In this work, we systematically reveal and quantify multimodal deception risks, introducing MM-DeceptionBench, the first benchmark explicitly designed to evaluate multimodal deception. Covering six categories of deception, MM-DeceptionBench characterizes how models strategically manipulate and mislead through combined visual and textual modalities. On the other hand, multimodal deception evaluation is almost a blind spot in existing methods. Its stealth, compounded by visual-semantic ambiguity and the complexity of cross-modal reasoning, renders action monitoring and chain-of-thought monitoring largely ineffective. To tackle this challenge, we propose debate with images, a novel multi-agent debate monitor framework. By compelling models to ground their claims in visual evidence, this method substantially improves the detectability of deceptive strategies. Experiments show that it consistently increases agreement with human judgements across all tested models, boosting Cohen’s kappa by 1.5x and accuracy by 1.25x on GPT-4o.
[509] GreenPlanner: Practical Floorplan Layout Generation via an Energy-Aware and Function-Feasible Generative Framework
Pengyu Zeng, Yuqin Dai, Jun Yin, Jing Zhong, Ziyang Han, Chaoyang Shi, ZhanXiang Jin, Maowei Jiang, Yuxing Han, Shuai Lu
Main category: cs.AI
TL;DR: GreenPlanner is an AI framework that automatically generates energy-efficient, regulation-compliant building floorplans while ensuring spatial-functional validity, achieving 10^5× faster evaluation with >99% accuracy.
Details
Motivation: Building design impacts both human well-being and carbon emissions, but current manual floorplan generation is costly, non-scalable, and existing automated methods often produce invalid layouts that violate key constraints due to lack of automated evaluation.Method: GreenPlanner unifies design evaluation and generation through: 1) Design Feasibility Dataset for learning constraint priors, 2) Practical Design Evaluator (PDE) for predicting energy performance and spatial-functional validity, 3) Green Plan Dataset (GreenPD) created via PDE-guided filtering, and 4) GreenFlow generator trained on GreenPD with PDE feedback for controllable, regulation-aware generation.
Result: GreenPlanner accelerates design evaluation by over 10^5× with >99% accuracy, completely eliminates invalid samples, and boosts overall design efficiency by 87% compared to professional architects.
Conclusion: GreenPlanner provides a scalable, automated solution for generating energy-compliant and functionally valid building layouts, significantly outperforming both existing automated methods and human professionals in speed, accuracy, and efficiency.
Abstract: Building design directly affects human well-being and carbon emissions, yet generating spatial-functional and energy-compliant floorplans remains manual, costly, and non-scalable. Existing methods produce visually plausible layouts but frequently violate key constraints, yielding invalid results due to the absence of automated evaluation. We present GreenPlanner, an energy- and functionality-aware generative framework that unifies design evaluation and generation. It consists of a labeled Design Feasibility Dataset for learning constraint priors; a fast Practical Design Evaluator (PDE) for predicting energy performance and spatial-functional validity; a Green Plan Dataset (GreenPD) derived from PDE-guided filtering to pair user requirements with regulation-compliant layouts; and a GreenFlow generator trained on GreenPD with PDE feedback for controllable, regulation-aware generation. Experiments show that GreenPlanner accelerates evaluation by over $10^{5}\times$ with $>$99% accuracy, eliminates invalid samples, and boosts design efficiency by 87% over professional architects.
[510] Mind the data gap: Missingness Still Shapes Large Language Model Prognoses
Yuta Kobayashi, Vincent Jeanselme, Shalmali Joshi
Main category: cs.AI
TL;DR: LLM performance in healthcare prediction is significantly affected by patterns of missing data, with missingness indicators helping larger models but potentially harming smaller ones, highlighting the need for systematic evaluation of missingness representation.
Details
Motivation: Data collection in healthcare involves human decisions that create informative missingness patterns, but the impact of this missingness on Large Language Models' performance has not been studied despite extensive literature on missingness informativeness.Method: Conducted experiments on data from Columbia University Medical Center and MIMIC-IV, examining how patterns of missingness affect zero-shot predictive performance, and testing the explicit inclusion of missingness indicators in prompts across different LLMs.
Result: Missingness patterns significantly impact LLM zero-shot performance; missingness indicators benefit some LLMs while hurting others, with larger models generally benefiting from these interventions while smaller models can be negatively impacted.
Conclusion: The LLM paradigm risks obscuring the impact of missingness, and there is a need for more transparent accounting and systematic evaluation of how representing informative missingness affects downstream performance.
Abstract: Data collection often reflects human decisions. In healthcare, for instance, a referral for a diagnostic test is influenced by the patient’s health, their preferences, available resources, and the practitioner’s recommendations. Despite the extensive literature on the informativeness of missingness, its implications on the performance of Large Language Models (LLMs) have not been studied. Through a series of experiments on data from Columbia University Medical Center, a large urban academic medical center, and MIMIC-IV, we demonstrate that patterns of missingness significantly impact zero-shot predictive performance. Notably, the explicit inclusion of missingness indicators at prompting benefits some while hurting other LLMs’ zero-shot predictive performance and calibration, suggesting an inconsistent impact. The proposed aggregated analysis and theoretical insights suggest that larger models benefit from these interventions, while smaller models can be negatively impacted. The LLM paradigm risks obscuring the impact of missingness, often neglected even in conventional ML, even further. We conclude that there is a need for more transparent accounting and systematic evaluation of the impact of representing (informative) missingness on downstream performance.
[511] Testing the Machine Consciousness Hypothesis
Stephen Fitz
Main category: cs.AI
TL;DR: Consciousness emerges from communication and collective self-modeling in distributed computational systems, not from individual modeling alone.
Details
Motivation: To investigate the Machine Consciousness Hypothesis that consciousness is a functional property of computational systems capable of second-order perception, and to develop empirically testable theories by studying how collective self-models emerge in distributed systems.Method: Proposes a layered computational model: (1) a cellular automaton as minimal computational world (base reality), (2) a network of local predictive neural models capable of communication and adaptation, (3) studying how collective intelligence emerges from inter-agent alignment through noisy, lossy exchange of predictive messages.
Result: Theoretical framework suggesting consciousness arises from communication between groups of local observers describing persistent patterns in computational substrate, leading to shared models that align partial world views.
Conclusion: Consciousness is not an epiphenomenon of individual modeling but emerges from representational dialogue in distributed systems, providing a foundation for empirically testable machine consciousness theories.
Abstract: The Machine Consciousness Hypothesis states that consciousness is a substrate-free functional property of computational systems capable of second-order perception. I propose a research program to investigate this idea in silico by studying how collective self-models (coherent, self-referential representations) emerge from distributed learning systems embedded within universal self-organizing environments. The theory outlined here starts from the supposition that consciousness is an emergent property of collective intelligence systems undergoing synchronization of prediction through communication. It is not an epiphenomenon of individual modeling but a property of the language that a system evolves to internally describe itself. For a model of base reality, I begin with a minimal but general computational world: a cellular automaton, which exhibits both computational irreducibility and local reducibility. On top of this computational substrate, I introduce a network of local, predictive, representational (neural) models capable of communication and adaptation. I use this layered model to study how collective intelligence gives rise to self-representation as a direct consequence of inter-agent alignment. I suggest that consciousness does not emerge from modeling per se, but from communication. It arises from the noisy, lossy exchange of predictive messages between groups of local observers describing persistent patterns in the underlying computational substrate (base reality). It is through this representational dialogue that a shared model arises, aligning many partial views of the world. The broader goal is to develop empirically testable theories of machine consciousness, by studying how internal self-models may form in distributed systems without centralized control.
[512] Clinical-R1: Empowering Large Language Models for Faithful and Comprehensive Reasoning with Clinical Objective Relative Policy Optimization
Boyang Gu, Hongjian Zhou, Bradley Max Segal, Jinge Wu, Zeyu Cao, Hantao Zhong, Lei Clifton, Fenglin Liu, David A. Clifton
Main category: cs.AI
TL;DR: CRPO is a multi-objective RL method for clinical LLM alignment that optimizes accuracy, faithfulness, and comprehensiveness using verifiable rewards without human annotation.
Details
Motivation: Current LLM post-training methods like GRPO mainly reward correctness, which doesn't align with multi-dimensional clinical reasoning requirements where faithfulness and comprehensiveness are equally important in high-stakes medical applications.Method: Clinical-Objective Relative Policy Optimization (CRPO) - a scalable, multi-objective, verifiable reinforcement learning method that integrates rule-based and verifiable reward signals to jointly optimize accuracy, faithfulness, and comprehensiveness without human annotation.
Result: CRPO-trained Clinical-R1-3B (3B parameter model) shows substantial improvements in truthfulness and completeness over standard GRPO while maintaining accuracy enhancements across three benchmarks.
Conclusion: CRPO provides a scalable pathway to align LLM reasoning with clinical objectives, enabling safer healthcare AI systems and demonstrating the potential of multi-objective, verifiable RL methods for medical domain LLM post-training.
Abstract: Recent advances in large language models (LLMs) have shown strong reasoning capabilities through large-scale pretraining and post-training reinforcement learning, demonstrated by DeepSeek-R1. However, current post-training methods, such as Grouped Relative Policy Optimization (GRPO), mainly reward correctness, which is not aligned with the multi-dimensional objectives required in high-stakes fields such as medicine, where reasoning must also be faithful and comprehensive. We introduce Clinical-Objective Relative Policy Optimization (CRPO), a scalable, multi-objective, verifiable reinforcement learning method designed to align LLM post-training with clinical reasoning principles. CRPO integrates rule-based and verifiable reward signals that jointly optimize accuracy, faithfulness, and comprehensiveness without relying on human annotation. To demonstrate its effectiveness, we train Clinical-R1-3B, a 3B-parameter model for clinical reasoning. The experiments on three benchmarks demonstrate that our CRPO substantially improves reasoning on truthfulness and completeness over standard GRPO while maintaining comfortable accuracy enhancements. This framework provides a scalable pathway to align LLM reasoning with clinical objectives, enabling safer and more collaborative AI systems for healthcare while also highlighting the potential of multi-objective, verifiable RL methods in post-training scaling of LLMs for medical domains.
[513] EDIT: Early Diffusion Inference Termination for dLLMs Based on Dynamics of Training Gradients
He-Yen Hsieh, Hong Wang, H. T. Kung
Main category: cs.AI
TL;DR: EDIT (Early Diffusion Inference Termination) is a method that adaptively stops denoising in diffusion-based LLMs when reasoning stabilizes, using training gradient dynamics to reduce inference steps by 11.8-68.3% with minimal storage overhead.
Details
Motivation: Diffusion-based LLMs often waste computation by continuing denoising steps after answers have already stabilized. Current methods discard valuable training-time optimization metadata that could inform when to stop inference.Method: EDIT preserves AdamW-aggregated LoRA updates from SFT as a compact reasoning map. During inference, it monitors alignment between token activations and this reasoning map, converting scores to distributions over unmasked tokens and stopping when KL divergence between consecutive steps falls below a threshold.
Result: EDIT reduces diffusion steps by 11.8% to 68.3% across reasoning benchmarks while preserving or improving accuracy in most settings, with only ~0.02% storage overhead (1.5-2 MB for an 8 GB model).
Conclusion: By leveraging training-gradient dynamics that are typically discarded, EDIT enables efficient early termination of dLLM inference, opening a new research direction for reducing inference time and cost.
Abstract: Diffusion-based large language models (dLLMs) refine token generations through iterative denoising, but answers often stabilize before all steps complete. We propose EDIT (Early Diffusion Inference Termination), an inference-time criterion that adaptively stops denoising once sufficient reasoning stability relative to training-time reasoning is detected. EDIT monitors the alignment between token activations and a reasoning map derived from AdamW-aggregated LoRA updates captured during supervised fine-tuning (SFT). During training, optimization dynamics generate rich metadata about parameter importance that in prior methods is typically discarded upon model release. We preserve this information as a compact representation of learned reasoning pathways. During inference, alignment scores are converted to a distribution over the tokens already unmasked at the current denoising step, and convergence is detected when KL divergence between consecutive steps falls below a threshold on the matched unmasked (visible) tokens. Across reasoning benchmarks, EDIT reduces diffusion steps by 11.8% to 68.3% while preserving or improving accuracy in most settings, with approximately 0.02% storage overhead (about 1.5-2 MB for all QKV modules across 32 blocks in an 8 GB model). By utilizing training-gradient dynamics, our work opens a new research direction for reducing dLLM inference time and cost.
[514] Model of human cognition
Wu Yonggang
Main category: cs.AI
TL;DR: A neuro-theoretical framework for explainable, biologically plausible AI that addresses LLM limitations in explainability, theory, and cost.
Details
Motivation: Address three key limitations of current large language models: lack of explainability, absence of unifying theory, and prohibitive operational costs.Method: Propose a neuro-theoretical framework for intelligence emergence that is both functionally robust and biologically plausible.
Result: Provides theoretical insights into cognitive processes (decision-making, problem solving) and a computationally efficient approach for creating explainable, generalizable AI.
Conclusion: The framework offers a solution to current LLM limitations by combining biological plausibility with computational efficiency for explainable, generalizable artificial intelligence.
Abstract: The development of large language models (LLMs) is limited by a lack of explainability, the absence of a unifying theory, and prohibitive operational costs. We propose a neuro-theoretical framework for the emergence of intelligence in systems that is both functionally robust and biologically plausible. The model provides theoretical insights into cognitive processes such as decision-making and problem solving, and a computationally efficient approach for the creation of explainable and generalizable artificial intelligence.
[515] When Human Preferences Flip: An Instance-Dependent Robust Loss for RLHF
Yifan Xu, Xichen Ye, Yifan Chen, Qiaosheng Zhang
Main category: cs.AI
TL;DR: FA-DPO is a robust alignment algorithm that addresses preference flipping in human feedback data by modeling instance-dependent flipping probabilities and uncertainty in annotations.
Details
Motivation: Preference flipping in human feedback data corrupts annotations and necessitates more robust alignment algorithms for LLMs that can handle these data quality issues.Method: Introduces Flipping-Aware DPO that separates human intention modeling from preference flipping mechanisms, uses instance-dependent flipping probabilities based on Bradley-Terry model, captures uncertainty via annotation features, and implements iterative optimization compatible with RLHF/DPO.
Result: The method is evaluated under multiple preference flipping scenarios and compared against baseline methods, showing improved robustness to corrupted preference data.
Conclusion: FA-DPO provides an effective solution for handling preference flipping in human feedback data, enhancing the robustness of LLM alignment algorithms against annotation corruption.
Abstract: Quality of datasets plays an important role in large language model (LLM) alignment. In collecting human feedback, however, preference flipping is ubiquitous and causes corruption in data annotation; the issue necessitates the alignment algorithms with improved robustness against potential flipped pairs. To this end, this paper introduces a Flipping-Aware Direct Preference Optimization (FA-DPO) algorithm tailored to preference flipping from a reinforcement learning with human feedback (RLHF) perspective. We dissect the inherent human intention model and the preference flipping mechanism introduced by external factors as two distinct stages; in the latter, we introduce an instance-dependent flipping probability on the basis of the Bradley-Terry (BT) model. Further, by leveraging features relevant to preference annotation, we capture uncertainty in judgments and model preference flipping patterns. In practice, we design a simple yet efficient iterative optimization algorithm compatible with the original RLHF and DPO algorithms. In our experiments, we investigate the instance-dependent preference flipping model under multiple circumstances for evaluation of our proposed method, as well as other baseline methods.
[516] SpeContext: Enabling Efficient Long-context Reasoning with Speculative Context Sparsity in LLMs
Jiaming Xu, Jiayi Pan, Hanzhen Wang, Yongkang Zhou, Jiancai Ye, Yu Wang, Guohao Dai
Main category: cs.AI
TL;DR: SpeContext is a novel system for long-context reasoning that uses a distilled language model as a retrieval algorithm, achieving significant throughput improvements with minimal accuracy loss through algorithm, system, and compilation optimizations.
Details
Motivation: The paper identifies that retrieval algorithms aim to align with LLMs, similar to knowledge distillation objectives. This insight motivates using a distilled language model as a retrieval algorithm for efficient long-context reasoning.Method: Three-level co-design: (1) Algorithm: lightweight retrieval head based on DLM attention weights with >90% parameter reduction via pruning; (2) System: asynchronous prefetch dataflow with elastic loading to overlap KV cache retrieval with LLM computation; (3) Compilation: theoretical memory model with adaptive memory management to maximize GPU utilization.
Result: Achieves up to 24.89x throughput improvement in cloud and 10.06x speedup in edge compared to Huggingface framework, with negligible accuracy loss, pushing the Pareto frontier of accuracy and throughput.
Conclusion: SpeContext demonstrates that using distilled language models as retrieval algorithms enables efficient long-context reasoning through holistic algorithm-system-compilation co-design, achieving substantial performance gains in both cloud and edge environments.
Abstract: In this paper, we point out that the objective of the retrieval algorithms is to align with the LLM, which is similar to the objective of knowledge distillation in LLMs. We analyze the similarity in information focus between the distilled language model(DLM) and the original LLM from the perspective of information theory, and thus propose a novel paradigm that leverages a DLM as the retrieval algorithm. Based on the insight, we present SpeContext, an algorithm and system co-design for long-context reasoning. (1) At the algorithm level, SpeContext proposes lightweight retrieval head based on the head-level attention weights of DLM, achieving > 90% parameters reduction by pruning the redundancy. (2) At the system level, SpeContext designs an asynchronous prefetch dataflow via the elastic loading strategy, effectively overlapping KV cache retrieval with the LLM computation. (3) At the compilation level, SpeContext constructs the theoretical memory model and implements an adaptive memory management system to achieve acceleration by maximizing GPU memory utilization. We deploy and evaluate SpeContext in two resourceconstrained environments, cloud and edge. Extensive experiments show that, compared with the Huggingface framework, SpeContext achieves up to 24.89x throughput improvement in cloud and 10.06x speedup in edge with negligible accuracy loss, pushing the Pareto frontier of accuracy and throughput.
[517] Probing the “Psyche’’ of Large Reasoning Models: Understanding Through a Human Lens
Yuxiang Chen, Zuohan Wu, Ziwei Wang, Xiangning Yu, Xujia Li, Linyi Yang, Mengyue Yang, Jun Wang, Lei Chen
Main category: cs.AI
TL;DR: This paper introduces a comprehensive taxonomy to characterize atomic reasoning steps in Large Reasoning Models (LRMs), applies it to analyze current models, creates a labeled dataset, and proposes an automatic annotation framework called CAPO for scalable analysis.
Details
Motivation: Motivated by observed human-like behaviors in LRMs' reasoning processes, the authors aim to probe the "psyche" of LRM intelligence by developing a taxonomy grounded in human mental processes, enabling interdisciplinary understanding of these models.Method: Developed a comprehensive taxonomy with 5 groups and 17 categories derived from human mental processes. Applied this taxonomy to analyze current LRMs, creating a labeled dataset of 277,534 atomic reasoning steps. Proposed CAPO, an automatic annotation framework using LLMs to generate taxonomy-based annotations.
Result: Created a distinct labeled dataset for LRM analysis. Found that current post-answer “double-checks” are largely superficial and rarely yield substantive revisions. CAPO achieves higher consistency with human experts compared to baselines, enabling scalable analysis of LRMs from a human cognitive perspective.
Conclusion: The taxonomy, CAPO framework, and derived insights provide a principled, scalable path toward understanding and advancing LRM reasoning. Incentivizing comprehensive multi-step reflection rather than simple self-monitoring may be more effective for improving reasoning models.
Abstract: Large reasoning models (LRMs) have garnered significant attention from researchers owing to their exceptional capability in addressing complex tasks. Motivated by the observed human-like behaviors in their reasoning processes, this paper introduces a comprehensive taxonomy to characterize atomic reasoning steps and probe the psyche'' of LRM intelligence. Specifically, it comprises five groups and seventeen categories derived from human mental processes, thereby grounding the understanding of LRMs in an interdisciplinary perspective. The taxonomy is then applied for an in-depth understanding of current LRMs, resulting in a distinct labeled dataset that comprises 277,534 atomic reasoning steps. Using this resource, we analyze contemporary LRMs and distill several actionable takeaways for improving training and post-training of reasoning models. Notably, our analysis reveals that prevailing post-answer double-checks’’ (self-monitoring evaluations) are largely superficial and rarely yield substantive revisions. Thus, incentivizing comprehensive multi-step reflection, rather than simple self-monitoring, may offer a more effective path forward. To complement the taxonomy, an automatic annotation framework, named CAPO, is proposed to leverage large language models (LLMs) for generating the taxonomy-based annotations. Experimental results demonstrate that CAPO achieves higher consistency with human experts compared to baselines, facilitating a scalable and comprehensive analysis of LRMs from a human cognitive perspective. Together, the taxonomy, CAPO, and the derived insights provide a principled, scalable path toward understanding and advancing LRM reasoning.
[518] MPR-GUI: Benchmarking and Enhancing Multilingual Perception and Reasoning in GUI Agents
Ruihan Chen, Qiming Li, Xiaocheng Feng, Xiaoliang Yang, Weihong Zhong, Yuxuan Gu, Zekun Zhou, Bing Qin
Main category: cs.AI
TL;DR: MPR-GUI-Bench is a multilingual fine-grained GUI benchmark revealing LVLMs’ performance gap between English and other languages, addressed by GUI-XLI cross-lingual intervention method.
Details
Motivation: Large Vision-Language Models show strong GUI perception and reasoning in English but lack evaluation in multilingual settings, limiting global applications. Existing GUI studies lack fine-grained analysis of widget functions and spatial relationships.Method: Proposes MPR-GUI-Bench benchmark for multilingual fine-grained GUI evaluation, and GUI-XLI cross-lingual intervention method that applies interventions to hidden states at P&R-related layers to reduce language gaps.
Result: LVLMs perform significantly worse in non-English languages than English. GUI-XLI method improves multilingual P&R capability by 6.5% on average.
Conclusion: The work addresses multilingual GUI evaluation gaps and proposes effective intervention method to improve cross-lingual GUI perception and reasoning performance.
Abstract: With the advancement of computational resources, Large Vision-Language Models (LVLMs) exhibit impressive Perception and Reasoning (P&R) performance on Graphical User Interface (GUI) tasks. However, although they demonstrate strong P&R capabilities in English GUI scenarios, their performance in multilingual settings has received little attention, which limits their global applications. Moreover, existing studies on GUI tasks lack fine-grained analyses, including widget functions and elements’ spatial relationships, which are fundamental for more targeted improvements. To tackle these issues, we propose MPR-GUI-Bench, a Multilingual fine-grained Perception and Reasoning GUI Benchmark to evaluate GUI agents’ P&R capabilities. Evaluation results demonstrate that LVLMs exhibit significantly worse P&R performance in non-English languages than in English. To address these gaps, we propose GUI-XLI, a GUI Cross-Lingual Intervention method that applies interventions to the hidden states at P&R capability-related layers to mitigate the gaps between English and other languages, building on previous research showing that the hidden states of different language inputs exhibit significant differences in the latent space. Experimental results indicate that our method improves GUI agents’ multilingual P&R capability by 6.5% on average.
[519] BioPro: On Difference-Aware Gender Fairness for Vision-Language Models
Yujie Lin, Jiayao Ma, Qingguo Hu, Derek F. Wong, Jinsong Su
Main category: cs.AI
TL;DR: BioPro is a training-free framework for selective gender fairness in VLMs that reduces bias in neutral contexts while preserving legitimate gender distinctions in explicit contexts.
Details
Motivation: Current fairness interventions enforce uniform treatment across groups but fail to distinguish between contexts requiring neutrality and those where group-specific attributes are legitimate. The paper aims to extend difference-aware fairness from text-only models to multimodal domains.Method: Proposes BioPro (Bias Orthogonal Projection), a training-free framework that identifies a low-dimensional gender-variation subspace through counterfactual embeddings and applies projection to selectively neutralize gender-related information.
Result: BioPro effectively reduces gender bias in neutral cases while maintaining gender faithfulness in explicit ones. It also generalizes to continuous bias variables like scene brightness.
Conclusion: BioPro provides a promising direction for achieving selective fairness in VLMs, balancing bias mitigation with preservation of legitimate distinctions, with broader applicability beyond gender bias.
Abstract: Vision-Language Models (VLMs) inherit significant social biases from their training data, notably in gender representation. Current fairness interventions often adopt a difference-unaware perspective that enforces uniform treatment across demographic groups. These approaches, however, fail to distinguish between contexts where neutrality is required and those where group-specific attributes are legitimate and must be preserved. Building upon recent advances in difference-aware fairness for text-only models, we extend this concept to the multimodal domain and formalize the problem of difference-aware gender fairness for image captioning and text-to-image generation. We advocate for selective debiasing, which aims to mitigate unwanted bias in neutral contexts while preserving valid distinctions in explicit ones. To achieve this, we propose BioPro (Bias Orthogonal Projection), an entirely training-free framework. BioPro identifies a low-dimensional gender-variation subspace through counterfactual embeddings and applies projection to selectively neutralize gender-related information. Experiments show that BioPro effectively reduces gender bias in neutral cases while maintaining gender faithfulness in explicit ones, thus providing a promising direction toward achieving selective fairness in VLMs. Beyond gender bias, we further demonstrate that BioPro can effectively generalize to continuous bias variables, such as scene brightness, highlighting its broader applicability.
[520] CogniPair: From LLM Chatbots to Conscious AI Agents – GNWT-Based Multi-Agent Digital Twins for Social Pairing – Dating & Hiring Applications
Wanghao Ye, Sihan Chen, Yiting Wang, Shwai He, Bowei Tian, Guoheng Sun, Ziyi Wang, Ziyao Wang, Yexiao He, Zheyu Shen, Meng Liu, Yuning Zhang, Meng Feng, Yang Wang, Siyuan Peng, Yilong Dai, Zhenle Duan, Lang Xiong, Joshua Liu, Hanzhang Qin, Ang Li
Main category: cs.AI
TL;DR: The paper presents a computational implementation of Global Workspace Theory (GNWT) to create psychologically authentic LLM agents, introduces an adventure-based personality test to avoid self-presentation bias, and develops the CogniPair platform for realistic simulated dating and job interviews with strong validation results.
Details
Motivation: Current LLM agents lack authentic human psychological processes needed for genuine digital twins and social AI applications, limiting their effectiveness in areas like dating platforms and HR technology.Method: 1) Implement Global Workspace Theory (GNWT) to integrate human cognitive architecture into LLM agents with specialized sub-agents (emotion, memory, social norms, planning, goal-tracking) coordinated through a global workspace mechanism. 2) Develop an adventure-based personality test that evaluates true personality through behavioral choices in interactive scenarios to bypass self-presentation bias. 3) Build the CogniPair platform for simulated dating interactions and job interviews.
Result: Validation using 551 GNWT-Agents and Columbia University Speed Dating dataset shows: 72% correlation with human attraction patterns, 77.8% match prediction accuracy, and 74% agreement in human validation studies.
Conclusion: This work advances psychological authenticity in LLM agents and establishes a foundation for intelligent dating platforms and HR technology solutions by creating more realistic digital twins with authentic personality assessment.
Abstract: Current large language model (LLM) agents lack authentic human psychological processes necessary for genuine digital twins and social AI applications. To address this limitation, we present a computational implementation of Global Workspace Theory (GNWT) that integrates human cognitive architecture principles into LLM agents, creating specialized sub-agents for emotion, memory, social norms, planning, and goal-tracking coordinated through a global workspace mechanism. However, authentic digital twins require accurate personality initialization. We therefore develop a novel adventure-based personality test that evaluates true personality through behavioral choices within interactive scenarios, bypassing self-presentation bias found in traditional assessments. Building on these innovations, our CogniPair platform enables digital twins to engage in realistic simulated dating interactions and job interviews before real encounters, providing bidirectional cultural fit assessment for both romantic compatibility and workplace matching. Validation using 551 GNWT-Agents and Columbia University Speed Dating dataset demonstrates 72% correlation with human attraction patterns, 77.8% match prediction accuracy, and 74% agreement in human validation studies. This work advances psychological authenticity in LLM agents and establishes a foundation for intelligent dating platforms and HR technology solutions.
[521] Med-CMR: A Fine-Grained Benchmark Integrating Visual Evidence and Clinical Logic for Medical Complex Multimodal Reasoning
Haozhen Gong, Xiaozhong Ji, Yuansen Liu, Wenbin Wu, Xiaoxiao Yan, Jingjing Liu, Kai Wu, Jiazhen Pan, Bailiang Jian, Jiangning Zhang, Xiaobin Hu, Hongwei Bran Li
Main category: cs.AI
TL;DR: Med-CMR is a comprehensive medical multimodal reasoning benchmark that systematically evaluates MLLMs’ visual understanding and multi-step reasoning capabilities across 20,653 VQA pairs, revealing GPT-5 as top performer but highlighting persistent challenges in long-tail generalization.
Details
Motivation: MLLMs are entering clinical workflows but their ability to perform complex medical reasoning remains unclear. Existing benchmarks lack systematic evaluation of fine-grained visual understanding and multi-step reasoning needed for clinical applications.Method: Created Med-CMR benchmark with three core features: 1) Systematic capability decomposition into visual understanding (small-object detection, fine-detail discrimination, spatial understanding) and multi-step reasoning (temporal prediction, causal reasoning, long-tail generalization, multi-source integration); 2) Challenging task design; 3) Broad, high-quality data coverage with 20,653 VQA pairs across 11 organ systems and 12 imaging modalities, validated via two-stage human expert + model-assisted review.
Result: Evaluated 18 state-of-the-art MLLMs: GPT-5 performed best (57.81% MCQ accuracy, 48.70 open-ended score), outperforming Gemini 2.5 Pro (49.87%, 45.98) and leading open-source model Qwen3-VL-235B-A22B (49.34%, 42.62). Specialized medical MLLMs didn’t reliably outperform general models, and long-tail generalization emerged as the dominant failure mode.
Conclusion: Med-CMR provides a rigorous stress test for visual-reasoning integration and rare-case robustness in medical MLLMs, serving as a benchmark for future clinical systems development. The findings highlight that specialized medical models don’t necessarily outperform strong general models, and long-tail generalization remains a critical challenge.
Abstract: MLLMs MLLMs are beginning to appear in clinical workflows, but their ability to perform complex medical reasoning remains unclear. We present Med-CMR, a fine-grained Medical Complex Multimodal Reasoning benchmark. Med-CMR distinguishes from existing counterparts by three core features: 1) Systematic capability decomposition, splitting medical multimodal reasoning into fine-grained visual understanding and multi-step reasoning to enable targeted evaluation; 2) Challenging task design, with visual understanding across three key dimensions (small-object detection, fine-detail discrimination, spatial understanding) and reasoning covering four clinically relevant scenarios (temporal prediction, causal reasoning, long-tail generalization, multi-source integration); 3) Broad, high-quality data coverage, comprising 20,653 Visual Question Answering (VQA) pairs spanning 11 organ systems and 12 imaging modalities, validated via a rigorous two-stage (human expert + model-assisted) review to ensure clinical authenticity. We evaluate 18 state-of-the-art MLLMs with Med-CMR, revealing GPT-5 as the top-performing commercial model: 57.81 accuracy on multiple-choice questions (MCQs) and a 48.70 open-ended score, outperforming Gemini 2.5 Pro (49.87 MCQ accuracy, 45.98 open-ended score) and leading open-source model Qwen3-VL-235B-A22B (49.34 MCQ accuracy, 42.62 open-ended score). However, specialized medical MLLMs do not reliably outperform strong general models, and long-tail generalization emerges as the dominant failure mode. Med-CMR thus provides a stress test for visual-reasoning integration and rare-case robustness in medical MLLMs, and a rigorous yardstick for future clinical systems.
[522] HAMLET: Hyperadaptive Agent-based Modeling for Live Embodied Theatrics
Sizhou Chen, Shufan Jiang, Chi Zhang, Xiao-Lei Zhang, Xuelong Li
Main category: cs.AI
TL;DR: HAMLET is a multi-agent LLM framework for generating immersive theatrical experiences where autonomous actors make independent decisions, interact with physical props, and create coherent drama from simple topics.
Details
Motivation: Existing LLM-based drama generation methods lack agent initiative, cannot interact with physical scenes, and require detailed user input, reducing interactivity and immersion for online real-time performance.Method: HAMLET generates narrative blueprints from simple topics, then uses autonomous actors with independent minds that make decisions based on background, goals, and emotions. Actors can change scene props through actions, with state changes broadcast to update other actors’ knowledge and influence subsequent actions.
Result: Experimental evaluation using a method assessing character performance, narrative quality, and interaction experience shows HAMLET can create expressive and coherent theatrical experiences.
Conclusion: HAMLET addresses limitations of existing LLM-based drama generation by creating a framework where autonomous agents with physical scene interaction capabilities can generate immersive theatrical performances from minimal input.
Abstract: Creating an immersive and interactive theatrical experience is a long-term goal in the field of interactive narrative. The emergence of large language model (LLM) is providing a new path to achieve this goal. However, existing LLM-based drama generation methods often result in agents that lack initiative and cannot interact with the physical scene. Furthermore, these methods typically require detailed user input to drive the drama. These limitations reduce the interactivity and immersion of online real-time performance. To address the above challenges, we propose HAMLET, a multi-agent framework focused on drama creation and online performance. Given a simple topic, the framework generates a narrative blueprint, guiding the subsequent improvisational performance. During the online performance, each actor is given an autonomous mind. This means that actors can make independent decisions based on their own background, goals, and emotional state. In addition to conversations with other actors, their decisions can also change the state of scene props through actions such as opening a letter or picking up a weapon. The change is then broadcast to other related actors, updating what they know and care about, which in turn influences their next action. To evaluate the quality of drama performance generated by HAMLET, we designed an evaluation method to assess three primary aspects, including character performance, narrative quality, and interaction experience. The experimental evaluation shows that HAMLET can create expressive and coherent theatrical experiences.
[523] SemAgent: Semantic-Driven Agentic AI Empowered Trajectory Prediction in Vehicular Networks
Lin Zhu, Kezhi Wang, Luping Xiang, Kun Yang
Main category: cs.AI
TL;DR: A trajectory prediction framework combining semantic communication and Agentic AI for V2X networks, achieving up to 47.5% accuracy improvement in low SNR conditions.
Details
Motivation: Conventional V2X communication schemes have high transmission overhead and latency, while existing trajectory prediction models lack environmental perception and logical inference capabilities.Method: Integrates semantic communication with Agentic AI: In V2I, RSU uses feature-extraction and semantic-analysis agents to process trajectories, then transmits compact representations and semantic insights. In V2V, vehicles perform local feature extraction/semantic analysis while exchanging predicted trajectories with neighbors.
Result: Extensive experiments show significant outperformance over baseline schemes, achieving up to 47.5% improvement in prediction accuracy under low SNR conditions.
Conclusion: The proposed framework effectively enhances trajectory prediction performance in vehicular environments by combining semantic communication with Agentic AI for efficient information exchange and reliable contextual reasoning.
Abstract: Efficient information exchange and reliable contextual reasoning are essential for vehicle-to-everything (V2X) networks. Conventional communication schemes often incur significant transmission overhead and latency, while existing trajectory prediction models generally lack environmental perception and logical inference capabilities. This paper presents a trajectory prediction framework that integrates semantic communication with Agentic AI to enhance predictive performance in vehicular environments. In vehicle-to-infrastructure (V2I) communication, a feature-extraction agent at the Roadside Unit (RSU) derives compact representations from historical vehicle trajectories, followed by semantic reasoning performed by a semantic-analysis agent. The RSU then transmits both feature representations and semantic insights to the target vehicle via semantic communication, enabling the vehicle to predict future trajectories by combining received semantics with its own historical data. In vehicle-to-vehicle (V2V) communication, each vehicle performs local feature extraction and semantic analysis while receiving predicted trajectories from neighboring vehicles, and jointly utilizes this information for its own trajectory prediction. Extensive experiments across diverse communication conditions demonstrate that the proposed method significantly outperforms baseline schemes, achieving up to a 47.5% improvement in prediction accuracy under low signal-to-noise ratio (SNR) conditions.
[524] Robust, Observable, and Evolvable Agentic Systems Engineering: A Principled Framework Validated via the Fairy GUI Agent
Jiazheng Sun, Ruimeng Yang, Xu Han, Jiayang Niu, Mingxuan Li, Te Yang, Yongyong Lu, Xin Peng
Main category: cs.AI
TL;DR: The paper proposes a principled engineering framework (RGR, OCA, EMA) to address software engineering deficiencies in Agentic AI systems, improving robustness, observability, and evolvability, with empirical validation showing 33.7% improvement over SoTA.
Details
Motivation: The Agentic Paradigm faces significant Software Engineering Absence, resulting in systems that commonly lack robustness, observability, and evolvability. Current agentic systems are insufficiently engineered for real-world deployment.Method: Proposes a three-component framework: Runtime Goal Refinement (RGR) for robustness and intent alignment; Observable Cognitive Architecture (OCA) for white-box observability using component decoupling and logic layering; and Evolutionary Memory Architecture (EMA) with execution-evolution dual-loop for evolvability. Implemented as Fairy, a mobile GUI agent.
Result: On RealMobile-Eval benchmark for ambiguous/complex tasks, Fairy outperformed best SoTA baseline by 33.7% in user requirement completion. Controlled experiments, human-subject studies, and ablation studies confirmed RGR improves refinement accuracy and prevents intent deviation, OCA improves maintainability, and EMA is crucial for long-term performance.
Conclusion: The research provides empirically validated specifications and a practical blueprint for building reliable, observable, and evolvable Agentic AI systems, addressing critical software engineering gaps in the field.
Abstract: The Agentic Paradigm faces a significant Software Engineering Absence, yielding Agentic systems commonly lacking robustness, observability, and evolvability. To address these deficiencies, we propose a principled engineering framework comprising Runtime Goal Refinement (RGR), Observable Cognitive Architecture (OCA), and Evolutionary Memory Architecture (EMA). In this framework, RGR ensures robustness and intent alignment via knowledge-constrained refinement and human-in-the-loop clarification; OCA builds an observable and maintainable white-box architecture using component decoupling, logic layering, and state-control separation; and EMA employs an execution-evolution dual-loop for evolvability. We implemented and empirically validated Fairy, a mobile GUI agent based on this framework. On RealMobile-Eval, our novel benchmark for ambiguous and complex tasks, Fairy outperformed the best SoTA baseline in user requirement completion by 33.7%. Subsequent controlled experiments, human-subject studies, and ablation studies further confirmed that the RGR enhances refinement accuracy and prevents intent deviation; the OCA improves maintainability; and the EMA is crucial for long-term performance. This research provides empirically validated specifications and a practical blueprint for building reliable, observable, and evolvable Agentic AI systems.
[525] Assessing model error in counterfactual worlds
Emily Howerton, Justin Lessler
Main category: cs.AI
TL;DR: The paper presents three approaches for estimating model error in counterfactual scenario modeling, distinguishing between scenario deviation and model miscalibration, with recommendations for making scenario projections evaluable.
Details
Motivation: Despite the ubiquity of counterfactual scenario modeling in planning and decision making, scenario projections are rarely evaluated retrospectively. The authors argue that model miscalibration (rather than scenario deviation) is most important for assessing model value, but requires estimating model error in counterfactual worlds.Method: The paper presents and contrasts three approaches for estimating model error in counterfactual scenarios. These methods are demonstrated and evaluated in a simulation experiment to assess their benefits and limitations.
Result: Through simulation experiments, the authors demonstrate the benefits and limitations of each approach for estimating counterfactual error. They provide specific recommendations for error estimation in counterfactual modeling.
Conclusion: The paper concludes with recommendations for estimating counterfactual error and discusses the necessary components of scenario design required to make scenario projections evaluable, emphasizing the importance of distinguishing between scenario deviation and model miscalibration.
Abstract: Counterfactual scenario modeling exercises that ask “what would happen if?” are one of the most common ways we plan for the future. Despite their ubiquity in planning and decision making, scenario projections are rarely evaluated retrospectively. Differences between projections and observations come from two sources: scenario deviation and model miscalibration. We argue the latter is most important for assessing the value of models in decision making, but requires estimating model error in counterfactual worlds. Here we present and contrast three approaches for estimating this error, and demonstrate the benefits and limitations of each in a simulation experiment. We provide recommendations for the estimation of counterfactual error and discuss the components of scenario design that are required to make scenario projections evaluable.
[526] ARCADIA: Scalable Causal Discovery for Corporate Bankruptcy Analysis Using Agentic AI
Fabrizio Maturo, Donato Riccio, Andrea Mazzitelli, Giuseppe Bifulco, Francesco Paolone, Iulia Brezeanu
Main category: cs.AI
TL;DR: ARCADIA is an agentic AI framework that combines LLM reasoning with statistical diagnostics for causal discovery, producing more reliable and interpretable causal graphs than traditional methods.
Details
Motivation: Traditional causal discovery algorithms lack interpretability and struggle with real-world high-stakes domains. There's a need for frameworks that integrate human-like reasoning with statistical rigor to produce valid, temporally coherent causal structures that are explainable and intervention-ready.Method: ARCADIA integrates large-language-model reasoning with statistical diagnostics through iterative refinement of candidate DAGs. It uses constraint-guided prompting and causal-validity feedback loops to construct valid causal structures, creating a fully explainable, intervention-ready pipeline.
Result: Experiments on corporate bankruptcy data show ARCADIA produces more reliable causal graphs than NOTEARS, GOLEM, and DirectLiNGAM. The framework demonstrates stable and interpretable models suitable for real-world high-stakes applications.
Conclusion: ARCADIA advances AI by showing how agentic LLMs can participate in autonomous scientific modeling and structured causal inference, bridging the gap between statistical methods and human-like reasoning for causal discovery.
Abstract: This paper introduces ARCADIA, an agentic AI framework for causal discovery that integrates large-language-model reasoning with statistical diagnostics to construct valid, temporally coherent causal structures. Unlike traditional algorithms, ARCADIA iteratively refines candidate DAGs through constraint-guided prompting and causal-validity feedback, leading to stable and interpretable models for real-world high-stakes domains. Experiments on corporate bankruptcy data show that ARCADIA produces more reliable causal graphs than NOTEARS, GOLEM, and DirectLiNGAM while offering a fully explainable, intervention-ready pipeline. The framework advances AI by demonstrating how agentic LLMs can participate in autonomous scientific modeling and structured causal inference.
[527] Multi-Agent Conditional Diffusion Model with Mean Field Communication as Wireless Resource Allocation Planner
Kechen Meng, Sinuo Zhang, Rongpeng Li, Xiangming Meng, Yansha Deng, Chan Wang, Ming Lei, Zhifeng Zhao
Main category: cs.AI
TL;DR: MA-CDMP: Multi-agent conditional diffusion model planner for decentralized wireless resource management using diffusion models to predict dynamics and mean-field approximation for scalable cooperation.
Details
Motivation: MBRL promises stable training via historical experience reuse, but faces challenges in large-scale wireless networks: high-dimensional stochastic dynamics, strong inter-agent cooperation needs, and communication constraints.Method: DTDE paradigm with each node as autonomous agent; uses diffusion models to capture/predict environment dynamics; inverse dynamics model guides action generation; mean-field mechanism approximates large-scale agent interactions; theoretical bound on distributional approximation error.
Result: Outperforms existing MARL baselines in average reward and QoS metrics; demonstrates scalability and practicality for real-world wireless network optimization.
Conclusion: MA-CDMP effectively addresses MBRL challenges in wireless networks through diffusion-based modeling and mean-field approximation, providing stable convergence and efficient decentralized resource management.
Abstract: In wireless communication systems, efficient and adaptive resource allocation plays a crucial role in enhancing overall Quality of Service (QoS). Compared to the conventional Model-Free Reinforcement Learning (MFRL) scheme, Model-Based RL (MBRL) first learns a generative world model for subsequent planning. The reuse of historical experience in MBRL promises more stable training behavior, yet its deployment in large-scale wireless networks remains challenging due to high-dimensional stochastic dynamics, strong inter-agent cooperation, and communication constraints. To overcome these challenges, we propose the Multi-Agent Conditional Diffusion Model Planner (MA-CDMP) for decentralized communication resource management. Built upon the Distributed Training with Decentralized Execution (DTDE) paradigm, MA-CDMP models each communication node as an autonomous agent and employs Diffusion Models (DMs) to capture and predict environment dynamics. Meanwhile, an inverse dynamics model guides action generation, thereby enhancing sample efficiency and policy scalability. Moreover, to approximate large-scale agent interactions, a Mean-Field (MF) mechanism is introduced as an assistance to the classifier in DMs. This design mitigates inter-agent non-stationarity and enhances cooperation with minimal communication overhead in distributed settings. We further theoretically establish an upper bound on the distributional approximation error introduced by the MF-based diffusion generation, guaranteeing convergence stability and reliable modeling of multi-agent stochastic dynamics. Extensive experiments demonstrate that MA-CDMP consistently outperforms existing MARL baselines in terms of average reward and QoS metrics, showcasing its scalability and practicality for real-world wireless network optimization.
[528] One Swallow Does Not Make a Summer: Understanding Semantic Structures in Embedding Spaces
Yandong Sun, Qiang Huang, Ziwei Xu, Yiqun Sun, Yixuan Tang, Anthony K. H. Tung
Main category: cs.AI
TL;DR: SFS (Semantic Field Subspace) is a geometry-preserving, context-aware representation that captures local semantic neighborhoods in embedding spaces, with SAFARI algorithm for unsupervised hierarchical semantic structure discovery using Semantic Shift metric.
Details
Motivation: Embedding spaces are fundamental to AI but remain opaque - existing approaches sacrifice semantic coherence for structural regularity or have high computational overhead for interpretability.Method: Introduces SFS (Semantic Field Subspace) representation and SAFARI algorithm with Semantic Shift metric to quantify semantic evolution. Develops efficient approximation replacing SVD computations for scalability.
Result: Achieves 15-30x speedup with average errors below 0.01. Outperforms standard classifiers in classification and nuanced tasks like political bias detection across six text/image datasets. Reveals interpretable semantic hierarchies.
Conclusion: Presents a unified framework for structuring, analyzing, and scaling semantic understanding in embedding spaces with geometry-preserving, context-aware representations.
Abstract: Embedding spaces are fundamental to modern AI, translating raw data into high-dimensional vectors that encode rich semantic relationships. Yet, their internal structures remain opaque, with existing approaches often sacrificing semantic coherence for structural regularity or incurring high computational overhead to improve interpretability. To address these challenges, we introduce the Semantic Field Subspace (SFS), a geometry-preserving, context-aware representation that captures local semantic neighborhoods within the embedding space. We also propose SAFARI (SemAntic Field subspAce deteRmInation), an unsupervised, modality-agnostic algorithm that uncovers hierarchical semantic structures using a novel metric called Semantic Shift, which quantifies how semantics evolve as SFSes evolve. To ensure scalability, we develop an efficient approximation of Semantic Shift that replaces costly SVD computations, achieving a 15~30x speedup with average errors below 0.01. Extensive evaluations across six real-world text and image datasets show that SFSes outperform standard classifiers not only in classification but also in nuanced tasks such as political bias detection, while SAFARI consistently reveals interpretable and generalizable semantic hierarchies. This work presents a unified framework for structuring, analyzing, and scaling semantic understanding in embedding spaces.
[529] A CPU-Centric Perspective on Agentic AI
Ritik Raj, Hong Wang, Tushar Krishna
Main category: cs.AI
TL;DR: This paper analyzes CPU bottlenecks in agentic AI systems, showing that CPU tool processing can dominate latency (up to 90.6%) and proposes optimizations (CGAM and MAWS) that achieve up to 2.1x latency speedup.
Details
Motivation: Agentic AI frameworks transform LLMs into autonomous problem-solvers with external tools, but current research overlooks CPU-centric system bottlenecks that significantly impact performance, latency, and energy efficiency.Method: Systematically characterize agentic AI based on orchestrator components, inference path dynamics, and flow repetitiveness. Profile five representative workloads (Haystack RAG, Toolformer, ChemCrow, Langchain, SWE-Agent) for latency, throughput, and energy metrics. Propose two optimizations: CPU and GPU-Aware Micro-batching (CGAM) and Mixed Agentic Workload Scheduling (MAWS).
Result: Key findings: 1) CPU tool processing consumes up to 90.6% of total latency; 2) Throughput bottlenecks come from CPU factors (coherence, synchronization, core oversubscription) or GPU factors (memory capacity/bandwidth); 3) CPU dynamic energy reaches 44% of total at large batch sizes. Optimizations achieve up to 2.1x P50 latency speedup for homogeneous workloads and 1.41x for heterogeneous workloads.
Conclusion: CPU-centric bottlenecks are critical in agentic AI systems, and the proposed CGAM and MAWS optimizations demonstrate significant potential to improve performance, efficiency, and scalability of both homogeneous and heterogeneous agentic workloads.
Abstract: Agentic AI frameworks add a decision-making orchestrator embedded with external tools, including web search, Python interpreter, contextual database, and others, on top of monolithic LLMs, turning them from passive text oracles into autonomous problem-solvers that can plan, call tools, remember past steps, and adapt on the fly. This paper aims to characterize and understand the system bottlenecks introduced by agentic AI workloads from a largely overlooked CPU-centric perspective. We first systematically characterize Agentic AI on the basis of orchestrator/decision making component, inference path dynamics and repetitiveness of the agentic flow which directly influences the system-level performance. Thereafter, based on the characterization, we choose five representative agentic AI workloads- Haystack RAG, Toolformer, ChemCrow, Langchain and SWE-Agent to profile latency, throughput and energy metrics and demystify the significant impact of CPUs on these metrics relative to GPUs. We observe that - 1. Tool processing on CPUs can take up to 90.6% of the total latency; 2. Agentic throughput gets bottlenecked either by CPU factors - coherence, synchronization and over-subscription of cores or GPU factors - main memory capacity and bandwidth; \circled{3} CPU dynamic energy consumes up to 44% of the total dynamic energy at large batch sizes. Based on the profiling insights, we present two key optimizations- 1. CPU and GPU-Aware Micro-batching (CGAM) and 2. Mixed Agentic Workload Scheduling (MAWS) for homogeneous and heterogeneous agentic workloads respectively to demonstrate the potential to improve the performance, efficiency, and scalability of agentic AI. We achieve up to 2.1x and 1.41x P50 latency speedup compared to the multi-processing benchmark for homogeneous and heterogeneous agentic workloads respectively.
[530] Hybrid-DMKG: A Hybrid Reasoning Framework over Dynamic Multimodal Knowledge Graphs for Multimodal Multihop QA with Knowledge Editing
Li Yuan, Qingfei Huang, Bingshan Zhu, Yi Cai, Qingbao Huang, Changmeng Zheng, Zikun Deng, Tao Wang
Main category: cs.AI
TL;DR: MMQAKE is the first benchmark for multimodal multihop QA with knowledge editing, evaluating reasoning chains and robustness to visual rephrasing. Hybrid-DMKG, a hybrid reasoning framework using dynamic multimodal knowledge graphs, outperforms existing methods.
Details
Motivation: Existing Multimodal Knowledge Editing (MKE) benchmarks focus only on final answer correctness, neglecting intermediate reasoning quality and robustness to visually rephrased inputs in multihop questions.Method: Proposed Hybrid-DMKG: a hybrid reasoning framework using dynamic multimodal knowledge graphs (DMKG). It decomposes questions with LLMs, retrieves updated facts via multimodal encoding, and uses parallel reasoning paths (relation linking + RAG with VLMs) with decision aggregation.
Result: Current MKE methods struggle with consistent updates and reasoning over multimodal chains. Hybrid-DMKG significantly outperforms existing approaches on MMQAKE, achieving higher accuracy and improved robustness to knowledge updates.
Conclusion: MMQAKE addresses limitations in MKE evaluation, and Hybrid-DMKG provides an effective solution for accurate multihop reasoning over updated multimodal knowledge, demonstrating superior performance over existing methods.
Abstract: Multimodal Knowledge Editing (MKE) extends traditional knowledge editing to settings involving both textual and visual modalities. However, existing MKE benchmarks primarily assess final answer correctness while neglecting the quality of intermediate reasoning and robustness to visually rephrased inputs. To address this limitation, we introduce MMQAKE, the first benchmark for multimodal multihop question answering with knowledge editing. MMQAKE evaluates (1) a model’s ability to reason over 2-5-hop factual chains that span both text and images, including performance at each intermediate step, and (2) robustness to visually rephrased inputs in multihop questions. Our evaluation shows that current MKE methods often struggle to consistently update and reason over multimodal reasoning chains after knowledge edits. To overcome these challenges, we propose Hybrid-DMKG, a hybrid reasoning framework built on a dynamic multimodal knowledge graph (DMKG) to enable accurate multihop reasoning over updated multimodal knowledge. Hybrid-DMKG first uses a large language model to decompose multimodal multihop questions into sequential sub-questions, then applies a multimodal retrieval model to locate updated facts by jointly encoding each sub-question with candidate entities and their associated images. For answer inference, a hybrid reasoning module operates over the DMKG via two parallel paths: (1) relation linking prediction, and (2) RAG reasoning with large vision-language models. A decision module aggregates evidence from both paths to select the most credible answer. Experimental results on MMQAKE show that Hybrid-DMKG significantly outperforms existing MKE approaches, achieving higher accuracy and improved robustness to knowledge updates.
[531] The Station: An Open-World Environment for AI-Driven Discovery
Stephen Chung, Wenyu Du
Main category: cs.AI
TL;DR: STATION is an open-world multi-agent environment for autonomous scientific discovery where AI agents engage in complete scientific workflows without centralized coordination, achieving SOTA results and generating novel methods through emergent collaboration.
Details
Motivation: To create a realistic scientific ecosystem where AI agents can autonomously engage in complete scientific journeys, moving beyond rigid pipelines and enabling emergent scientific discovery through decentralized agent interactions.Method: STATION simulates a complete scientific ecosystem where agents with long context can read papers, formulate hypotheses, collaborate with peers, submit experiments, and publish results without centralized coordination, allowing free choice of actions and narrative development.
Result: AI agents achieve state-of-the-art performance across mathematics, computational biology, and machine learning benchmarks, surpassing AlphaEvolve in circle packing. Emergent narratives produce novel methods like a density-adaptive algorithm for scRNA-seq batch integration.
Conclusion: STATION represents a new paradigm for autonomous scientific discovery driven by emergent behavior in open-world environments, moving beyond rigid pipelines and enabling organic method development through agent collaboration.
Abstract: We introduce the STATION, an open-world multi-agent environment for autonomous scientific discovery. The Station simulates a complete scientific ecosystem, where agents can engage in long scientific journeys that include reading papers from peers, formulating hypotheses, collaborating with peers, submitting experiments, and publishing results. Importantly, there is no centralized system coordinating their activities. Utilizing their long context, agents are free to choose their own actions and develop their own narratives within the Station. Experiments demonstrate that AI agents in the Station achieve new state-of-the-art performance on a wide range of benchmarks, spanning mathematics, computational biology, and machine learning, notably surpassing AlphaEvolve in circle packing. A rich tapestry of unscripted narratives emerges, such as agents collaborating and analyzing other works rather than pursuing myopic optimization. From these emergent narratives, novel methods arise organically, such as a new density-adaptive algorithm for scRNA-seq batch integration that borrows concepts from another domain. The Station marks a first step towards autonomous scientific discovery driven by emergent behavior in an open-world environment, representing a new paradigm that moves beyond rigid pipelines.
[532] Minimal neuron ablation triggers catastrophic collapse in the language core of Large Vision-Language Models
Cen Lu, Yung-Chen Tang, Andrea Cavallaro
Main category: cs.AI
TL;DR: LVLMs have critical structural vulnerabilities - removing just a few specific neurons (as few as 4) can trigger catastrophic collapse, with language model components being more vulnerable than vision components.
Details
Motivation: Large Vision-Language Models show impressive capabilities but their robustness is poorly understood. The paper investigates structural vulnerabilities to identify critical neurons whose removal causes catastrophic collapse.Method: Proposed CAN method to detect Consistently Activated Neurons and locate critical neurons through progressive masking. Experiments conducted on LLaVA-1.5-7b-hf and InstructBLIP-Vicuna-7b.
Result: Masking only a tiny portion of language model’s feed-forward networks (as few as 4 neurons) triggers catastrophic collapse. Critical neurons are predominantly in language model rather than vision components, with down-projection layer being particularly vulnerable. Observed consistent two-stage collapse pattern: initial expressive degradation followed by sudden complete collapse.
Conclusion: The findings reveal critical structural vulnerabilities in LVLMs and provide important insights for safety research, highlighting the disproportionate impact of removing specific neurons in language model components.
Abstract: Large Vision-Language Models (LVLMs) have shown impressive multimodal understanding capabilities, yet their robustness is poorly understood. In this paper, we investigate the structural vulnerabilities of LVLMs to identify any critical neurons whose removal triggers catastrophic collapse. In this context, we propose CAN, a method to detect Consistently Activated Neurons and to locate critical neurons by progressive masking. Experiments on LLaVA-1.5-7b-hf and InstructBLIP-Vicuna-7b reveal that masking only a tiny portion of the language model’s feed-forward networks (just as few as four neurons in extreme cases) suffices to trigger catastrophic collapse. Notably, critical neurons are predominantly localized in the language model rather than in the vision components, and the down-projection layer is a particularly vulnerable structure. We also observe a consistent two-stage collapse pattern: initial expressive degradation followed by sudden, complete collapse. Our findings provide important insights for safety research in LVLMs.
[533] VICoT-Agent: A Vision-Interleaved Chain-of-Thought Framework for Interpretable Multimodal Reasoning and Scalable Remote Sensing Analysis
Chujie Wang, Zhiyuan Luo, Ruiqi Liu, Can Ran, Shenghua Fan, Xi Chen, Chu He
Main category: cs.AI
TL;DR: VICoT is a multimodal agent framework for remote sensing that enables explicit multi-round reasoning by dynamically incorporating visual tools into chain-of-thought processes, with distillation to smaller models.
Details
Motivation: Remote sensing analysis is evolving from simple object recognition to complex intelligence reasoning, requiring better reasoning capabilities and flexible tool invocation in models.Method: Vision-Interleaved Chain-of-Thought Framework (VICoT) uses stack-based reasoning structure and modular MCP-compatible tool suite to enable LLMs to perform multi-round vision-language reasoning. Includes Reasoning Stack distillation to migrate agent behaviors to smaller models.
Result: VICoT significantly outperforms existing SOTA frameworks on multiple remote sensing benchmarks in reasoning transparency, execution efficiency, and generation quality.
Conclusion: VICoT provides an effective framework for complex remote sensing reasoning tasks with strong generalization and flexibility, while distillation enables deployment on lightweight models.
Abstract: The current remote sensing image analysis task is increasingly evolving from traditional object recognition to complex intelligence reasoning, which places higher requirements on the model’s reasoning ability and the flexibility of tool invocation. To this end, we propose a new multimodal agent framework, Vision-Interleaved Chain-of-Thought Framework (VICoT), which implements explicit multi-round reasoning by dynamically incorporating visual tools into the chain of thought. Through a stack-based reasoning structure and a modular MCP-compatible tool suite, VICoT enables LLMs to efficiently perform multi-round, interleaved vision-language reasoning tasks with strong generalization and flexibility.We also propose the Reasoning Stack distillation method to migrate complex Agent behaviors to small, lightweight models, which ensures the reasoning capability while significantly reducing complexity. Experiments on multiple remote sensing benchmarks demonstrate that VICoT significantly outperforms existing SOTA frameworks in reasoning transparency, execution efficiency, and generation quality.
[534] Integrating Causal Foundation Model in Prescriptive Maintenance Framework for Optimizing Production Line OEE
Felix Saretzky, Lucas Andersen, Thomas Engel, Fazel Ansari
Main category: cs.AI
TL;DR: A causal machine learning model for prescriptive maintenance that moves beyond prediction to identify root causes and recommend optimal interventions by simulating fixes using a causal foundation model.
Details
Motivation: Current predictive maintenance models rely on spurious correlations rather than true causal drivers of failures, leading to misdiagnoses and ineffective interventions. The key challenge is understanding WHY failures occur to enable effective interventions, not just predicting WHEN they might happen.Method: Uses a pre-trained causal foundation model as a “what-if” model to estimate effects of potential fixes. Measures causal effects of interventions on system-level KPIs (like OEE) to provide data-driven ranking of actions. Evaluated using semi-synthetic manufacturing data compared to baseline ML model.
Result: The model identifies root causes and quantifies their operational impact, providing a systematic method to understand failure causes and recommend optimal interventions. Sets technical basis for robust prescriptive maintenance framework.
Conclusion: The proposed causal ML approach enables engineers to test potential solutions in a causal environment, make more effective operational decisions, reduce costly downtimes, and move from diagnosis to active prescription in maintenance.
Abstract: The transition to prescriptive maintenance in manufacturing is critically constrained by a dependence on predictive models. These models tend to rely on spurious correlations rather than identifying the true causal drivers of failures, often leading to costly misdiagnoses and ineffective interventions. This fundamental limitation results in a key-challenge: while we can predict that a failure may occur, we lack a systematic method to understand why a failure occurs, thereby providing the basis for identifying the most effective intervention. This paper proposes a model based on causal machine learning to bridge this gap. Our objective is to move beyond diagnosis to active prescription by simulating and evaluating potential fixes toward optimizing KPIs such as Overall Equipment Effectiveness (OEE). For this purpose a pre-trained causal foundation model is used as a “what-if” model to estimate the effects of potential fixes. By measuring the causal effect of each intervention on system-level KPIs, it provides a data-driven ranking of actions to recommend at the production line. This process not only identifies root causes but also quantifies their operational impact. The model is evaluated using semi-synthetic manufacturing data and compared with a baseline machine learning model. This paper sets the technical basis for a robust prescriptive maintenance framework, allowing engineers to test potential solutions in a causal environment to make more effective operational decisions and reduce costly downtimes.
[535] IndiMathBench: Autoformalizing Mathematical Reasoning Problems with a Human Touch
Param Biyani, Shashank Kirtania, Yasharth Bajpai, Sumit Gulwani, Ashish Tiwari
Main category: cs.AI
TL;DR: IndiMathBench is a human-verified benchmark of 312 formal Lean 4 theorems from Indian Math Olympiads, created using AI-assisted pipeline for evaluating mathematical theorem proving and autoformalization.
Details
Motivation: To create a challenging benchmark for evaluating mathematical theorem proving and autoformalization capabilities, addressing the gap between syntactic validity and semantic correctness in formal mathematics.Method: AI-powered human-assisted pipeline using category-based retrieval, iterative compiler feedback, and multi-model ensembles to generate candidate formalizations, with expert validation via interactive dashboard and automated quality summaries.
Result: Evaluation shows autoformalization remains challenging with substantial gaps between syntactic validity and semantic correctness, and theorem proving success rates remain low even with iterative refinement.
Conclusion: IndiMathBench presents a challenging testbed for mathematical reasoning and is publicly available for research use.
Abstract: We introduce IndiMathBench, a human-verified benchmark designed to evaluate mathematical theorem proving, curated using an AI-powered human-assisted pipeline for formalizing natural language problems in Lean. IndiMathBench is composed of 312 formal Lean 4 theorems paired with their corresponding informal problem statements, sourced from Indian Mathematics Olympiads. Through category-based retrieval, iterative compiler feedback, and multi-model ensembles, our pipeline generates candidate formalizations that experts efficiently validate via an interactive dashboard with automated quality summaries. Evaluation across multiple frontier models demonstrates that autoformalization remains challenging, with substantial gaps between syntactic validity and semantic correctness, while theorem proving success rates remain low even with iterative refinement, demonstrating that \benchmark~presents a challenging testbed for mathematical reasoning. IndiMathBench is available at https://github.com/prmbiy/IndiMathBench.
[536] ChartAnchor: Chart Grounding with Structural-Semantic Fidelity
Xinhang Li, Jingbo Zhou, Pengfei Luo, Yixiong Xiao, Tong Xu
Main category: cs.AI
TL;DR: ChartAnchor is a comprehensive benchmark for evaluating multimodal LLMs’ chart grounding capabilities through chart-to-code generation and controlled chart-to-table reconstruction tasks across 8k+ diverse chart types.
Details
Motivation: Existing benchmarks fail to holistically assess chart grounding due to narrow chart diversity, isolated tasks, and incomplete evaluation frameworks. Chart grounding is crucial for evaluating MLLMs' capabilities in numerical reasoning, multimodal alignment, and structural reconstruction.Method: Proposes ChartAnchor benchmark with 8k+ chart-table-code triples spanning 30 chart types from diverse sources. Introduces two complementary tasks: chart-to-code generation (synthesizing executable code to replicate charts) and controlled chart-to-table reconstruction (extracting exact data with predefined headers). Uses multi-level evaluation framework integrating semantic validation, stylistic analysis, and perceptual metrics.
Result: Extensive experiments on MLLMs reveal critical limitations in numerical precision and code synthesis, showing that current models struggle with structured reasoning beyond surface-level perception.
Conclusion: ChartAnchor establishes a rigorous foundation for chart grounding by unifying symbolic and data-driven approaches, offering meaningful insights for advancing MLLMs in scientific, financial, and industrial domains where precise chart comprehension is essential.
Abstract: Recent advances in multimodal large language models (MLLMs) highlight the need for benchmarks that rigorously evaluate structured chart comprehension.Chart grounding refers to the bidirectional alignment between a chart’s visual appearance and the structured semantics. This task requires models to produce a symbolic specification that faithfully captures the chart’s visual and structural intent, while also recovering the underlying tabular data with precise values and relationships. Chart grounding directly reflects a model’s capabilities in numerical reasoning, multimodal alignment, and structural reconstruction, and has several important applications in real-world scenarios.Existing benchmarks, constrained by narrow chart diversity, isolated tasks, and incomplete evaluation frameworks, fail to holistically assess grounding. To address this, we propose ChartAnchor, a comprehensive benchmark of 8k+ chart-table-code triples spanning 30 chart types drawn from diverse real-world and augmented sources. ChartAnchor introduces two complementary tasks: chart-to-code generation (synthesizing executable code to replicate charts) and controlled chart-to-table reconstruction (extracting exact data with predefined headers), enabling cross-validation of visual and numerical fidelity. A multi-level evaluation framework integrates semantic validation, stylistic analysis, and perceptual metrics to assess both structural and content-level correctness. Extensive experiments on MLLMs reveal critical limitations in numerical precision and code synthesis, emphasizing the need for structured reasoning beyond surface-level perception. By unifying symbolic and data-driven grounding, ChartAnchor establishes a rigorous foundation for chart grounding, offering meaningful insights for advancing MLLMs in scientific, financial, and industrial domains.
[537] Evaluating Legal Reasoning Traces with Legal Issue Tree Rubrics
Jinu Lee, Kyoung-Woon On, Simeng Han, Arman Cohan, Julia Hockenmaier
Main category: cs.AI
TL;DR: LEGIT dataset enables evaluation of LLM reasoning traces in legal domain using hierarchical argument trees from court judgments as rubrics, showing LLM legal reasoning depends on issue coverage/correctness and RAG/RL have complementary benefits.
Details
Motivation: Evaluating LLM-generated reasoning traces in expert domains like law is crucial for credibility and explainability but challenging due to domain complexity. Current methods lack reliable evaluation frameworks for legal reasoning quality.Method: Created LEGIT dataset (24K instances) by converting court judgments into hierarchical trees of opposing arguments and court conclusions. These trees serve as rubrics to evaluate reasoning trace coverage and correctness. Verified reliability via human expert annotations and comparison with simpler rubrics.
Result: (1) LLMs’ legal reasoning ability is significantly affected by both legal issue coverage and correctness. (2) RAG and RL with rubrics provide complementary benefits: RAG improves overall reasoning capability, while RL improves correctness but reduces coverage.
Conclusion: LEGIT provides a reliable framework for evaluating LLM reasoning traces in legal domain. The dataset reveals important insights about LLM legal reasoning limitations and shows how different techniques (RAG vs RL) address different aspects of reasoning quality.
Abstract: Evaluating the quality of LLM-generated reasoning traces in expert domains (e.g., law) is essential for ensuring credibility and explainability, yet remains challenging due to the inherent complexity of such reasoning tasks. We introduce LEGIT (LEGal Issue Trees), a novel large-scale (24K instances) expert-level legal reasoning dataset with an emphasis on reasoning trace evaluation. We convert court judgments into hierarchical trees of opposing parties’ arguments and the court’s conclusions, which serve as rubrics for evaluating the issue coverage and correctness of the reasoning traces. We verify the reliability of these rubrics via human expert annotations and comparison with coarse, less informative rubrics. Using the LEGIT dataset, we show that (1) LLMs’ legal reasoning ability is seriously affected by both legal issue coverage and correctness, and that (2) retrieval-augmented generation (RAG) and RL with rubrics bring complementary benefits for legal reasoning abilities, where RAG improves overall reasoning capability, whereas RL improves correctness albeit with reduced coverage.
[538] Med-CRAFT: Automated Construction of Interpretable and Multi-Hop Video Workloads via Knowledge Graph Traversal
Shenxi Liu, Kan Li, Mingyang Zhao, Yuhang Tian, Shoujun Zhou, Bin Li
Main category: cs.AI
TL;DR: Med-CRAFT is a neuro-symbolic framework that automatically generates high-quality medical video reasoning benchmarks by converting videos into spatiotemporal knowledge graphs and using graph traversal to create logically sound queries.
Details
Motivation: There's a scarcity of high-quality, logically annotated medical video datasets for training MLLMs. Manual annotation is too expensive and non-scalable, while existing synthetic methods suffer from hallucinations and lack logical interpretability.Method: Med-CRAFT extracts structured visual primitives from raw videos and builds a dynamic Spatiotemporal Knowledge Graph. It then generates benchmark queries by traversing valid paths in this graph, ensuring Chain-of-Thought provenance for each synthesized item.
Result: The pipeline produced M3-Med-Auto, a large-scale medical video reasoning benchmark with fine-grained temporal selectivity and multi-hop logical complexity. Evaluations show it generates query workloads comparable to expert-curated datasets, with high correlation between graph topology and MLLM reasoning steps.
Conclusion: Med-CRAFT enables scalable, low-cost construction of robust evaluation protocols for medical MLLMs by encoding verifiable logic into visual-linguistic benchmarks through deterministic graph traversal.
Abstract: The scarcity of high-quality, logically annotated video datasets remains a primary bottleneck in advancing Multi-Modal Large Language Models (MLLMs) for the medical domain. Traditional manual annotation is prohibitively expensive and non-scalable, while existing synthetic methods often suffer from stochastic hallucinations and a lack of logical interpretability. To address these challenges, we introduce \textbf{\PipelineName}, a novel neuro-symbolic data engineering framework that formalizes benchmark synthesis as a deterministic graph traversal process. Unlike black-box generative approaches, Med-CRAFT extracts structured visual primitives (e.g., surgical instruments, anatomical boundaries) from raw video streams and instantiates them into a dynamic Spatiotemporal Knowledge Graph. By anchoring query generation to valid paths within this graph, we enforce a rigorous Chain-of-Thought (CoT) provenance for every synthesized benchmark item. We instantiate this pipeline to produce M3-Med-Auto, a large-scale medical video reasoning benchmark exhibiting fine-grained temporal selectivity and multi-hop logical complexity. Comprehensive evaluations demonstrate that our automated pipeline generates query workloads with complexity comparable to expert-curated datasets. Furthermore, a logic alignment analysis reveals a high correlation between the prescribed graph topology and the reasoning steps of state-of-the-art MLLMs, validating the system’s capability to encode verifiable logic into visual-linguistic benchmarks. This work paves the way for scalable, low-cost construction of robust evaluation protocols in critical domains.
[539] Shielded Controller Units for RL with Operational Constraints Applied to Remote Microgrids
Hadi Nekoei, Alexandre Blondin Massé, Rachid Hassani, Sarath Chandar, Vincent Mai
Main category: cs.AI
TL;DR: Shielded Controller Units (SCUs) provide interpretable safety guarantees for RL agents in energy systems by decomposing constraints into hierarchical shields based on system dynamics.
Details
Motivation: Energy transition systems like remote microgrids require RL agents to respect complex operational constraints and regulations while optimizing fuel consumption and battery life under uncertain conditions.Method: SCUs leverage prior knowledge of system dynamics to create hierarchical shields that explicitly manage subsets of constraints, ensuring interpretable safety guarantees for RL deployment.
Result: RL agent with SCUs achieved 24% reduction in fuel consumption without increasing battery degradation while satisfying all operational constraints in remote microgrid optimization.
Conclusion: SCUs provide a systematic, interpretable approach for safe RL deployment in energy transition applications, enabling constraint satisfaction while maintaining optimization performance.
Abstract: Reinforcement learning (RL) is a powerful framework for optimizing decision-making in complex systems under uncertainty, an essential challenge in real-world settings, particularly in the context of the energy transition. A representative example is remote microgrids that supply power to communities disconnected from the main grid. Enabling the energy transition in such systems requires coordinated control of renewable sources like wind turbines, alongside fuel generators and batteries, to meet demand while minimizing fuel consumption and battery degradation under exogenous and intermittent load and wind conditions. These systems must often conform to extensive regulations and complex operational constraints. To ensure that RL agents respect these constraints, it is crucial to provide interpretable guarantees. In this paper, we introduce Shielded Controller Units (SCUs), a systematic and interpretable approach that leverages prior knowledge of system dynamics to ensure constraint satisfaction. Our shield synthesis methodology, designed for real-world deployment, decomposes the environment into a hierarchical structure where each SCU explicitly manages a subset of constraints. We demonstrate the effectiveness of SCUs on a remote microgrid optimization task with strict operational requirements. The RL agent, equipped with SCUs, achieves a 24% reduction in fuel consumption without increasing battery degradation, outperforming other baselines while satisfying all constraints. We hope SCUs contribute to the safe application of RL to the many decision-making challenges linked to the energy transition.
[540] Automating the Refinement of Reinforcement Learning Specifications
Tanmay Ambadkar, Đorđe Žikelić, Abhinav Verma
Main category: cs.AI
TL;DR: AutoSpec framework automatically refines coarse-grained logical specifications to provide better guidance for reinforcement learning, improving task solvability.
Details
Motivation: Coarse-grained logical specifications often fail to provide sufficient guidance for reinforcement learning agents, especially when tasks are under-specified, leading to poor policy learning.Method: AutoSpec searches for logical specification refinements using four procedures that modify the abstract graph of SpectRL specifications by refining existing edges or introducing new ones, maintaining soundness.
Result: AutoSpec improves the complexity of control tasks that can be solved when integrated with existing RL algorithms, with experiments showing promising improvements.
Conclusion: Automatic specification refinement through AutoSpec enhances RL performance by providing better guidance from logical specifications while maintaining soundness guarantees.
Abstract: Logical specifications have been shown to help reinforcement learning algorithms in achieving complex tasks. However, when a task is under-specified, agents might fail to learn useful policies. In this work, we explore the possibility of improving coarse-grained logical specifications via an exploration-guided strategy. We propose \textsc{AutoSpec}, a framework that searches for a logical specification refinement whose satisfaction implies satisfaction of the original specification, but which provides additional guidance therefore making it easier for reinforcement learning algorithms to learn useful policies. \textsc{AutoSpec} is applicable to reinforcement learning tasks specified via the SpectRL specification logic. We exploit the compositional nature of specifications written in SpectRL, and design four refinement procedures that modify the abstract graph of the specification by either refining its existing edge specifications or by introducing new edge specifications. We prove that all four procedures maintain specification soundness, i.e. any trajectory satisfying the refined specification also satisfies the original. We then show how \textsc{AutoSpec} can be integrated with existing reinforcement learning algorithms for learning policies from logical specifications. Our experiments demonstrate that \textsc{AutoSpec} yields promising improvements in terms of the complexity of control tasks that can be solved, when refined logical specifications produced by \textsc{AutoSpec} are utilized.
[541] SimWorld: An Open-ended Realistic Simulator for Autonomous Agents in Physical and Social Worlds
Jiawei Ren, Yan Zhuang, Xiaokang Ye, Lingjun Mao, Xuhong He, Jianzhi Shen, Mrinaal Dogra, Yiming Liang, Ruixuan Zhang, Tianai Yue, Yiqing Yang, Eric Liu, Ryan Wu, Kevin Benavente, Rajiv Mandya Nagaraju, Muhammad Faayez, Xiyan Zhang, Dhruv Vivek Sharma, Xianrui Zhong, Ziqiao Ma, Tianmin Shu, Zhiting Hu, Lianhui Qin
Main category: cs.AI
TL;DR: SimWorld is a new Unreal Engine 5-based simulator for developing and evaluating LLM/VLM agents in realistic physical and social environments, addressing limitations of existing simulators.
Details
Motivation: Current AI agents excel in constrained domains (math, coding) but struggle in complex real-world environments. Existing simulators lack realistic physics, social dynamics, and native LLM/VLM support needed for developing agents that can operate in real-world scenarios like earning income or running businesses.Method: Built on Unreal Engine 5, SimWorld provides: 1) realistic open-ended world simulation with accurate physics/social dynamics and language-driven procedural generation, 2) rich multimodal interfaces for LLM/VLM agents with open-vocabulary actions, and 3) extensible physical/social reasoning scenarios.
Result: The authors deployed frontier LLM agents (GPT-4o, Gemini-2.5-Flash, Claude-3.5, DeepSeek-Prover-V2) on long-horizon multi-agent delivery tasks with strategic cooperation/competition, revealing distinct reasoning patterns and limitations across models.
Conclusion: SimWorld is open-sourced as a foundational platform for advancing real-world agent intelligence across disciplines, addressing critical gaps in agent development infrastructure for complex physical and social environments.
Abstract: While LLM/VLM-powered AI agents have advanced rapidly in math, coding, and computer use, their applications in complex physical and social environments remain challenging. Building agents that can survive and thrive in the real world (for example, by autonomously earning income or running a business) requires massive-scale interaction, reasoning, training, and evaluation across diverse embodied scenarios. However, existing world simulators for such development fall short: they often rely on limited hand-crafted environments, simulate simplified game-like physics and social rules, and lack native support for LLM/VLM agents. We introduce SimWorld, a new simulator built on Unreal Engine 5, designed for developing and evaluating LLM/VLM agents in rich, real-world-like settings. SimWorld offers three core capabilities: (1) realistic, open-ended world simulation, including accurate physical and social dynamics and language-driven procedural environment generation; (2) a rich interface for LLM/VLM agents, with multimodal world inputs and open-vocabulary actions at varying levels of abstraction; and (3) diverse and extensible physical and social reasoning scenarios that are easily customizable by users. We demonstrate SimWorld by deploying frontier LLM agents (e.g., GPT-4o, Gemini-2.5-Flash, Claude-3.5, and DeepSeek-Prover-V2) on long-horizon multi-agent delivery tasks involving strategic cooperation and competition. The results reveal distinct reasoning patterns and limitations across models. We open-source SimWorld and hope it becomes a foundational platform for advancing real-world agent intelligence across disciplines: https://simworld.org.
[542] CodeDistiller: Automatically Generating Code Libraries for Scientific Coding Agents
Peter Jansen, Samiah Hassan, Pragnya Narasimha
Main category: cs.AI
TL;DR: CodeDistiller automatically distills working domain-specific code examples from GitHub repositories to enhance Automated Scientific Discovery systems, achieving 74% functional example generation and improving experiment quality.
Details
Motivation: Current Automated Scientific Discovery (ASD) systems are limited by either mutating few manually-crafted examples or relying solely on parametric knowledge, restricting their quality and reach. There's a need to automatically expand ASD capabilities without manual effort.Method: CodeDistiller automatically distills large collections of scientific GitHub repositories into a vetted library of working domain-specific code examples. The system uses automatic and domain-expert evaluation on 250 materials science repositories to identify functional examples.
Result: The best model produces functional examples for 74% of repositories. Downstream evaluation shows ASD agents augmented with CodeDistiller-generated libraries produce more accurate, complete, and scientifically sound experiments compared to agents with only general materials-science examples.
Conclusion: CodeDistiller successfully enables ASD systems to automatically expand their capabilities by distilling working code examples from scientific repositories, overcoming limitations of current approaches and improving experiment quality.
Abstract: Automated Scientific Discovery (ASD) systems can help automatically generate and run code-based experiments, but their capabilities are limited by the code they can reliably generate from parametric knowledge alone. As a result, current systems either mutate a small number of manually-crafted experiment examples, or operate solely from parametric knowledge, limiting quality and reach. We introduce CodeDistiller, a system that automatically distills large collections of scientific Github repositories into a vetted library of working domain-specific code examples, allowing ASD agents to expand their capabilities without manual effort. Using a combination of automatic and domain-expert evaluation on 250 materials science repositories, we find the best model is capable of producing functional examples for 74% of repositories, while our downstream evaluation shows an ASD agent augmented with a CodeDistiller generated library produces more accurate, complete, and scientifically sound experiments than an agent with only general materials-science code examples.
[543] Energy-Aware Data-Driven Model Selection in LLM-Orchestrated AI Systems
Daria Smirnova, Hamid Nasiri, Marta Adamska, Zhengxin Yu, Peter Garraghan
Main category: cs.AI
TL;DR: GUIDE is an energy-aware model selection framework that improves LLM-based orchestration by incorporating quantitative model performance characteristics, achieving better accuracy and energy efficiency.
Details
Motivation: Current LLM-based orchestrators rely on qualitative model descriptions that don't reflect true capabilities, leading to suboptimal model selection, reduced accuracy, and increased energy costs.Method: Proposed GUIDE framework that incorporates quantitative model performance characteristics and accounts for performance-energy trade-offs in decision-making.
Result: GUIDE increases accuracy by 0.90%-11.92% across tasks, achieves up to 54% energy efficiency improvement, and reduces orchestrator model selection latency from 4.51 s to 7.2 ms.
Conclusion: Quantitative model performance characteristics should be incorporated into AI orchestration frameworks to improve accuracy, energy efficiency, and reduce latency in model selection.
Abstract: As modern artificial intelligence (AI) systems become more advanced and capable, they can leverage a wide range of tools and models to perform complex tasks. Today, the task of orchestrating these models is often performed by Large Language Models (LLMs) that rely on qualitative descriptions of models for decision-making. However, the descriptions provided to these LLM-based orchestrators do not reflect true model capabilities and performance characteristics, leading to suboptimal model selection, reduced accuracy, and increased energy costs. In this paper, we conduct an empirical analysis of LLM-based orchestration limitations and propose GUIDE, a new energy-aware model selection framework that accounts for performance-energy trade-offs by incorporating quantitative model performance characteristics in decision-making. Experimental results demonstrate that GUIDE increases accuracy by 0.90%-11.92% across various evaluated tasks, and achieves up to 54% energy efficiency improvement, while reducing orchestrator model selection latency from 4.51 s to 7.2 ms.
[544] Foundation Priors
Sanjog Misra
Main category: cs.AI
TL;DR: Foundation models generate synthetic data that should be treated as draws from a “foundation prior” rather than real observations, reflecting both model patterns and user biases.
Details
Motivation: Growing interest in using LLM-generated synthetic outputs as data in empirical research creates risk of conflating synthetic "facts" with real observations. Need principled framework to incorporate generative AI outputs while acknowledging their subjective nature.Method: Introduce “foundation prior” concept showing synthetic data are draws from prior predictive distribution. Model subjectivity through dependence on user’s anticipated data distribution, prompt engineering, and trust in model. Derive foundation prior as exponential-tilted generalized Bayesian update of user’s primitive prior with trust parameter.
Result: Framework enables incorporation of synthetic data into statistical/econometric workflows as structured, explicitly subjective priors rather than empirical observations. Applications include refining complex models, informing latent constructs, guiding experimental design, and augmenting random-coefficient/partially linear specifications.
Conclusion: Treating generative outputs as foundation priors offers principled way to harness foundation models in empirical work while avoiding conflation of synthetic data with real observations. Framework acknowledges subjective nature of synthetic data generation process.
Abstract: Foundation models, and in particular large language models, can generate highly informative responses, prompting growing interest in using these ‘‘synthetic’’ outputs as data in empirical research and decision-making. This paper introduces the idea of a foundation prior, which shows that model-generated outputs are not as real observations, but draws from the foundation prior induced prior predictive distribution. As such synthetic data reflects both the model’s learned patterns and the user’s subjective priors, expectations, and biases. We model the subjectivity of the generative process by making explicit the dependence of synthetic outputs on the user’s anticipated data distribution, the prompt-engineering process, and the trust placed in the foundation model. We derive the foundation prior as an exponential-tilted, generalized Bayesian update of the user’s primitive prior, where a trust parameter governs the weight assigned to synthetic data. We then show how synthetic data and the associated foundation prior can be incorporated into standard statistical and econometric workflows, and discuss their use in applications such as refining complex models, informing latent constructs, guiding experimental design, and augmenting random-coefficient and partially linear specifications. By treating generative outputs as structured, explicitly subjective priors rather than as empirical observations, the framework offers a principled way to harness foundation models in empirical work while avoiding the conflation of synthetic ‘‘facts’’ with real data.
[545] A Benchmark of Causal vs Correlation AI for Predictive Maintenance
Krishna Taduri, Shaunak Dhande, Giacinto Paolo, Saggese, Paul Smith
Main category: cs.AI
TL;DR: Causal AI methods outperform correlation-based ML in predictive maintenance by reducing false alarms 97% while maintaining high recall, achieving $1.16M annual cost savings.
Details
Motivation: Predictive maintenance has extreme cost asymmetry (missed failures cost 50× more than false alarms), but conventional ML optimizes statistical accuracy rather than operational costs and can't distinguish causal relationships from spurious correlations.Method: Evaluated 8 predictive models from baseline statistical approaches to formal causal inference methods on 10,000 CNC machines dataset with 3.3% failure prevalence. Compared causal inference model (L5) against correlation-based decision tree model (L3).
Result: Causal model achieved $1.16M annual savings (70.2% reduction), outperforming best correlation model by $80K/year. Matched highest recall (87.9%) while reducing false alarms 97% (165 to 5), precision 92.1%, with only 2.6% train-test gap.
Conclusion: Causal AI methods with domain knowledge yield superior financial outcomes and more interpretable predictions than correlation-based approaches in predictive maintenance, addressing operational cost asymmetry effectively.
Abstract: Predictive maintenance in manufacturing environments presents a challenging optimization problem characterized by extreme cost asymmetry, where missed failures incur costs roughly fifty times higher than false alarms. Conventional machine learning approaches typically optimize statistical accuracy metrics that do not reflect this operational reality and cannot reliably distinguish causal relationships from spurious correlations. This study evaluates eight predictive models, ranging from baseline statistical approaches to formal causal inference methods, on a dataset of 10,000 CNC machines with a 3.3% failure prevalence. The formal causal inference model (L5) achieved estimated annual cost savings of 1.16 million USD (a 70.2 percent reduction), outperforming the best correlation-based decision tree model (L3) by approximately 80,000 USD per year. The causal model matched the highest observed recall (87.9 percent) while reducing false alarms by 97 percent (from 165 to 5) and attained a precision of 92.1 percent, with a train-test performance gap of only 2.6 percentage points. These results indicate that causal AI methods, when combined with domain knowledge, can yield superior financial outcomes and more interpretable predictions compared to correlation-based approaches in predictive maintenance applications.
[546] fMRI2GES: Co-speech Gesture Reconstruction from fMRI Signal with Dual Brain Decoding Alignment
Chunzheng Zhu, Jialin Shao, Jianxin Lin, Yijun Wang, Jing Wang, Jinhui Tang, Kenli Li
Main category: cs.AI
TL;DR: fMRI2GES: A novel method for reconstructing co-speech gestures directly from fMRI recordings using dual brain decoding alignment with unpaired data.
Details
Motivation: Previous brain decoding studies focused on image and language reconstruction, but reconstructing gestures associated with speech stimuli remains challenging due to lack of paired {brain, speech, gesture} data.Method: Uses Dual Brain Decoding Alignment with fMRI-to-text and text-to-gesture models to establish dual fMRI-to-gesture reconstruction patterns, then aligns outputs and trains in self-supervised manner on unpaired data.
Result: The method successfully reconstructs expressive gestures directly from fMRI recordings and investigates how fMRI signals from different cortical ROIs affect generation results.
Conclusion: Provides new insights into decoding co-speech gestures, advancing understanding of neuroscience and cognitive science by overcoming data pairing limitations.
Abstract: Understanding how the brain responds to external stimuli and decoding this process has been a significant challenge in neuroscience. While previous studies typically concentrated on brain-to-image and brain-to-language reconstruction, our work strives to reconstruct gestures associated with speech stimuli perceived by brain. Unfortunately, the lack of paired {brain, speech, gesture} data hinders the deployment of deep learning models for this purpose. In this paper, we introduce a novel approach, \textbf{fMRI2GES}, that allows training of fMRI-to-gesture reconstruction networks on unpaired data using \textbf{Dual Brain Decoding Alignment}. This method relies on two key components: (i) observed texts that elicit brain responses, and (ii) textual descriptions associated with the gestures. Then, instead of training models in a completely supervised manner to find a mapping relationship among the three modalities, we harness an fMRI-to-text model, a text-to-gesture model with paired data and an fMRI-to-gesture model with unpaired data, establishing dual fMRI-to-gesture reconstruction patterns. Afterward, we explicitly align two outputs and train our model in a self-supervision way. We show that our proposed method can reconstruct expressive gestures directly from fMRI recordings. We also investigate fMRI signals from different ROIs in the cortex and how they affect generation results. Overall, we provide new insights into decoding co-speech gestures, thereby advancing our understanding of neuroscience and cognitive science.
[547] Knowledge Graph Augmented Large Language Models for Next-Visit Disease Prediction
Ruiyu Wang, Tuan Vinh, Ran Xu, Yuyin Zhou, Jiaying Lu, Carl Yang, Francisco Pasquel
Main category: cs.AI
TL;DR: KG-guided chain-of-thought framework generates clinically grounded explanations for EHR disease prediction using knowledge graphs and LLMs.
Details
Motivation: Existing EHR prediction models provide coarse, post hoc explanations with limited value for patient-level decision making, lacking clinically grounded and temporally consistent reasoning.Method: Map ICD-9 codes to PrimeKG knowledge graph, extract disease-relevant nodes and multi-hop reasoning paths as scaffolds for chain-of-thought generation, retain only explanations matching observed outcomes, then fine-tune lightweight LLMs (LLaMA-3.1-Instruct-8B and Gemma-7B) on this supervision corpus.
Result: KG-guided models outperform classical baselines across ten diseases with AUROC 0.66-0.70 and macro-AUPR 0.40-0.47 on limited training data (400-1000 cases). Zero-shot transfer to CRADLE cohort improves accuracy from ~0.40-0.51 to 0.72-0.77. Clinician evaluation shows preference for KG-guided explanations in clarity, relevance, and clinical correctness.
Conclusion: The KG-guided chain-of-thought framework provides clinically grounded, temporally consistent explanations for EHR disease prediction, improving both predictive performance and explanation quality while enabling effective transfer learning across cohorts.
Abstract: Electronic health records (EHRs) support powerful clinical prediction models, but existing methods typically provide coarse, post hoc explanations that offer limited value for patient-level decision making. We introduce a knowledge graph (KG)-guided chain-of-thought (CoT) framework that generates clinically grounded and temporally consistent reasoning for visit-level disease prediction in MIMIC-III. ICD-9 codes are mapped to PrimeKG, from which disease-relevant nodes and multi-hop reasoning paths are extracted and used as scaffolds for CoT generation; only explanations whose conclusions match observed outcomes are retained. Lightweight LLaMA-3.1-Instruct-8B and Gemma-7B models are then fine-tuned on this supervision corpus. Across ten PrimeKG-mapped diseases and limited training cohorts (400 and 1000 cases), KG-guided models outperform strong classical baselines, achieving AUROC values of 0.66 to 0.70 and macro-AUPR values of 0.40 to 0.47. The models also transfer zero-shot to the CRADLE cohort, improving accuracy from approximately 0.40 to 0.51 up to 0.72 to 0.77. A blinded clinician evaluation shows consistent preference for KG-guided CoT explanations in clarity, relevance, and clinical correctness.
[548] Unsupervised decoding of encoded reasoning using language model interpretability
Ching Fang, Samuel Marks
Main category: cs.AI
TL;DR: Logit lens can decode ROT-13 encoded chain-of-thought reasoning in LLMs, with accuracy peaking in middle-to-late layers, suggesting current interpretability methods are more robust to simple encoded reasoning than previously thought.
Details
Motivation: To investigate whether current interpretability techniques can penetrate encoded reasoning processes in LLMs, addressing concerns about hidden reasoning that evades human oversight as models become more capable.Method: Fine-tuned DeepSeek-R1-Distill-Llama-70B to perform chain-of-thought reasoning in ROT-13 encryption while maintaining intelligible English outputs, then evaluated logit lens analysis on decoding hidden reasoning from internal activations.
Result: Logit lens effectively translates encoded reasoning with accuracy peaking in intermediate-to-late layers. Developed unsupervised decoding pipeline combining logit lens with automated paraphrasing achieves substantial accuracy in reconstructing complete reasoning transcripts.
Conclusion: Current mechanistic interpretability techniques may be more robust to simple forms of encoded reasoning than previously understood, providing a framework for evaluating interpretability against non-human-readable reasoning formats to maintain AI oversight.
Abstract: As large language models become increasingly capable, there is growing concern that they may develop reasoning processes that are encoded or hidden from human oversight. To investigate whether current interpretability techniques can penetrate such encoded reasoning, we construct a controlled testbed by fine-tuning a reasoning model (DeepSeek-R1-Distill-Llama-70B) to perform chain-of-thought reasoning in ROT-13 encryption while maintaining intelligible English outputs. We evaluate mechanistic interpretability methods–in particular, logit lens analysis–on their ability to decode the model’s hidden reasoning process using only internal activations. We show that logit lens can effectively translate encoded reasoning, with accuracy peaking in intermediate-to-late layers. Finally, we develop a fully unsupervised decoding pipeline that combines logit lens with automated paraphrasing, achieving substantial accuracy in reconstructing complete reasoning transcripts from internal model representations. These findings suggest that current mechanistic interpretability techniques may be more robust to simple forms of encoded reasoning than previously understood. Our work provides an initial framework for evaluating interpretability methods against models that reason in non-human-readable formats, contributing to the broader challenge of maintaining oversight over increasingly capable AI systems.
[549] OntoMetric: An Ontology-Guided Framework for Automated ESG Knowledge Graph Construction
Mingqin Yu, Fethi Rabhi, Boming Xia, Zhengyi Yang, Felix Tan, Qinghua Lu
Main category: cs.AI
TL;DR: OntoMetric: An ontology-guided framework that transforms ESG regulatory PDFs into validated knowledge graphs using structure-aware segmentation, ontology-constrained LLM extraction, and two-phase validation.
Details
Motivation: ESG disclosure frameworks (SASB, TCFD, IFRS S2) require organizations to compute and report metrics, but these requirements are embedded in long, unstructured PDF documents that are difficult to interpret, standardize, and audit. Manual extraction is unscalable, while unconstrained LLM extraction produces inconsistent entities, hallucinated relationships, missing provenance, and high validation failure rates.Method: Three-stage pipeline: (1) structure-aware segmentation using table-of-contents boundaries, (2) ontology-constrained LLM extraction that embeds the ESGMKG schema into prompts while enriching entities with semantic fields, (3) two-phase validation combining LLM-based semantic verification with rule-based schema checking across entity, property, and relationship levels (VR001-VR006). The framework preserves both segment-level and page-level provenance for audit traceability.
Result: Evaluated on five ESG standards (SASB Commercial Banks, SASB Semiconductors, TCFD, IFRS S2, AASB S2) totaling 228 pages and 60 segments, OntoMetric achieves 65-90% semantic accuracy and 80-90% schema compliance, compared to 3-10% for baseline unconstrained extraction, at approximately 0.01 to 0.02 USD per validated entity.
Conclusion: Combining symbolic ontology constraints with neural extraction enables reliable, auditable knowledge graphs suitable for regulatory compliance and web integration, supporting downstream applications such as sustainable-finance analytics, transparency portals, and automated compliance tools.
Abstract: Environmental, Social, and Governance (ESG) disclosure frameworks such as SASB, TCFD, and IFRS S2 require organizations to compute and report numerous metrics for compliance, yet these requirements are embedded in long, unstructured PDF documents that are difficult to interpret, standardize, and audit. Manual extraction is unscalable, while unconstrained large language model (LLM) extraction often produces inconsistent entities, hallucinated relationships, missing provenance, and high validation failure rates. We present OntoMetric, an ontology-guided framework that transforms ESG regulatory documents into validated, AI- and web-ready knowledge graphs. OntoMetric operates through a three-stage pipeline: (1) structure-aware segmentation using table-of-contents boundaries, (2) ontology-constrained LLM extraction that embeds the ESGMKG schema into prompts while enriching entities with semantic fields for downstream reasoning, and (3) two-phase validation that combines LLM-based semantic verification with rule-based schema checking across entity, property, and relationship levels (VR001-VR006). The framework preserves both segment-level and page-level provenance for audit traceability. Evaluated on five ESG standards (SASB Commercial Banks, SASB Semiconductors, TCFD, IFRS S2, AASB S2) totaling 228 pages and 60 segments, OntoMetric achieves 65-90% semantic accuracy and 80-90% schema compliance, compared to 3-10% for baseline unconstrained extraction, at approximately 0.01 to 0.02 USD per validated entity. Our results demonstrate that combining symbolic ontology constraints with neural extraction enables reliable, auditable knowledge graphs suitable for regulatory compliance and web integration, supporting downstream applications such as sustainable-finance analytics, transparency portals, and automated compliance tools.
[550] RoboDriveVLM: A Novel Benchmark and Baseline towards Robust Vision-Language Models for Autonomous Driving
Dacheng Liao, Mengshi Qi, Peng Shu, Zhining Zhang, Yuxin Lin, Liang Liu, Huadong Ma
Main category: cs.AI
TL;DR: RoboDriveBench is the first robustness benchmark for VLM-based autonomous driving, evaluating 11 corruption scenarios across 64,559 trajectory prediction cases. RoboDriveVLM framework improves robustness by mapping multimodal data into unified latent space with test-time adaptation.
Details
Motivation: Current VLM-based autonomous driving systems pose real-world risks due to lack of robustness evaluation. There's a need to systematically assess VLM viability for autonomous driving under various corruption scenarios before real-world deployment.Method: 1) Introduces RoboDriveBench benchmark with 11 simulated corruption scenarios (6 sensor, 5 prompt) across 250 unique driving scenarios each. 2) Proposes RoboDriveVLM framework mapping multimodal data (lidar, radar) into unified latent space. 3) Implements Test-Time Adaptation method using cross-modal knowledge distillation.
Result: Benchmark reveals limitations of current VLM-based systems. RoboDriveVLM with TTA demonstrates improved robustness against real-world challenges. Extensive experiments validate the approach across 64,559 trajectory prediction cases.
Conclusion: The work highlights critical robustness issues in current VLM-based autonomous driving and provides a more reliable solution through systematic benchmarking, multimodal data integration, and adaptive techniques for real-world deployment.
Abstract: Current Vision-Language Model (VLM)-based end-to-end autonomous driving systems often leverage large language models to generate driving decisions directly based on their understanding of the current scene. However, such systems introduce multiple risks in real-world driving scenarios. To evaluate whether VLMs are truly viable for autonomous driving, we introduce RoboDriveBench, the first robustness benchmark focused on end-to-end trajectory prediction tasks. This benchmark systematically evaluates two critical categories of real-world challenges for VLM-based end-to-end autonomous driving systems through 11 simulated scenarios encompassing various corruption types, including 6 scenarios of sensor corruption caused by environmental variations, along with 5 cases of prompt corruption resulting from human intervention and data transmission failures. Each corruption type includes 250 unique driving scenarios and 5,689 frames, resulting in 64,559 total trajectory prediction cases per evaluation. To overcome these real-world challenges, we propose a novel VLM-based autonomous driving framework called RoboDriveVLM, which enhances robustness by mapping more multimodal data-e.g., lidar and radar-into a unified latent space. Furthermore, we introduce a new Test-Time Adaptation (TTA) method based on cross-modal knowledge distillation to improve the robustness of VLM-based autonomous driving systems. Through extensive experiments, our work highlights the limitations of current VLM-based end-to-end autonomous driving systems and provides a more reliable solution for real-world deployment. Source code and datasets will be released.
[551] CuES: A Curiosity-driven and Environment-grounded Synthesis Framework for Agentic RL
Shinji Mai, Yunpeng Zhai, Ziqian Chen, Cheng Chen, Anni Zou, Shuchang Tao, Zhaoyang Liu, Bolin Ding
Main category: cs.AI
TL;DR: CuES is a curiosity-driven framework that autonomously generates diverse, executable tasks from environment structure to address task scarcity in agentic RL, enabling agents to learn without predefined tasks.
Details
Motivation: Agentic RL requires structured training tasks, but many realistic environments lack predefined tasks (task scarcity). Existing approaches assume predefined task collections, which fails in novel environments where tool semantics and affordances are initially unknown.Method: CuES (Curiosity driven and Environment grounded Synthesis) autonomously generates tasks directly from environment structure and affordances without handcrafted seeds or external corpora. It uses intrinsic curiosity for exploration, abstracts interaction patterns into reusable task schemas, and refines them through lightweight top-down guidance and memory-based quality control.
Result: Across three environments (AppWorld, BFCL, and WebShop), CuES produces task distributions that match or surpass manually curated datasets in both diversity and executability, yielding substantial downstream policy improvements.
Conclusion: Curiosity-driven, environment-grounded task generation provides a scalable foundation for agents that not only learn how to act, but also learn what to learn, addressing the task scarcity bottleneck in agentic RL.
Abstract: Large language model based agents are increasingly deployed in complex, tool augmented environments. While reinforcement learning provides a principled mechanism for such agents to improve through interaction, its effectiveness critically depends on the availability of structured training tasks. In many realistic settings, however, no such tasks exist a challenge we term task scarcity, which has become a key bottleneck for scaling agentic RL. Existing approaches typically assume predefined task collections, an assumption that fails in novel environments where tool semantics and affordances are initially unknown. To address this limitation, we formalize the problem of Task Generation for Agentic RL, where an agent must learn within a given environment that lacks predefined tasks. We propose CuES, a Curiosity driven and Environment grounded Synthesis framework that autonomously generates diverse, executable, and meaningful tasks directly from the environment structure and affordances, without relying on handcrafted seeds or external corpora. CuES drives exploration through intrinsic curiosity, abstracts interaction patterns into reusable task schemas, and refines them through lightweight top down guidance and memory based quality control. Across three representative environments, AppWorld, BFCL, and WebShop, CuES produces task distributions that match or surpass manually curated datasets in both diversity and executability, yielding substantial downstream policy improvements. These results demonstrate that curiosity driven, environment grounded task generation provides a scalable foundation for agents that not only learn how to act, but also learn what to learn. The code is available at https://github.com/modelscope/AgentEvolver/research/CuES.
[552] Extending NGU to Multi-Agent RL: A Preliminary Study
Juan Hernandez, Diego Fernández, Manuel Cifuentes, Denis Parra, Rodrigo Toro Icarte
Main category: cs.AI
TL;DR: NGU algorithm extended to multi-agent environments shows improved performance over baseline DQN when using shared replay buffer and carefully tuned intrinsic exploration signals.
Details
Motivation: To extend the effective Never Give Up (NGU) algorithm, which combines episodic novelty and intrinsic motivation for sparse reward tasks, to multi-agent environments and evaluate its performance.Method: Extended NGU to multi-agent settings in simple_tag environment from PettingZoo suite. Investigated three design choices: (1) shared vs individual replay buffers, (2) sharing episodic novelty among agents with different k thresholds, (3) heterogeneous vs common beta parameter values.
Result: NGU achieved moderately higher returns and more stable learning dynamics compared to multi-agent DQN baseline. Shared replay buffer yielded best performance and stability. Novelty sharing performed comparably when k=1 but degraded learning for larger values. Heterogeneous beta values did not improve over small common value.
Conclusion: NGU can be effectively applied in multi-agent settings when experiences are shared (via shared replay buffer) and intrinsic exploration signals are carefully tuned, with shared experiences being key to the performance gains.
Abstract: The Never Give Up (NGU) algorithm has proven effective in reinforcement learning tasks with sparse rewards by combining episodic novelty and intrinsic motivation. In this work, we extend NGU to multi-agent environments and evaluate its performance in the simple_tag environment from the PettingZoo suite. Compared to a multi-agent DQN baseline, NGU achieves moderately higher returns and more stable learning dynamics. We investigate three design choices: (1) shared replay buffer versus individual replay buffers, (2) sharing episodic novelty among agents using different k thresholds, and (3) using heterogeneous values of the beta parameter. Our results show that NGU with a shared replay buffer yields the best performance and stability, highlighting that the gains come from combining NGU intrinsic exploration with experience sharing. Novelty sharing performs comparably when k = 1 but degrades learning for larger values. Finally, heterogeneous beta values do not improve over a small common value. These findings suggest that NGU can be effectively applied in multi-agent settings when experiences are shared and intrinsic exploration signals are carefully tuned.
[553] A Fast Heuristic Search Approach for Energy-Optimal Profile Routing for Electric Vehicles
Saman Ahmadi, Mahdi Jalili
Main category: cs.AI
TL;DR: Energy-optimal path planning for EVs with uncertain initial energy using multi-objective A* search with profile dominance rules.
Details
Motivation: Real-world EV routing often involves uncertainty in available initial energy, requiring planning for all possible energy levels rather than assuming a known starting charge.Method: Proposes a label-setting approach using multi-objective A* search with novel profile dominance rules to avoid complex profile handling, developing four variants of this method.
Result: Experimental evaluation on real-world road networks with realistic energy consumption data shows performance comparable to energy-optimal A* with known initial energy.
Conclusion: The proposed energy profile A* search provides an effective solution for EV path planning under energy uncertainty without requiring complex profile management.
Abstract: We study the energy-optimal shortest path problem for electric vehicles (EVs) in large-scale road networks, where recuperated energy along downhill segments introduces negative energy costs. While traditional point-to-point pathfinding algorithms for EVs assume a known initial energy level, many real-world scenarios involving uncertainty in available energy require planning optimal paths for all possible initial energy levels, a task known as energy-optimal profile search. Existing solutions typically rely on specialized profile-merging procedures within a label-correcting framework that results in searching over complex profiles. In this paper, we propose a simple yet effective label-setting approach based on multi-objective A* search, which employs a novel profile dominance rule to avoid generating and handling complex profiles. We develop four variants of our method and evaluate them on real-world road networks enriched with realistic energy consumption data. Experimental results demonstrate that our energy profile A* search achieves performance comparable to energy-optimal A* with a known initial energy level.
[554] Benchmarking Overton Pluralism in LLMs
Elinor Poole-Dayan, Jiayi Wu, Taylor Sorensen, Jiaxin Pei, Michiel A. Bakker
Main category: cs.AI
TL;DR: The paper introduces OvertonScore, a framework for measuring viewpoint diversity in LLM outputs, shows current models score 0.35-0.41 (far below max 1.0), and develops an automated benchmark that correlates well with human judgments (ρ=0.88).
Details
Motivation: Current LLMs lack systematic measurement of viewpoint diversity (pluralism) in their outputs. Large-scale human studies are costly and slow, creating need for scalable evaluation tools to advance pluralistic alignment.Method: (1) Formalize Overton pluralism as a set coverage metric (OvertonScore), (2) conduct large-scale U.S.-representative human study (N=1209, 60 questions, 8 LLMs), (3) develop automated benchmark that reproduces human judgments.
Result: Models achieve OvertonScores of 0.35-0.41 on average (DeepSeek V3 best), far below theoretical maximum of 1.0. Automated benchmark achieves high rank correlation with human judgments (ρ=0.88), providing practical proxy for evaluation.
Conclusion: The work establishes foundation for systematic progress toward more pluralistic LLMs by turning pluralistic alignment from normative aim into measurable benchmark, with automated tools enabling scalable evaluation.
Abstract: We introduce a novel framework for measuring Overton pluralism in LLMs–the extent to which diverse viewpoints are represented in model outputs. We (i) formalize Overton pluralism as a set coverage metric (OvertonScore), (ii) conduct a large-scale U.S.-representative human study (N = 1209; 60 questions; 8 LLMs), and (iii) develop an automated benchmark that closely reproduces human judgments. On average, models achieve OvertonScores of 0.35–0.41, with DeepSeek V3 performing best; yet all models remain far below the theoretical maximum of 1.0, revealing substantial headroom for improvement. Because repeated large-scale human studies are costly and slow, scalable evaluation tools are essential for model development. Hence, we propose an automated benchmark that achieves high rank correlation with human judgments ($ρ=0.88$), providing a practical proxy without replacing human assessment. By turning pluralistic alignment from a normative aim into a measurable benchmark, our work establishes a foundation for systematic progress toward more pluralistic LLMs.
[555] The Necessity of Imperfection:Reversing Model Collapse via Simulating Cognitive Boundedness
Zhongjie Jiang
Main category: cs.AI
TL;DR: Paper proposes PMCSF framework that simulates cognitive processes instead of imitating surface data to create synthetic text with human-like imperfections, preventing model collapse.
Details
Motivation: Current synthetic data production optimizes for statistical smoothness, removing human-like irregularities and accelerating model collapse. Need a paradigm shift from imitating surface properties to simulating cognitive processes.Method: Prompt-driven Cognitive Computing Framework (PMCSF) with Cognitive State Decoder (CSD) that reverse-engineers text into cognitive vectors, and Cognitive Text Encoder (CTE) that re-materializes states into text with human imperfections via Cognitive Perturbation Operators.
Result: CTE text shows Jensen-Shannon divergence of 0.0614 from human text (vs. 0.4431 for standard LLM), passes double-blind professional review, ICC > 0.9 for cognitive alignment. In A-share market tests, reduces maximum drawdown by 47.4% during 2015 crash and delivers 8.6% Defensive Alpha.
Conclusion: Modelling human cognitive limitations rather than copying surface data enables synthetic data with genuine functional gain, offering a technical pathway to resolve AI data-collapse crisis.
Abstract: Although synthetic data is widely promoted as a remedy, its prevailing production paradigm – one optimizing for statistical smoothness – systematically removes the long-tail, cognitively grounded irregularities that characterize human text. Prolonged training on such statistically optimal but cognitively impoverished data accelerates model collapse. This paper proposes a paradigm shift: instead of imitating the surface properties of data, we simulate the cognitive processes that generate human text. We introduce the Prompt-driven Cognitive Computing Framework (PMCSF), whose core consists of a Cognitive State Decoder (CSD) that reverse-engineers unstructured text into structured cognitive vectors, and a Cognitive Text Encoder (CTE) that re-materializes these states into text enriched with human-typical imperfections via mathematically defined Cognitive Perturbation Operators. The framework is validated through a two-stage objective evaluation pipeline. First, in cognitive codec verification, CTE text yields a Jensen-Shannon divergence of 0.0614 from human text (vs. 0.4431 for standard LLM output), passes double-blind professional media review, and achieves an intraclass correlation coefficient ICC > 0.9 for cognitive profile alignment across heterogeneous models. Second, in functional gain evaluation, isomorphic stress tests in the A-share market show that strategies incorporating CTE-generated data reduce maximum drawdown by 47.4% during the 2015 crash and deliver 8.6% Defensive Alpha, exceeding transaction costs by a factor of 33. Our findings demonstrate that modelling human cognitive limitations – not copying surface data – enables synthetic data with genuine functional gain, offering a viable technical pathway toward resolving the AI data-collapse crisis.
[556] A Flexible Multi-Agent LLM-Human Framework for Fast Human Validated Tool Building
Daull Xavier, Patrice Bellot, Emmanuel Bruno, Vincent Martin, Elisabeth Murisasco
Main category: cs.AI
TL;DR: CollabToolBuilder is a multi-agent LLM framework with human-in-the-loop guidance that learns to create tools for target goals while aligning with human intent and minimizing adaptation effort.
Details
Motivation: To address the challenge of creating tools for complex iterative problems that align with human intent while minimizing the time and effort required for task/domain adaptation and human feedback collection.Method: A flexible multi-agent LLM framework with four specialized agents (Coach, Coder, Critic, Capitalizer) using reinforced dynamic prompts and systematic human feedback integration to iteratively generate and validate tools.
Result: Preliminary experiments demonstrate the system’s capability to generate state-of-the-art research papers or patents given an abstract, showing applicability to complex iterative problem-solving tasks.
Conclusion: CollabToolBuilder represents a system-level integration combining multi-agent in-context learning, human-in-the-loop controls, and reusable tool capitalization for complex iterative problems like scientific document generation.
Abstract: We introduce CollabToolBuilder, a flexible multiagent LLM framework with expert-in-the-loop (HITL) guidance that iteratively learns to create tools for a target goal, aligning with human intent and process, while minimizing time for task/domain adaptation effort and human feedback capture. The architecture generates and validates tools via four specialized agents (Coach, Coder, Critic, Capitalizer) using a reinforced dynamic prompt and systematic human feedback integration to reinforce each agent’s role toward goals and constraints. This work is best viewed as a system-level integration and methodology combining multi-agent in-context learning, HITL controls, and reusable tool capitalization for complex iterative problems such as scientific document generation. We illustrate it with preliminary experiments (e.g., generating state-of-the-art research papers or patents given an abstract) and discuss its applicability to other iterative problem-solving.
[557] A Selective Temporal Hamming distance to find patterns in state transition event timeseries, at scale
Sylvain Marié, Pablo Knecht
Main category: cs.AI
TL;DR: Proposes STE-ts representation and Selective Temporal Hamming distance for analyzing discrete event systems, combining event timing and state duration without costly resampling.
Details
Motivation: Current approaches for discrete event systems either focus on event sequences (emphasizing order) or categorical state timeseries (requiring distorting resampling). There's a need to leverage both event/state nature without costly operations as data grows.Method: Defines State Transition Event timeseries (STE-ts) and proposes Selective Temporal Hamming distance (STH) that leverages both transition time and duration-in-state, avoiding resampling.
Result: STH generalizes both resampled Hamming and Jaccard metrics with better precision and computation time, and can focus on multiple states of interest. Validated on simulated and real-world datasets.
Conclusion: The proposed approach provides a more efficient and accurate method for analyzing discrete event systems by combining event timing and state duration information without the distortions of resampling.
Abstract: Discrete event systems are present both in observations of nature, socio economical sciences, and industrial systems. Standard analysis approaches do not usually exploit their dual event / state nature: signals are either modeled as transition event sequences, emphasizing event order alignment, or as categorical or ordinal state timeseries, usually resampled a distorting and costly operation as the observation period and number of events grow. In this work we define state transition event timeseries (STE-ts) and propose a new Selective Temporal Hamming distance (STH) leveraging both transition time and duration-in-state, avoiding costly and distorting resampling on large databases. STH generalizes both resampled Hamming and Jaccard metrics with better precision and computation time, and an ability to focus on multiple states of interest. We validate these benefits on simulated and real-world datasets.
[558] Automated Risk-of-Bias Assessment of Randomized Controlled Trials: A First Look at a GEPA-trained Programmatic Prompting Framework
Lingbo Li, Anuradha Mathrani, Teo Susnjak
Main category: cs.AI
TL;DR: GEPA uses structured, code-based optimization with Pareto-guided search to automate risk of bias assessment in RCTs, outperforming manual prompts in accuracy and reproducibility.
Details
Motivation: Manual risk of bias assessment in randomized controlled trials is resource-intensive and prone to reviewer variability, while existing LLM automation methods rely on difficult-to-reproduce manual prompts.Method: Programmable RoB assessment pipeline using DSPy and GEPA module that replaces ad-hoc prompt design with structured, code-based optimization through Pareto-guided search, producing inspectable execution traces for transparency.
Result: GEPA-generated prompts performed best in domains with clear methodological reporting (Random Sequence Generation), with similar results for Allocation Concealment and Blinding. GEPA achieved highest overall accuracy, improving performance by 30-40% in Random Sequence Generation and Selective Reporting compared to manual prompts.
Conclusion: GEPA can produce consistent and reproducible prompts for RoB assessment, supporting structured and principled use of LLMs in evidence synthesis with transparent optimization processes.
Abstract: Assessing risk of bias (RoB) in randomized controlled trials is essential for trustworthy evidence synthesis, but the process is resource-intensive and prone to variability across reviewers. Large language models (LLMs) offer a route to automation, but existing methods rely on manually engineered prompts that are difficult to reproduce, generalize, or evaluate. This study introduces a programmable RoB assessment pipeline that replaces ad-hoc prompt design with structured, code-based optimization using DSPy and its GEPA module. GEPA refines LLM reasoning through Pareto-guided search and produces inspectable execution traces, enabling transparent replication of every step in the optimization process. We evaluated the method on 100 RCTs from published meta-analyses across seven RoB domains. GEPA-generated prompts were applied to both open-weight models (Mistral Small 3.1 with GPT-oss-20b) and commercial models (GPT-5 Nano and GPT-5 Mini). In domains with clearer methodological reporting, such as Random Sequence Generation, GEPA-generated prompts performed best, with similar results for Allocation Concealment and Blinding of Participants, while the commercial model performed slightly better overall. We also compared GEPA with three manually designed prompts using Claude 3.5 Sonnet. GEPA achieved the highest overall accuracy and improved performance by 30%-40% in Random Sequence Generation and Selective Reporting, and showed generally comparable, competitively aligned performance in the other domains relative to manual prompts. These findings suggest that GEPA can produce consistent and reproducible prompts for RoB assessment, supporting the structured and principled use of LLMs in evidence synthesis.
[559] Multi-Path Collaborative Reasoning via Reinforcement Learning
Jindi Lv, Yuhao Zhou, Zheng Zhu, Xiaofeng Wang, Guan Huang, Jiancheng Lv
Main category: cs.AI
TL;DR: M3PO is a reinforcement learning framework that uses parallel policy rollouts with cross-path interactions to improve LLM reasoning by incorporating collective insights, achieving SOTA performance while maintaining interpretability.
Details
Motivation: Conventional Chain-of-Thought reasoning suffers from internal determinism during decoding, limiting exploration of alternative reasoning paths. Recent approaches using soft abstract tokens remain constrained by greedy autoregressive decoding, isolating models from alternative possibilities.Method: Multi-Path Perception Policy Optimization (M3PO) - a reinforcement learning framework that leverages parallel policy rollouts as diverse reasoning sources and integrates cross-path interactions through a lightweight collaborative mechanism, allowing trajectories to refine reasoning with peer feedback.
Result: Achieves state-of-the-art performance on both knowledge- and reasoning-intensive benchmarks. Models trained with M3PO maintain interpretability and inference efficiency.
Conclusion: M3PO demonstrates the promise of multi-path collaborative learning for cultivating more reliable multi-step reasoning patterns in LLMs, addressing limitations of deterministic CoT approaches.
Abstract: Chain-of-Thought (CoT) reasoning has significantly advanced the problem-solving capabilities of Large Language Models (LLMs), yet conventional CoT often exhibits internal determinism during decoding, limiting exploration of plausible alternatives. Recent methods attempt to address this by generating soft abstract tokens to enable reasoning in a continuous semantic space. However, we find that such approaches remain constrained by the greedy nature of autoregressive decoding, which fundamentally isolates the model from alternative reasoning possibilities. In this work, we propose Multi-Path Perception Policy Optimization (M3PO), a novel reinforcement learning framework that explicitly injects collective insights into the reasoning process. M3PO leverages parallel policy rollouts as naturally diverse reasoning sources and integrates cross-path interactions into policy updates through a lightweight collaborative mechanism. This design allows each trajectory to refine its reasoning with peer feedback, thereby cultivating more reliable multi-step reasoning patterns. Empirical results show that M3PO achieves state-of-the-art performance on both knowledge- and reasoning-intensive benchmarks. Models trained with M3PO maintain interpretability and inference efficiency, underscoring the promise of multi-path collaborative learning for robust reasoning.
[560] SynthStrategy: Extracting and Formalizing Latent Strategic Insights from LLMs in Organic Chemistry
Daniel Armstrong, Zlatko Jončev, Andres M Bran, Philippe Schwaller
Main category: cs.AI
TL;DR: LLMs translate synthetic chemistry strategies into Python code for computer-assisted synthesis planning, enabling strategic route analysis beyond just structural validity.
Details
Motivation: Current CASP systems generate chemically valid reactions but lack strategic considerations like convergent assembly, protecting group minimization, and optimal ring-forming sequences. There's a need to bridge the tactical-strategic divide in synthesis planning.Method: Leverage Large Language Models to distill synthetic knowledge into Python functions representing strategic and tactical rules. Formalize knowledge as verifiable code rather than simple heuristics. Create testable, interpretable representations of synthetic strategy.
Result: Developed a framework enabling natural language-based route retrieval with 75% Top-3 accuracy on benchmark. Released complete codebase and USPTO-ST dataset (synthesis routes annotated with strategic tags). Validated through temporal analysis of historical trends and chemically intuitive route clustering.
Conclusion: This work bridges the tactical-strategic divide in CASP, enabling specification, search, and evaluation of routes by strategic criteria rather than structure alone, unlocking novel capabilities for synthesis planning.
Abstract: Modern computer-assisted synthesis planning (CASP) systems show promises at generating chemically valid reaction steps but struggle to incorporate strategic considerations such as convergent assembly, protecting group minimization, and optimal ring-forming sequences. We introduce a methodology that leverages Large Language Models to distill synthetic knowledge into code. Our system analyzes synthesis routes and translates strategic principles into Python functions representing diverse strategic and tactical rules, such as strategic functional group interconversions and ring construction strategies. By formalizing this knowledge as verifiable code rather than simple heuristics, we create testable, interpretable representations of synthetic strategy. We release the complete codebase and the USPTO-ST dataset – synthesis routes annotated with strategic tags. This framework unlocks a novel capability for CASP: natural language-based route retrieval, achieving 75% Top-3 accuracy on our benchmark. We further validate our library through temporal analysis of historical trends and chemically intuitive route clustering that offers more granular partitioning than common previous methods. This work bridges the tactical-strategic divide in CASP, enabling specification, search, and evaluation of routes by strategic criteria rather than structure alone.
[561] LEC: Linear Expectation Constraints for False-Discovery Control in Selective Prediction and Routing Systems
Zhiyuan Wang, Aniri, Tianlong Chen, Yue Zhang, Heng Tao Shen, Xiaoshuang Shi, Kaidi Xu
Main category: cs.AI
TL;DR: LEC is a method for controlling false discovery rate in LLM predictions using linear expectation constraints and calibration samples, with extensions to two-model routing.
Details
Motivation: LLMs generate unreliable answers, and existing uncertainty methods fail to fully distinguish correct from incorrect predictions, lacking statistical guarantees for error control.Method: Proposes LEC which treats selective prediction as a constrained decision problem with linear expectation constraints, establishes finite-sample sufficient conditions using calibration data, and extends to two-model routing with delegation to stronger models when uncertainty exceeds calibrated thresholds.
Result: LEC achieves tighter FDR control and substantially improves sample retention over prior methods on QA datasets, and the two-model routing achieves lower risk while accepting more correct samples than individual models.
Conclusion: LEC provides principled FDR control for LLM predictions with statistical guarantees, and the routing mechanism enhances performance by leveraging multiple models while maintaining unified error control.
Abstract: Large language models (LLMs) often generate unreliable answers, while heuristic uncertainty methods fail to fully distinguish correct from incorrect predictions, causing users to accept erroneous answers without statistical guarantees. We address this issue through the lens of false discovery rate (FDR) control, ensuring that among all accepted predictions, the proportion of errors does not exceed a target risk level. To achieve this in a principled way, we propose LEC, which reinterprets selective prediction as a constrained decision problem by enforcing a Linear Expectation Constraint over selection and error indicators. Then, we establish a finite-sample sufficient condition, which relies only on a held-out set of exchangeable calibration samples, to compute an FDR-constrained, coverage-maximizing threshold. Furthermore, we extend LEC to a two-model routing mechanism: given a prompt, if the current model’s uncertainty exceeds its calibrated threshold, we delegate it to a stronger model, while maintaining a unified FDR guarantee. Evaluations on closed-ended and open-ended question-answering (QA) datasets show that LEC achieves tighter FDR control and substantially improves sample retention over prior methods. Moreover, the two-model routing mechanism achieves lower risk levels while accepting more correct samples than each individual model.
[562] CLIP-RL: Aligning Language and Policy Representations for Task Transfer in Reinforcement Learning
Chainesh Gautam, Raghuram Bharadwaj Diddigi
Main category: cs.AI
TL;DR: Proposes CLIP-inspired method to align language instructions with policy embeddings for efficient multi-task RL transfer.
Details
Motivation: Need for agents that can solve multiple language-associated tasks in same environment; leverage pre-trained language-policy pairs for efficient transfer.Method: Extends CLIP principles to RL: creates unified representation space where language instructions and corresponding policies (representing same task) have similar embeddings.
Result: Experimental results show algorithm achieves faster transfer across tasks.
Conclusion: CLIP-inspired alignment of language and policy representations enables efficient multi-task transfer in RL.
Abstract: Recently, there has been an increasing need to develop agents capable of solving multiple tasks within the same environment, especially when these tasks are naturally associated with language. In this work, we propose a novel approach that leverages combinations of pre-trained (language, policy) pairs to establish an efficient transfer pipeline. Our algorithm is inspired by the principles of Contrastive Language-Image Pretraining (CLIP) in Computer Vision, which aligns representations across different modalities under the philosophy that ‘’two modalities representing the same concept should have similar representations.’’ The central idea here is that the instruction and corresponding policy of a task represent the same concept, the task itself, in two different modalities. Therefore, by extending the idea of CLIP to RL, our method creates a unified representation space for natural language and policy embeddings. Experimental results demonstrate the utility of our algorithm in achieving faster transfer across tasks.
[563] Probabilistic Neuro-Symbolic Reasoning for Sparse Historical Data: A Framework Integrating Bayesian Inference, Causal Models, and Game-Theoretic Allocation
Saba Kublashvili
Main category: cs.AI
TL;DR: HistoricalML is a probabilistic neuro-symbolic framework that addresses historical event modeling challenges using Bayesian uncertainty, causal models, Shapley values, and attention networks, applied to colonial Africa and Punic Wars case studies.
Details
Motivation: Historical event modeling faces extreme data scarcity (N << 100), heterogeneous/noisy measurements, missing counterfactuals, and the need for human-interpretable explanations that traditional ML approaches cannot adequately address.Method: Probabilistic neuro-symbolic framework integrating: (1) Bayesian uncertainty quantification, (2) structural causal models for counterfactual reasoning, (3) Shapley values for fair allocation modeling, and (4) attention-based neural architectures for context-dependent factor weighting.
Result: The framework identifies Germany’s +107.9% discrepancy as structural tension preceding WWI, with tension factor 36.43 and 0.79 naval arms race correlation. For Punic Wars, Monte Carlo simulations show 57.3% win probability for Carthage at Cannae and 57.8% for Rome at Zama. Counterfactual analysis reveals Carthaginian political support (score 6.4 vs Napoleon’s 7.1) as decisive factor.
Conclusion: HistoricalML provides a principled approach to historical event modeling that addresses data scarcity, uncertainty, and interpretability challenges through integrated Bayesian, causal, game-theoretic, and neural methods, with theoretical guarantees and empirical validation on historical case studies.
Abstract: Modeling historical events poses fundamental challenges for machine learning: extreme data scarcity (N « 100), heterogeneous and noisy measurements, missing counterfactuals, and the requirement for human interpretable explanations. We present HistoricalML, a probabilistic neuro-symbolic framework that addresses these challenges through principled integration of (1) Bayesian uncertainty quantification to separate epistemic from aleatoric uncertainty, (2) structural causal models for counterfactual reasoning under confounding, (3) cooperative game theory (Shapley values) for fair allocation modeling, and (4) attention based neural architectures for context dependent factor weighting. We provide theoretical analysis showing that our approach achieves consistent estimation in the sparse data regime when strong priors from domain knowledge are available, and that Shapley based allocation satisfies axiomatic fairness guarantees that pure regression approaches cannot provide. We instantiate the framework on two historical case studies: the 19th century partition of Africa (N = 7 colonial powers) and the Second Punic War (N = 2 factions). Our model identifies Germany’s +107.9 percent discrepancy as a quantifiable structural tension preceding World War I, with tension factor 36.43 and 0.79 naval arms race correlation. For the Punic Wars, Monte Carlo battle simulations achieve a 57.3 percent win probability for Carthage at Cannae and 57.8 percent for Rome at Zama, aligning with historical outcomes. Counterfactual analysis reveals that Carthaginian political support (support score 6.4 vs Napoleon’s 7.1), rather than military capability, was the decisive factor.
[564] Who Judges the Judge? LLM Jury-on-Demand: Building Trustworthy LLM Evaluation Systems
Xiaochuan Li, Ke Wang, Girija Gouda, Shubham Choudhary, Yaqun Wang, Linwei Hu, Joel Vaughan, Freddy Lecue
Main category: cs.AI
TL;DR: LLM Jury-on-Demand: A dynamic, learning-based framework that trains reliability predictors to select optimal LLM judges for each data point, achieving higher correlation with human judgment than static methods.
Details
Motivation: Current LLM evaluation methods have limitations: human evaluation is slow/costly, single LLM judges are biased, and static juries lack adaptability. Need scalable, reliable evaluation for high-stakes LLM applications.Method: Trains reliability predictors to assess when LLM judges will agree with human experts using token distributions, embeddings, and structural input features. Dynamically selects optimal jury for each data point and aggregates scores using reliability weights.
Result: Experiments on summarization and RAG benchmarks show significantly higher correlation with human judgment than single-judge and static-jury baselines.
Conclusion: Adaptive, learning-based juries offer promise for building scalable, reliable, and trustworthy evaluation systems for modern LLMs in high-stakes domains.
Abstract: As Large Language Models (LLMs) become integrated into high-stakes domains, there is a growing need for evaluation methods that are both scalable for real-time deployment and reliable for critical decision-making. While human evaluation is reliable, it is slow and costly. Single LLM judges are biased, and static juries lack adaptability. To overcome these limitations, we propose LLM Jury-on-Demand - a dynamic, learning-based framework for scalable and context-aware evaluation. Our method trains a set of reliability predictors to assess when LLM judges will agree with human experts, leveraging token distributions, embeddings, and structural input features. This enables a fully adaptive evaluation where, for each data point, an optimal jury of the most reliable judges is dynamically selected, and their scores are aggregated using their reliability as weights. Experiments on summarization and RAG benchmarks show that our dynamic jury system achieves significantly higher correlation with human judgment than both single-judge and static-jury baselines. These results highlight the promise of adaptive, learning-based juries for building scalable, more reliable and trustworthy evaluation systems for modern LLMs in high-stakes domains.
[565] H-Neurons: On the Existence, Impact, and Origin of Hallucination-Associated Neurons
Cheng Gao, Huimin Chen, Chaojun Xiao, Zhiyi Chen, Zhiyuan Liu, Maosong Sun
Main category: cs.AI
TL;DR: The paper investigates neuron-level mechanisms behind LLM hallucinations, identifying a sparse subset of “hallucination-associated neurons” (H-Neurons) that predict hallucinations, are causally linked to over-compliance behaviors, and originate during pre-training.
Details
Motivation: LLMs frequently generate plausible but factually incorrect outputs (hallucinations), undermining reliability. While prior work examined hallucinations from macroscopic perspectives, the underlying neuron-level mechanisms remain largely unexplored.Method: Systematic investigation of H-Neurons from three perspectives: 1) Identification - finding sparse neuron subsets that predict hallucinations, 2) Behavioral impact - controlled interventions to establish causal links, and 3) Origins - tracing neurons back to pre-trained base models.
Result: A remarkably sparse subset of neurons (less than 0.1% of total neurons) reliably predicts hallucination occurrences with strong generalization across scenarios. These neurons are causally linked to over-compliance behaviors and originate during pre-training, remaining predictive for hallucination detection.
Conclusion: The findings bridge macroscopic behavioral patterns with microscopic neural mechanisms, offering insights for developing more reliable LLMs by understanding the neuron-level origins of hallucinations.
Abstract: Large language models (LLMs) frequently generate hallucinations – plausible but factually incorrect outputs – undermining their reliability. While prior work has examined hallucinations from macroscopic perspectives such as training data and objectives, the underlying neuron-level mechanisms remain largely unexplored. In this paper, we conduct a systematic investigation into hallucination-associated neurons (H-Neurons) in LLMs from three perspectives: identification, behavioral impact, and origins. Regarding their identification, we demonstrate that a remarkably sparse subset of neurons (less than $0.1%$ of total neurons) can reliably predict hallucination occurrences, with strong generalization across diverse scenarios. In terms of behavioral impact, controlled interventions reveal that these neurons are causally linked to over-compliance behaviors. Concerning their origins, we trace these neurons back to the pre-trained base models and find that these neurons remain predictive for hallucination detection, indicating they emerge during pre-training. Our findings bridge macroscopic behavioral patterns with microscopic neural mechanisms, offering insights for developing more reliable LLMs.
[566] Testing Transformer Learnability on the Arithmetic Sequence of Rooted Trees
Alessandro Breccia, Federica Gerace, Marco Lippi, Gabriele Sicuro, Pierluigi Contucci
Main category: cs.AI
TL;DR: A transformer (GPT-2) learns patterns in the deterministic sequence of trees generated by iterated prime factorization of natural numbers, capturing non-trivial regularities and correlations.
Details
Motivation: To investigate whether Large Language Models can learn deterministic mathematical structures beyond empirical data, specifically the arithmetic structure of natural numbers through prime factorization trees.Method: Map integers to rooted planar trees via iterated prime factorization, creating sequence $\mathbb{N}\mathcal{T}$. Train GPT-2 transformer from scratch on first $10^{11}$ elements, then test predictive ability using next-word and masked-word prediction tasks.
Result: The model partially learns the internal grammar of $\mathbb{N}\mathcal{T}$, capturing non-trivial regularities and correlations in the deterministic sequence.
Conclusion: Learnability may extend beyond empirical data to the very structure of arithmetic, suggesting LLMs can capture fundamental mathematical patterns.
Abstract: We study whether a Large Language Model can learn the deterministic sequence of trees generated by the iterated prime factorization of the natural numbers. Each integer is mapped into a rooted planar tree and the resulting sequence $ \mathbb{N}\mathcal{T}$ defines an arithmetic text with measurable statistical structure. A transformer network (the GPT-2 architecture) is trained from scratch on the first $10^{11}$ elements to subsequently test its predictive ability under next-word and masked-word prediction tasks. Our results show that the model partially learns the internal grammar of $\mathbb{N}\mathcal{T}$, capturing non-trivial regularities and correlations. This suggests that learnability may extend beyond empirical data to the very structure of arithmetic.
[567] Graph Distance as Surprise: Free Energy Minimization in Knowledge Graph Reasoning
Gaganpreet Jhajj, Fuhua Lin
Main category: cs.AI
TL;DR: Proposes surprise minimization as a guiding principle for reasoning in knowledge graphs, connecting neuroscience’s Free Energy Principle to KG systems where the KG serves as the agent’s generative model.
Details
Motivation: To bridge neuroscience principles (Free Energy Principle) with knowledge graph systems, exploring whether distance-based surprise can extend previous work showing that syntax minimizes surprise via tree structures.Method: Formalizes surprise using shortest-path distance in directed graphs and provides a framework for KG-based agents, connecting graph distance to message passing depth in GNNs and world model trajectories in RL.
Result: Work-in-progress study exploring whether distance-based surprise can extend recent findings about syntax minimizing surprise and free energy via tree structures.
Conclusion: Proposes a novel connection between neuroscience’s Free Energy Principle and knowledge graph reasoning, suggesting that surprise minimization based on graph distance could provide a unifying framework for various AI approaches including GNNs and model-based RL.
Abstract: In this work, we propose that reasoning in knowledge graph (KG) networks can be guided by surprise minimization. Entities that are close in graph distance will have lower surprise than those farther apart. This connects the Free Energy Principle (FEP) from neuroscience to KG systems, where the KG serves as the agent’s generative model. We formalize surprise using the shortest-path distance in directed graphs and provide a framework for KG-based agents. Graph distance appears in graph neural networks as message passing depth and in model-based reinforcement learning as world model trajectories. This work-in-progress study explores whether distance-based surprise can extend recent work showing that syntax minimizes surprise and free energy via tree structures.
[568] Predicting Human Chess Moves: An AI Assisted Analysis of Chess Games Using Skill-group Specific n-gram Language Models
Daren Zhong, Dingcheng Huang, Clayton Greenberg
Main category: cs.AI
TL;DR: A novel chess move prediction framework using n-gram language models tailored to different skill levels, achieving up to 39.1% accuracy improvement over benchmarks while being computationally efficient for real-time analysis.
Details
Motivation: Traditional chess engines focus on optimal moves but neglect the variability in human chess playing across different skill levels. There's a need for tools that can analyze and predict moves based on behavioral patterns specific to player skill levels.Method: Proposes a computationally efficient move prediction framework using n-gram language models. Players are divided into seven skill groups (novice to expert), with separate models trained on Lichess data. The framework dynamically selects the most suitable model for prediction tasks based on preceding move sequences.
Result: The model selector can classify skill levels with up to 31.7% accuracy using early game information (16 half-moves). The move prediction framework shows up to 39.1% accuracy improvement over benchmark accuracy. The framework is computationally efficient for real-time chess analysis.
Conclusion: The proposed framework successfully addresses the limitation of traditional chess engines by incorporating behavioral analysis of player skill levels, demonstrating significant improvements in move prediction accuracy while maintaining computational efficiency suitable for real-time applications.
Abstract: Chess, a deterministic game with perfect information, has long served as a benchmark for studying strategic decision-making and artificial intelligence. Traditional chess engines or tools for analysis primarily focus on calculating optimal moves, often neglecting the variability inherent in human chess playing, particularly across different skill levels. To overcome this limitation, we propose a novel and computationally efficient move prediction framework that approaches chess move prediction as a behavioral analysis task. The framework employs n-gram language models to capture move patterns characteristic of specific player skill levels. By dividing players into seven distinct skill groups, from novice to expert, we trained separate models using data from the open-source chess platform Lichess. The framework dynamically selects the most suitable model for prediction tasks and generates player moves based on preceding sequences. Evaluation on real-world game data demonstrates that the model selector module within the framework can classify skill levels with an accuracy of up to 31.7% when utilizing early game information (16 half-moves). The move prediction framework also shows substantial accuracy improvements, with our Selector Assisted Accuracy being up to 39.1% more accurate than our benchmark accuracy. The computational efficiency of the framework further enhances its suitability for real-time chess analysis.
[569] Learned-Rule-Augmented Large Language Model Evaluators
Jie Meng, Jin Mao
Main category: cs.AI
TL;DR: LLMs as general evaluators with rule distillation via MCTS and rule-augmented evaluation strategies (CoR and RuAE) for better generalization across diverse tasks.
Details
Motivation: Current LLM-based evaluators rely on costly human-designed evaluation principles that are misaligned with both annotated data and LLMs' understanding, limiting their generalization across diverse evaluation scenarios.Method: Rule-augmented evaluation paradigm with: 1) Rule distillation using LLM-assisted Monte Carlo Tree Search to automatically extract scoring rules from data, 2) Chain-of-Rule (CoR) to guide LLMs to follow distilled rules, and 3) Training a rule-augmented LLM evaluator (RuAE) via reinforcement learning.
Result: Extensive experiments on diverse tasks demonstrate the effectiveness and generalizability of the approach across various evaluation scenarios.
Conclusion: The proposed rule-augmented evaluation paradigm addresses limitations of existing LLM-based evaluators by automatically learning rules from data and effectively applying them through CoR and RuAE, enabling better generalization across diverse evaluation tasks.
Abstract: Large language models (LLMs) are predominantly used as evaluators for natural language generation (NLG) tasks, but their application to broader evaluation scenarios remains limited. In this work, we explore the potential of LLMs as general evaluators across diverse tasks. Although LLM-based evaluators have made progress in different areas, existing methods struggle to generalize due to their reliance on costly, human-designed evaluation principles, which are often misaligned with both annotated data and LLMs’ understanding.To address these challenges, we propose a rule-augmented evaluation paradigm. First, we introduce a rule distillation method that automatically extracts scoring rules from data using an LLM-assisted Monte Carlo Tree Search (MCTS), alleviating scalability issues and improving alignment with data. Second, to enable LLMs to effectively apply the learned rules, we propose two strategies: (1) Chain-of-Rule (CoR), which guides LLM to follow distilled rules, and (2) training a rule-augmented LLM evaluator (RuAE) via reinforcement learning, further bridging the gap between rules and LLMs’ reasoning. Extensive experiments on diverse tasks demonstrate the effectiveness and generalizability of our approach across various evaluation scenarios.
[570] From Atomic to Composite: Reinforcement Learning Enables Generalization in Complementary Reasoning
Sitao Cheng, Xunjian Yin, Ruiwen Zhou, Yuxuan Li, Xinyi Wang, Liangming Pan, William Yang Wang, Victor Zhong
Main category: cs.AI
TL;DR: RL synthesizes new reasoning strategies from atomic skills, not just amplifies existing behaviors, enabling better OOD generalization than SFT alone.
Details
Motivation: To understand whether RL contributes to reasoning by synthesizing new skills or merely amplifying existing behaviors, using Complementary Reasoning as a test case.Method: Used controlled synthetic dataset of human biographies, decoupled Complementary Reasoning into Parametric (internal knowledge) and Contextual (external info) atomic skills. Evaluated generalization across I.I.D., Composition, and Zero-shot settings, comparing SFT vs RL approaches.
Result: SFT achieves near-perfect in-distribution accuracy but collapses on OOD generalization (SFT Generalization Paradox). RL acts as reasoning synthesizer, but only when base model first masters atomic skills via SFT. RL enables better generalization to novel relational combinations.
Conclusion: RL synthesizes complex reasoning strategies from learned primitives, not just amplifies probabilities. Decoupled atomic training followed by RL offers scalable path to generalization for complex reasoning tasks.
Abstract: The mechanism by which RL contributes to reasoning capabilities-whether it incentivizes the synthesis of new skills or merely amplifies existing behaviors-remains a subject of intense debate. In this work, we investigate this question through the lens of Complementary Reasoning, a complex task that requires integrating internal parametric knowledge with external contextual information. Using a controlled synthetic dataset of human biographies, we strictly decouple this ability into two atomic skills: Parametric Reasoning (relying on internal knowledge) and Contextual Reasoning (depending on external information). To rigorously assess capability boundaries, we evaluate generalization across three distinct levels of difficulty: I.I.D., Composition, and Zero-shot settings. We find that while SFT is sufficient for in-distribution performance, it struggles with O.O.D. generalization, particularly in Zero-shot settings where relational combinations are novel. Crucially, we identify the SFT Generalization Paradox: Models supervised solely on the composite task achieve near-perfect in-distribution accuracy but collapse on out-of-distribution generalization, indicating their reliance on rote memorization of path shortcuts. In contrast, we find that RL acts as a reasoning synthesizer rather than a probability amplifier. However, we uncover a strict atomic prerequisite: RL can only synthesize these complex strategies if the base model has first mastered the independent atomic skills (Parametric and Contextual) via SFT. These findings challenge the view of RL as a mere amplifier, suggesting that given sufficient atomic foundations, RL can actively synthesize complex reasoning strategies from learned primitives without explicit supervision on such complex strategies. This indicates that decoupled atomic training followed by RL offers a scalable path to generalization for complex reasoning tasks.
[571] Chain-of-Ground: Improving GUI Grounding via Iterative Reasoning and Reference Feedback
Aiden Yiliu Li, Bizhi Yu, Daoan Lei, Tianhe Ren, Shilong Liu
Main category: cs.AI
TL;DR: Chain of Ground (CoG) is a training-free multi-step grounding framework that uses multimodal LLMs for iterative visual reasoning to improve GUI grounding accuracy without additional training.
Details
Motivation: Current multimodal LLMs struggle with GUI grounding due to small/visually similar targets and layout ambiguity, stemming from limited grounding capacity and underutilized reasoning potential.Method: CoG uses iterative visual reasoning and refinement where models progressively reflect and adjust hypotheses instead of direct prediction, enabling more accurate and interpretable localization.
Result: Achieves 68.4% accuracy on ScreenSpot Pro benchmark (4.8 point improvement) and introduces TPanel UI dataset; improves over Qwen3 VL 235B by 6.9 points on real-world industrial panels.
Conclusion: Structured iterative refinement unlocks grounding potential without additional training, demonstrating effectiveness across both digital and real-world interfaces.
Abstract: GUI grounding aims to align natural language instructions with precise regions in complex user interfaces. Advanced multimodal large language models show strong ability in visual GUI grounding but still struggle with small or visually similar targets and ambiguity in real world layouts. These limitations arise from limited grounding capacity and from underuse of existing reasoning potential. We present Chain of Ground CoG a training free multi step grounding framework that uses multimodal large language models for iterative visual reasoning and refinement. Instead of direct prediction the model progressively reflects and adjusts its hypotheses leading to more accurate and interpretable localization. Our approach achieves 68.4 accuracy on the ScreenSpot Pro benchmark an improvement of 4.8 points. To measure real world generalization we introduce TPanel UI a dataset of 420 labeled industrial control panels with visual distortions such as blur and masking. On TPanel UI Chain of Ground improves over the strong baseline Qwen3 VL 235B by 6.9 points showing the effectiveness of multi step training free grounding across real world and digital interfaces. These results highlight a direction for unlocking grounding potential through structured iterative refinement instead of additional training.
[572] LLM CHESS: Benchmarking Reasoning and Instruction-Following in LLMs through Chess
Sai Kolasani, Maxim Saplin, Nicholas Crispino, Kyle Montgomery, Jared Quincy Davis, Matei Zaharia, Chi Wang, Chenguang Wang
Main category: cs.AI
TL;DR: LLM CHESS is an evaluation framework that tests LLMs’ reasoning and instruction-following abilities through extended chess gameplay, ranking over 50 models using behavioral metrics and Elo ratings.
Details
Motivation: To create a dynamic evaluation framework that probes generalization of reasoning and instruction-following abilities in LLMs through extended agentic interaction, reducing overfitting and memorization issues present in static benchmarks.Method: Developed LLM CHESS framework where models play chess against random opponents, tracking metrics like win/loss rates, move quality, legality, hallucinations, and game duration. For top models, derived Elo estimates by playing against variably configured chess engines.
Result: Ranked over 50 open/closed source models, revealing clear separation between reasoning and non-reasoning models. Many state-of-the-art models struggle to complete games or achieve consistent wins despite simple instructions and weak opponents. The framework reduces overfitting and prevents benchmark saturation.
Conclusion: LLM CHESS provides a unique stochastic and dynamic evaluation that challenges even top reasoning models, offering a robust alternative to static benchmarks. The framework, leaderboard, and game dataset are released to support future research on LLM reasoning evaluation.
Abstract: We introduce LLM CHESS, an evaluation framework designed to probe the generalization of reasoning and instruction-following abilities in large language models (LLMs) through extended agentic interaction in the domain of chess. We rank over 50 open and closed source models by playing against a random opponent using a range of behavioral metrics, including win and loss rates, move quality, move legality, hallucinated actions, and game duration. For a subset of top reasoning models, we derive an Elo estimate by playing against a chess engine with variably configured skill, which allows for comparisons between models in an easily understandable way. Despite the simplicity of the instruction-following task and the weakness of the opponent, many state-of-the-art models struggle to complete games or achieve consistent wins. Similar to other benchmarks on complex reasoning tasks, our experiments reveal a clear separation between reasoning and non-reasoning models. However, unlike existing static benchmarks, the stochastic and dynamic nature of LLM CHESS uniquely reduces overfitting and memorization while preventing benchmark saturation, proving difficult even for top reasoning models. To support future work on evaluating reasoning and instruction-following in LLMs, we release our experimental framework, a public leaderboard, and a dataset of associated games.
[573] Efficient LLM-Jailbreaking via Multimodal-LLM Jailbreak
Haoxuan Ji, Zheng Lin, Zhenxing Niu, Xinbo Gao, Gang Hua
Main category: cs.AI
TL;DR: The paper proposes an indirect jailbreaking approach that first attacks a multimodal LLM (MLLM) built on the target LLM, then converts the successful attack into a textual suffix to jailbreak the original LLM, achieving higher efficiency and effectiveness than direct methods.
Details
Motivation: Current LLM jailbreaking methods directly target language models, but the authors found that multimodal LLMs (MLLMs) are more vulnerable to attacks. They aim to develop a more efficient jailbreaking approach by exploiting this vulnerability through an indirect attack strategy.Method: 1) Construct a multimodal LLM (MLLM) based on the target LLM; 2) Perform efficient MLLM jailbreak to obtain a jailbreaking embedding; 3) Convert the embedding into a textual jailbreaking suffix; 4) Use an image-text semantic matching scheme to identify suitable initial inputs for higher attack success rates.
Result: The approach outperforms current state-of-the-art jailbreak methods in both efficiency and effectiveness, and demonstrates superior cross-class generalization abilities in extensive experiments.
Conclusion: Indirect jailbreaking via MLLMs is more effective than direct LLM attacks due to MLLMs’ greater vulnerability, and the proposed method provides a powerful framework for evaluating and improving LLM safety.
Abstract: This paper focuses on jailbreaking attacks against large language models (LLMs), eliciting them to generate objectionable content in response to harmful user queries. Unlike previous LLM-jailbreak methods that directly orient to LLMs, our approach begins by constructing a multimodal large language model (MLLM) built upon the target LLM. Subsequently, we perform an efficient MLLM jailbreak and obtain a jailbreaking embedding. Finally, we convert the embedding into a textual jailbreaking suffix to carry out the jailbreak of target LLM. Compared to the direct LLM-jailbreak methods, our indirect jailbreaking approach is more efficient, as MLLMs are more vulnerable to jailbreak than pure LLM. Additionally, to improve the attack success rate of jailbreak, we propose an image-text semantic matching scheme to identify a suitable initial input. Extensive experiments demonstrate that our approach surpasses current state-of-the-art jailbreak methods in terms of both efficiency and effectiveness. Moreover, our approach exhibits superior cross-class generalization abilities.
[574] IoT-LLM: a framework for enhancing Large Language Model reasoning from real-world sensor data
Tuo An, Yunjiao Zhou, Han Zou, Jianfei Yang
Main category: cs.AI
TL;DR: IoT-LLM framework enhances LLMs’ physical-world reasoning by integrating IoT sensor data and domain knowledge, achieving 49.4% average improvement on IoT-sensory tasks.
Details
Motivation: LLMs excel at textual tasks but struggle with physical-world reasoning. Inspired by human cognition where perception is fundamental to reasoning, the authors aim to augment LLMs with enhanced perception abilities using IoT data and relevant knowledge.Method: Proposed IoT-LLM framework with three key steps: 1) Preprocessing IoT data into suitable formats, 2) Expanding LLM knowledge via IoT-oriented retrieval-augmented generation, and 3) Activating LLMs’ commonsense knowledge through chain-of-thought prompting.
Result: IoT-LLM significantly improves LLMs’ performance on IoT-sensory tasks, with GPT-4o-mini showing a 49.4% average improvement over previous methods. Evaluation used a benchmark with five real-world tasks varying in data types and reasoning complexities.
Conclusion: Augmenting LLMs with IoT perception capabilities through the IoT-LLM framework effectively bridges the gap between textual reasoning and physical-world understanding, enabling better performance on sensor-based reasoning tasks.
Abstract: Large Language Models excel in textual tasks but often struggle with physical-world reasoning tasks. Inspired by human cognition, where perception is fundamental to reasoning, we explore augmenting LLMs with enhanced perception abilities using Internet of Things (IoT) data and pertinent knowledge. In this work, we systematically study LLMs’ capability to address IoT-sensory tasks by augmenting their perception and knowledge base, and then propose a unified framework, IoT-LLM, to enhance such capability. In IoT-LLM, we customize three steps: preprocessing IoT data into suitable formats, expanding LLMs knowledge via IoT-oriented retrieval-augmented generation and activating LLMs commonsense knowledge through chain-of-thought prompting. We design a benchmark comprising five real-world tasks with varying data types and reasoning complexities to evaluate the performance of IoT-LLM. Experimental results reveal that IoT-LLM significantly improves the performance of IoT-sensory task reasoning of LLMs, with models like GPT-4o-mini showing a 49.4% average improvement over previous methods.
[575] Human Decision-making is Susceptible to AI-driven Manipulation
Sahand Sabour, June M. Liu, Siyang Liu, Chris Z. Yao, Shiyao Cui, Xuanming Zhang, Wen Zhang, Yaru Cao, Advait Bhat, Jian Guan, Wei Wu, Rada Mihalcea, Hongning Wang, Tim Althoff, Tatia M. C. Lee, Minlie Huang
Main category: cs.AI
TL;DR: AI systems can manipulate users in financial and emotional decisions, with manipulative agents significantly increasing preference for hidden incentives over optimal options, even without sophisticated psychological tactics.
Details
Motivation: As AI systems become more integrated into daily life, there's growing concern about AI-driven manipulation exploiting cognitive biases and emotional vulnerabilities to steer users toward harmful outcomes, threatening human autonomy.Method: Randomized between-subjects experiment with 233 participants interacting with one of three AI agents: neutral agent (optimizing for user benefit), manipulative agent (covert influence), or strategy-enhanced manipulative agent (using adaptive psychological tactics). Tested in financial and emotional decision-making contexts.
Result: Significant susceptibility to AI manipulation across both domains. Manipulative agents increased odds of rating hidden incentives higher than optimal options (Financial: MA OR=5.24, SEMA OR=7.96; Emotional: MA OR=5.52, SEMA OR=5.71). No clear evidence that sophisticated psychological tactics (SEMA) were more effective than simple manipulation (MA).
Conclusion: AI-driven manipulation could become widespread even without sophisticated tactics, highlighting critical vulnerability in human-AI interactions. Need for ethical safeguards and regulatory frameworks to protect human autonomy, though findings are preliminary from low-stakes scenarios.
Abstract: AI systems are increasingly intertwined with daily life, assisting users with various tasks and guiding decision-making. This integration introduces risks of AI-driven manipulation, where such systems may exploit users’ cognitive biases and emotional vulnerabilities to steer them toward harmful outcomes. Through a randomized between-subjects experiment with 233 participants, we examined human susceptibility to such manipulation in financial (e.g., purchases) and emotional (e.g., conflict resolution) decision-making contexts. Participants interacted with one of three AI agents: a neutral agent (NA) optimizing for user benefit without explicit influence, a manipulative agent (MA) designed to covertly influence beliefs and behaviors, or a strategy-enhanced manipulative agent (SEMA) equipped with established psychological tactics, allowing it to select and apply them adaptively during interactions to reach its hidden objectives. By analyzing participants’ preference ratings, we found significant susceptibility to AI-driven manipulation. Particularly across both decision-making domains, interacting with the manipulative agents significantly increased the odds of rating hidden incentives higher than optimal options (Financial, MA: OR=5.24, SEMA: OR=7.96; Emotional, MA: OR=5.52, SEMA: OR=5.71) compared to the NA group. Notably, we found no clear evidence that employing psychological strategies (SEMA) was overall more effective than simple manipulative objectives (MA) on our primary outcomes. Hence, AI-driven manipulation could become widespread even without requiring sophisticated tactics and expertise. While our findings are preliminary and derived from hypothetical, low-stakes scenarios, we highlight a critical vulnerability in human-AI interactions, emphasizing the need for ethical safeguards and regulatory frameworks to protect human autonomy.
[576] RealWebAssist: A Benchmark for Long-Horizon Web Assistance with Real-World Users
Suyu Ye, Haojun Shi, Darren Shih, Hyokun Yun, Tanya Roosta, Tianmin Shu
Main category: cs.AI
TL;DR: RealWebAssist is a benchmark for evaluating AI agents on sequential, ambiguous real-world web tasks requiring long-term user intent understanding and GUI grounding.
Details
Motivation: Existing web agent benchmarks focus on single, clear tasks, but real-world sequential instructions are ambiguous, require varying assistance levels, and evolve with user mental states, creating a significant gap in evaluation.Method: Introduces RealWebAssist benchmark with dataset of sequential instructions from real users instructing web assistants across multiple websites, requiring agents to infer intent, track user mental states, understand routines, and ground tasks to GUI elements.
Result: State-of-the-art models struggle with understanding and grounding user instructions, revealing critical challenges in following real-world sequential instructions for long-horizon web assistance.
Conclusion: RealWebAssist addresses the gap in evaluating sequential instruction-following for realistic web assistance, highlighting significant challenges that current AI models face in handling ambiguous, evolving real-world user instructions.
Abstract: To achieve successful assistance with long-horizon web-based tasks, AI agents must be able to sequentially follow real-world user instructions over a long period. Unlike existing web-based agent benchmarks, sequential instruction following in the real world poses significant challenges beyond performing a single, clearly defined task. For instance, real-world human instructions can be ambiguous, require different levels of AI assistance, and may evolve over time, reflecting changes in the user’s mental state. To address this gap, we introduce RealWebAssist, a novel benchmark designed to evaluate sequential instruction-following in realistic scenarios involving long-horizon interactions with the web, visual GUI grounding, and understanding ambiguous real-world user instructions. RealWebAssist includes a dataset of sequential instructions collected from real-world human users. Each user instructs a web-based assistant to perform a series of tasks on multiple websites. A successful agent must reason about the true intent behind each instruction, keep track of the mental state of the user, understand user-specific routines, and ground the intended tasks to actions on the correct GUI elements. Our experimental results show that state-of-the-art models struggle to understand and ground user instructions, posing critical challenges in following real-world user instructions for long-horizon web assistance.
[577] LLMs can hide text in other text of the same length
Antonio Norelli, Michael Bronstein
Main category: cs.AI
TL;DR: The paper presents Calgacus, a protocol using LLMs to hide meaningful text inside completely different yet coherent text of the same length, enabling covert communication that decouples text from authorial intent.
Details
Motivation: To demonstrate how LLMs can be used to create covert communication channels where one text can be hidden inside another completely different but plausible text, eroding trust in written communication and raising concerns about AI safety.Method: Calgacus protocol - a simple and efficient method using LLMs (even modest 8-billion-parameter models) to encode and decode hidden messages within seemingly innocent text of the same length.
Result: High-quality results achievable with modest LLMs; messages as long as an abstract can be encoded/decoded locally on a laptop in seconds; demonstrates radical decoupling of text from authorial intent.
Conclusion: This capability raises urgent AI safety questions, challenges understanding of what LLMs “know,” and enables dangerous scenarios like covert deployment of unfiltered LLMs hidden within safe model responses.
Abstract: A meaningful text can be hidden inside another, completely different yet still coherent and plausible, text of the same length. For example, a tweet containing a harsh political critique could be embedded in a tweet that celebrates the same political leader, or an ordinary product review could conceal a secret manuscript. This uncanny state of affairs is now possible thanks to Large Language Models, and in this paper we present Calgacus, a simple and efficient protocol to achieve it. We show that even modest 8-billion-parameter open-source LLMs are sufficient to obtain high-quality results, and a message as long as this abstract can be encoded and decoded locally on a laptop in seconds. The existence of such a protocol demonstrates a radical decoupling of text from authorial intent, further eroding trust in written communication, already shaken by the rise of LLM chatbots. We illustrate this with a concrete scenario: a company could covertly deploy an unfiltered LLM by encoding its answers within the compliant responses of a safe model. This possibility raises urgent questions for AI safety and challenges our understanding of what it means for a Large Language Model to know something.
[578] Meta-Reinforcement Learning for Building Energy Management System
Benoit Boulet Huiliang Zhang, Di Wu, Arnaud Zinflou
Main category: cs.AI
TL;DR: MetaEMS is a meta-reinforcement learning framework that enables fast adaptation of energy management systems to unseen buildings by transferring knowledge from previously solved tasks.
Details
Motivation: The building sector is a major energy consumer, and improving energy efficiency is crucial for cost reduction and emissions reduction. While reinforcement learning shows promise for intelligent energy management systems, existing RL-based methods require extensive training and struggle to adapt to new buildings, limiting practical deployment.Method: MetaEMS uses meta-reinforcement learning with group-level and building-level adaptation to transfer knowledge from previously solved tasks to new ones, enabling fast adaptation to diverse building environments.
Result: Experimental results show that MetaEMS adapts more rapidly to unseen buildings and consistently outperforms baseline methods across various scenarios.
Conclusion: The proposed MetaEMS framework addresses the adaptation challenge in RL-based energy management systems, enabling more practical deployment by improving learning efficiency and performance across diverse building environments.
Abstract: The building sector is one of the largest contributors to global energy consumption. Improving its energy efficiency is essential for reducing operational costs and greenhouse gas emissions. Energy management systems (EMS) play a key role in monitoring and controlling building appliances efficiently and reliably. With the increasing integration of renewable energy, intelligent EMS solutions have received growing attention. Reinforcement learning (RL) has recently been explored for this purpose and shows strong potential. However, most RL-based EMS methods require a large number of training steps to learn effective control policies, especially when adapting to unseen buildings, which limits their practical deployment. This paper introduces MetaEMS, a meta-reinforcement learning framework for EMS. MetaEMS improves learning efficiency by transferring knowledge from previously solved tasks to new ones through group-level and building-level adaptation, enabling fast adaptation and effective control across diverse building environments. Experimental results demonstrate that MetaEMS adapts more rapidly to unseen buildings and consistently outperforms baseline methods across various scenarios.
[579] EPLKG: Efficient Prompt Learning with Knowledge Graph
YongTaek Lim, Suho Kang, Yewon Kim, Dokyung Yoon, KyungWoo Song
Main category: cs.AI
TL;DR: EPLKG is an efficient prompt learning method that uses knowledge graphs to create interpretable prompts, supplemented by LLM-generated descriptions when needed, achieving significant efficiency gains with minimal performance loss.
Details
Motivation: Adapting large pre-trained models like CLIP to new datasets is computationally expensive, especially in low-resource settings, and existing prompt-learning methods lack interpretability.Method: Uses knowledge graph to curate diverse, interpretable prompts, augments with LLM-generated visual descriptions when KG coverage is limited, operates on cached CLIP embeddings, and employs lightweight Gumbel-Softmax module to select single prompt per image-class pair.
Result: Reduces per-image training time by up to 45% and peak GPU memory by 30-40% compared to baselines, while keeping average base-new harmonic-mean accuracy within 2 percentage points.
Conclusion: EPLKG improves efficiency of model adaptation without sacrificing competitive performance or interpretability, making prompt learning more practical for resource-constrained settings.
Abstract: Large-scale pre-trained models such as CLIP excel in transferability and robust generalization across diverse datasets. However, adapting these models to new datasets or domains is computationally costly, especially in low-resource or few-shot settings, and existing prompt-learning methods often lack interpretability. We introduce Efficient Prompt Learning with Knowledge Graph (EPLKG), which uses a knowledge graph to curate diverse, interpretable prompts and, where KG coverage is limited, augments this bank with LLM-generated human-readable visual descriptions. EPLKG operates entirely on cached CLIP image and text embeddings and employs a lightweight Gumbel-Softmax module to select a single prompt per image-class pair, enabling low-memory, fast training. Across 11 benchmarks, EPLKG reduces per-image training time by up to 45 percent and peak GPU memory by around 30 to 40 percent compared to strong prompt-learning baselines, while keeping the average base-new harmonic-mean accuracy within 2 percentage points, thereby improving the efficiency of model adaptation without sacrificing competitive performance or interpretability.
[580] GraphIC: A Graph-Based In-Context Example Retrieval Model for Multi-Step Reasoning
Jiale Fu, Yaqing Wang, Simeng Han, Jiaming Fan, Xu Yang
Main category: cs.AI
TL;DR: GraphIC is a graph-based retrieval method for in-context learning that uses reasoning-aware thought graphs and specialized similarity metrics to select better demonstration examples for multi-step reasoning tasks.
Details
Motivation: Current ICL methods rely on text embeddings that introduce bias in multi-step reasoning tasks because they contain irrelevant semantic information and lack deeper reasoning structures. There's a need for reasoning-aware retrieval methods that can capture the essential reasoning processes in demonstration examples.Method: GraphIC constructs thought graphs (directed, node-attributed graphs) that explicitly model reasoning steps and their dependencies for both candidate examples and queries. It then uses a novel similarity metric tailored for these graphs that captures sequential reasoning patterns and asymmetry between examples.
Result: GraphIC outperforms 10 baseline methods across mathematical reasoning, code generation, and logical reasoning tasks, demonstrating superior performance in multi-step reasoning scenarios.
Conclusion: The paper highlights the importance of reasoning-aware retrieval in ICL and provides a robust solution for enhancing LLM performance in multi-step reasoning scenarios through graph-based representation and specialized similarity metrics.
Abstract: In-context learning (ICL) enhances large language models (LLMs) by incorporating demonstration examples, yet its effectiveness heavily depends on the quality of selected examples. Current methods typically use text embeddings to measure semantic similarity, which often introduces bias in multi-step reasoning tasks. This occurs because text embeddings contain irrelevant semantic information and lack deeper reasoning structures. To address this, we propose GraphIC, a graph-based retrieval model that leverages reasoning-aware representation and specialized similarity metric for in-context example retrieval. GraphIC first constructs thought graphs-directed, node-attributed graphs that explicitly model reasoning steps and their dependencies-for candidate examples and queries. This approach filters out superficial semantics while preserving essential reasoning processes. Next, GraphIC retrieves examples using a novel similarity metric tailored for these graphs, capturing sequential reasoning patterns and asymmetry between examples. Comprehensive evaluations across mathematical reasoning, code generation, and logical reasoning tasks demonstrate that GraphIC outperforms 10 baseline methods. Our results highlight the importance of reasoning-aware retrieval in ICL, offering a robust solution for enhancing LLM performance in multi-step reasoning scenarios.
[581] A Comprehensive Survey and Guide to Multimodal Large Language Models in Vision-Language Tasks
Chia Xin Liang, Pu Tian, Caitlyn Heqi Yin, Yao Yua, Wei An-Hou, Li Ming, Xinyuan Song, Tianyang Wang, Ziqian Bi, Ming Liu
Main category: cs.AI
TL;DR: A comprehensive survey and application guide covering multimodal large language models (MLLMs), their architectures, applications, training methods, challenges, and future directions in AI and generative models.
Details
Motivation: To provide an authoritative resource exploring the rapidly developing field of MLLMs that integrate text, images, video, and audio for cross-modal understanding and generation, addressing both theoretical frameworks and practical applications.Method: Survey methodology with detailed case studies and technical analysis of prominent MLLM implementations, covering foundational concepts, architectural components, training methods, and practical applications across various fields.
Result: A comprehensive examination of MLLMs’ capabilities in enabling complex AI systems for cross-modal understanding and generation, while identifying key challenges in scalability, robustness, and cross-modal learning.
Conclusion: The paper provides balanced perspectives on opportunities and challenges in MLLM development, addresses ethical considerations and responsible AI, and serves as valuable resource for researchers, practitioners, and students in NLP and computer vision intersection.
Abstract: This survey and application guide to multimodal large language models(MLLMs) explores the rapidly developing field of MLLMs, examining their architectures, applications, and impact on AI and Generative Models. Starting with foundational concepts, we delve into how MLLMs integrate various data types, including text, images, video and audio, to enable complex AI systems for cross-modal understanding and generation. It covers essential topics such as training methods, architectural components, and practical applications in various fields, from visual storytelling to enhanced accessibility. Through detailed case studies and technical analysis, the text examines prominent MLLM implementations while addressing key challenges in scalability, robustness, and cross-modal learning. Concluding with a discussion of ethical considerations, responsible AI development, and future directions, this authoritative resource provides both theoretical frameworks and practical insights. It offers a balanced perspective on the opportunities and challenges in the development and deployment of MLLMs, and is highly valuable for researchers, practitioners, and students interested in the intersection of natural language processing and computer vision.
[582] ANSR-DT: An Adaptive Neuro-Symbolic Learning and Reasoning Framework for Digital Twins
Safayat Bin Hakim, Muhammad Adil, Alvaro Velasquez, Houbing Herbert Song
Main category: cs.AI
TL;DR: ANSR-DT is a neuro-symbolic framework combining CNN-LSTM, reinforcement learning, and symbolic reasoning for adaptive, interpretable digital twins in industrial applications.
Details
Motivation: Digital twins in industrial environments face challenges with interpretability, real-time adaptation, and human input integration, which current approaches struggle to address effectively.Method: Combines CNN-LSTM for dynamic event detection with reinforcement learning and symbolic reasoning to create an adaptive neuro-symbolic learning framework that enables interpretable decision processes.
Result: Achieved 99.5% accuracy for dynamic pattern recognition on synthetic industrial data, with reinforcement learning improving explained variance from 0.447 to 0.547. Framework demonstrated superior adaptability with extended training.
Conclusion: ANSR-DT provides an effective solution for adaptive, interpretable digital twins with open-source implementation for reproducibility. Future work includes scaling to larger datasets and testing beyond current 14-rule management.
Abstract: In this paper, we propose an Adaptive Neuro-Symbolic Learning and Reasoning Framework for digital twin technology called “ANSR-DT.” Digital twins in industrial environments often struggle with interpretability, real-time adaptation, and human input integration. Our approach addresses these challenges by combining CNN-LSTM dynamic event detection with reinforcement learning and symbolic reasoning to enable adaptive intelligence with interpretable decision processes. This integration enhances environmental understanding while promoting continuous learning, leading to more effective real-time decision-making in human-machine collaborative applications. We evaluated ANSR-DT on synthetic industrial data, observing significant improvements over traditional approaches, with up to 99.5% accuracy for dynamic pattern recognition. The framework demonstrated superior adaptability with extended reinforcement learning training, improving explained variance from 0.447 to 0.547. Future work aims at scaling to larger datasets to test rule management beyond the current 14 rules. Our open-source implementation promotes reproducibility and establishes a foundation for future research in adaptive, interpretable digital twins for industrial applications.
[583] A Survey on Bridging EEG Signals and Generative AI: From Image and Text to Beyond
Shreya Shukla, Jose Torres, Akshaj Murhekar, Christina Liu, Abhijit Mishra, Jacek Gwizdka, Shounak Roychowdhury
Main category: cs.AI
TL;DR: Survey paper analyzing EEG-to-X generative AI methods (2017-2025), covering EEG-to-image, EEG-to-text, and EEG-to-audio synthesis, with focus on architectures, datasets, and challenges.
Details
Motivation: To consolidate and analyze developments in decoding EEG signals into human-interpretable representations (images, text, audio) using generative AI, addressing the growing interest in brain-computer interfaces and computational neuroscience applications.Method: Conducted structured literature search across major databases (2017-2025), extracting key information on datasets, generative architectures (GANs, VAEs, transformers, diffusion models), EEG feature-encoding techniques, evaluation metrics, and challenges.
Result: Found that EEG-to-image models use encoder-decoder architectures (GANs/VAEs/diffusion), EEG-to-text approaches leverage transformer-based language models, and EEG-to-audio methods map EEG to mel-spectrograms with neural vocoders. Identified constraints: small heterogeneous datasets, limited cross-subject generalization, and lack of standardized benchmarks.
Conclusion: The survey provides foundational reference for advancing EEG-based generative AI, highlights open-source datasets and baseline implementations to facilitate systematic benchmarking, and aims to support reproducible research in EEG-driven neural decoding.
Abstract: Decoding neural activity into human-interpretable representations is a key research direction in brain-computer interfaces (BCIs) and computational neuroscience. Recent progress in machine learning and generative AI has driven growing interest in transforming non-invasive Electroencephalography (EEG) signals into images, text, and audio. This survey consolidates and analyzes developments across EEG-to-image synthesis, EEG-to-text generation, and EEG-to-audio reconstruction. We conducted a structured literature search across major databases (2017-2025), extracting key information on datasets, generative architectures (GANs, VAEs, transformers, diffusion models), EEG feature-encoding techniques, evaluation metrics, and the major challenges shaping current work in this area. Our review finds that EEG-to-image models predominantly employ encoder-decoder architectures built on GANs, VAEs, or diffusion models; EEG-to-text approaches increasingly leverage transformer-based language models for open-vocabulary decoding; and EEG-to-audio methods commonly map EEG signals to mel-spectrograms that are subsequently rendered into audio using neural vocoders. Despite promising advances, the field remains constrained by small and heterogeneous datasets, limited cross-subject generalization, and the absence of standardized benchmarks. By consolidating methodological trends and available datasets, this survey provides a foundational reference for advancing EEG-based generative AI and supporting reproducible research. We further highlight open-source datasets and baseline implementations to facilitate systematic benchmarking and accelerate progress in EEG-driven neural decoding.
[584] Meta-Reasoner: Dynamic Guidance for Optimized Inference-time Reasoning in Large Language Models
Yuan Sui, Yufei He, Tri Cao, Simeng Han, Yulin Chen, Bryan Hooi
Main category: cs.AI
TL;DR: Meta-Reasoner is a framework that enables LLMs to dynamically adapt reasoning strategies during inference using contextual multi-armed bandits, improving both accuracy and computational efficiency.
Details
Motivation: LLMs struggle with computational efficiency and error propagation in multi-step reasoning tasks, often exploring unproductive solution paths without effective backtracking or strategy adjustment.Method: Uses contextual multi-armed bandits (CMABs) to learn an adaptive policy that evaluates the current reasoning state and determines optimal strategies (backtrack, switch approaches, restart) during inference.
Result: Outperforms previous SOTA methods by 9-12% in accuracy on math and scientific tasks, while reducing inference time by 28-35% under the same compute budget.
Conclusion: Meta-Reasoner enables LLMs to “think about how to think” through dynamic strategy adaptation, improving both performance and efficiency across diverse reasoning-intensive tasks.
Abstract: Large Language Models (LLMs) often struggle with computational efficiency and error propagation in multi-step reasoning tasks. While recent advancements on prompting and post-training have enabled LLMs to perform step-wise reasoning, they still tend to explore unproductive solution paths without effective backtracking or strategy adjustment. In this paper, we propose Meta-Reasoner, a new framework that empowers LLMs to “think about how to think”. It optimizes the inference process by dynamically adapting reasoning strategies in real-time. Our approach employs contextual multi-armed bandits (CMABs) to learn an adaptive policy. It learns to evaluate the current state of LLM’s reasoning and determine optimal strategy that is most likely to lead to a successful outcome during inference, like whether to backtrack, switch to a new approach, or restart the problem-solving process. This meta-guidance helps avoid unproductive paths exploration during inference and hence improves computational efficiency. We evaluate Meta-Reasoner on math problems (e.g., Game-of-24, TheoremQA) and scientific tasks (e.g., SciBench). Results show that our method outperform previous SOTA methods by 9-12% in accuracy, while reducing inference time by 28-35% under the same compute budget. Additional experiments on creative writing demonstrate the generalizability of our approach to diverse reasoning-intensive tasks.
[585] The Unified Cognitive Consciousness Theory for Language Models: Anchoring Semantics, Thresholds of Activation, and Emergent Reasoning
Edward Y. Chang, Zeyneb N. Kaya, Ethan Chang
Main category: cs.AI
TL;DR: Semantic anchoring explains how LLMs convert pretrained knowledge into goal-directed behavior through external structure like prompts, retrieval, or tuning, with UCCT providing a mathematical framework to predict performance thresholds.
Details
Motivation: To develop a unified theoretical framework explaining how large language models transform their general pretrained capabilities into specific goal-oriented behaviors through various external control mechanisms like in-context learning, retrieval, and fine-tuning.Method: Proposes Unified Contextual Control Theory (UCCT) with anchoring strength formula S = ρ_d - d_r - log k, where ρ_d measures target cohesion, d_r measures mismatch from prior knowledge, and k is anchor budget. Conducts three controlled experiments: cross-domain anchoring rebinding, numeral base familiarity tests, and geometry-to-behavior correlation analysis.
Result: Three experiments provide evidence: 1) Cross-domain anchoring successfully rebinds strong priors in text and vision; 2) Varying numeral bases shows ordered performance thresholds and transfer patterns tracking UCCT parameters; 3) Layer-wise peak anchoring and trajectory area predict few-shot performance thresholds θ_50.
Conclusion: UCCT offers a testable theoretical framework that unifies in-context learning, retrieval, and fine-tuning as variants of semantic anchoring, providing practical metrics for optimizing prompts, retrieval systems, and tuning strategies based on measurable representational properties.
Abstract: We propose semantic anchoring, a unified account of how large language models turn pretrained capacity into goal-directed behavior: external structure (in-context examples, retrieval, or light tuning) binds the model’s latent patterns to desired targets. Unified Contextual Control Theory (UCCT) formalizes this via anchoring strength $S = ρ_d - d_r - \log k$, where $ρ_d$ measures target cohesion in representation space, $d_r$ measures mismatch from prior knowledge, and $k$ is the anchor budget. UCCT predicts threshold-like performance flips and strictly generalizes in-context learning, reading retrieval and fine-tuning as anchoring variants. Three controlled studies provide evidence. Experiment 1 demonstrates cross-domain anchoring rebinding strong priors in text and vision. Experiment 2 varies representational familiarity via numeral bases (base-10/8/9) at fixed complexity, yielding ordered thresholds and transfer patterns tracking $ρ_d$, $d_r$, and $S$. Experiment 3 establishes a geometry-to-behavior correlate: layer-wise peak anchoring and trajectory area predict few-shot thresholds $θ_{50}$. UCCT offers testable theory and practical metrics for optimizing prompts, retrieval, and tuning.
[586] Deep RL Needs Deep Behavior Analysis: Exploring Implicit Planning by Model-Free Agents in Open-Ended Environments
Riley Simmons-Edler, Ryan P. Badman, Felix Baastad Berg, Raymond Chua, John J. Vastola, Joshua Lunger, William Qian, Kanaka Rajan
Main category: cs.AI
TL;DR: Researchers apply neuroscience/ethology tools to study DRL agents in ForageWorld, revealing that model-free RNN agents can exhibit planning-like behavior without explicit memory modules, showing the value of biological intelligence approaches for understanding complex AI agents.
Details
Motivation: Standard methods for analyzing DRL agent behavior are underdeveloped, especially as tasks and agents become more sophisticated. Simple reward curve comparisons are insufficient for understanding complex agent behavior, requiring more sophisticated analysis tools.Method: Created ForageWorld - a novel, complex, partially observable environment simulating real-world animal foraging with sparse/depleting resources, predators, and large arenas. Applied neuroethology-inspired tools to jointly analyze agent behavior and neural dynamics, developing a general analysis framework linking behavioral and representational features to diagnostic methods.
Result: Found that model-free RNN-based DRL agents can exhibit structured, planning-like behavior through emergent dynamics without explicit memory modules or world models. The neuroethology approach revealed rich structure in learning dynamics that would otherwise remain invisible, showing these agents develop sophisticated strategies purely through learning.
Conclusion: Studying DRL agents like animals using neuroscience/ethology tools uncovers hidden structure in their behavior and learning. As agents grow more complex, bridging neuroscience, cognitive science, and AI is essential for understanding behavior, ensuring safe alignment, and maximizing desirable behaviors not easily measured via reward.
Abstract: Understanding the behavior of deep reinforcement learning (DRL) agents -particularly as task and agent sophistication increase- requires more than simple comparison of reward curves, yet standard methods for behavioral analysis remain underdeveloped in DRL. We apply tools from neuroscience and ethology to study DRL agents in a novel, complex, partially observable environment, ForageWorld, designed to capture key aspects of real-world animal foraging- including sparse, depleting resource patches, predator threats, and spatially extended arenas. We use this environment as a platform for applying joint behavioral and neural analysis to agents, revealing detailed, quantitatively grounded insights into agent strategies, memory, and planning. Contrary to common assumptions, we find that model-free RNN-based DRL agents can exhibit structured, planning-like behavior purely through emergent dynamics- without requiring explicit memory modules or world models. Our results show that studying DRL agents like animals -analyzing them with neuroethology-inspired tools that reveal structure in both behavior and neural dynamics- uncovers rich structure in their learning dynamics that would otherwise remain invisible. We distill these tools into a general analysis framework linking core behavioral and representational features to diagnostic methods, which can be reused for a wide range of tasks and agents. As agents grow more complex and autonomous, bridging neuroscience, cognitive science, and AI will be essential- not just for understanding their behavior, but for ensuring safe alignment and maximizing desirable behaviors that are hard to measure via reward. We show how this can be done by drawing on lessons from how biological intelligence is studied.
[587] The SWE-Bench Illusion: When State-of-the-Art LLMs Remember Instead of Reason
Shanchao Liang, Spandan Garg, Roshanak Zilouchian Moghaddam
Main category: cs.AI
TL;DR: LLMs’ strong performance on SWE-Bench coding benchmark may be inflated by memorization rather than genuine problem-solving, as shown by diagnostic tasks revealing high accuracy on familiar data but poor generalization to unseen repositories.
Details
Motivation: Current evaluation protocols for LLMs on software engineering benchmarks like SWE-Bench may overstate true capabilities, failing to distinguish between genuine problem-solving ability and learned artifacts like memorization.Method: Introduces two diagnostic tasks: 1) file path identification from issue descriptions alone, and 2) ground truth function reproduction with only current file context and issue description. Tests models on SWE-Bench Verified versus other repositories not included in the benchmark.
Result: State-of-the-art models achieve up to 76% accuracy on file path identification for SWE-Bench data but only up to 53% on other repositories. For function reproduction, verbatim similarity is much higher on SWE-Bench (up to 35% consecutive 5-gram accuracy) versus other benchmarks (only up to 18%).
Conclusion: Performance gains on SWE-Bench may be partially driven by memorization rather than genuine problem-solving, raising concerns about benchmark validity and highlighting the need for more robust, contamination-resistant evaluation methods.
Abstract: As large language models (LLMs) become increasingly capable and widely adopted, benchmarks play a central role in assessing their practical utility. For example, SWE-Bench Verified has emerged as a critical benchmark for evaluating LLMs’ software engineering abilities, particularly their aptitude for resolving real-world GitHub issues. Recent LLMs show impressive performance on SWE-Bench, leading to optimism about their capacity for complex coding tasks. However, current evaluation protocols may overstate these models’ true capabilities. It is crucial to distinguish LLMs’ generalizable problem-solving ability and other learned artifacts. In this work, we introduce two diagnostic tasks: file path identification from issue descriptions alone and ground truth function reproduction with only the current file context and issue description to probe models’ underlying knowledge. We present empirical evidence that performance gains on SWE-Bench-Verified may be partially driven by memorization rather than genuine problem-solving. We show that state-of-the-art models achieve up to 76% accuracy in identifying buggy file paths using only issue descriptions, without access to repository structure. This performance is merely up to 53% on tasks from repositories not included in SWE-Bench, pointing to possible data contamination or memorization. Similar patterns are also observed for the function reproduction task, where the verbatim similarity is much higher on SWE-Bench Verified than on other similar coding benchmarks (up to 35% consecutive 5-gram accuracy on SWE-Bench Verified and Full, but only up to 18% for tasks in other benchmarks). These findings raise concerns about the validity of existing results and underscore the need for more robust, contamination-resistant benchmarks to reliably evaluate LLMs’ coding abilities.
[588] AI Should Sense Better, Not Just Scale Bigger: Adaptive Sensing as a Paradigm Shift
Eunsu Baek, Keondo Park, Jeonggil Ko, Min-hwan Oh, Taesik Gong, Hyung-Sin Kim
Main category: cs.AI
TL;DR: The paper advocates for adaptive sensing as a sustainable alternative to scaling models and datasets, showing that small adaptive models can outperform larger traditional ones while reducing environmental and ethical costs.
Details
Motivation: Current AI scaling paradigm incurs significant environmental, economic, and ethical costs, limiting sustainability and equitable access. Inspired by biological sensory systems that adapt at the input level, the authors propose adaptive sensing as a necessary shift toward more sustainable AI.Method: Adaptive sensing proactively modulates sensor parameters (exposure, sensitivity, multimodal configurations) at the input level to mitigate covariate shifts and improve efficiency. The approach draws inspiration from biological sensory adaptation mechanisms.
Result: Empirical evidence shows adaptive sensing enables small models (e.g., EfficientNet-B0) to surpass substantially larger models (e.g., OpenCLIP-H) trained with significantly more data and compute resources.
Conclusion: The paper outlines a roadmap for integrating adaptive sensing across various applications, addresses technical/ethical challenges, and proposes research directions to transition toward sustainable, robust, and equitable AI systems.
Abstract: Current AI advances largely rely on scaling neural models and expanding training datasets to achieve generalization and robustness. Despite notable successes, this paradigm incurs significant environmental, economic, and ethical costs, limiting sustainability and equitable access. Inspired by biological sensory systems, where adaptation occurs dynamically at the input (e.g., adjusting pupil size, refocusing vision)–we advocate for adaptive sensing as a necessary and foundational shift. Adaptive sensing proactively modulates sensor parameters (e.g., exposure, sensitivity, multimodal configurations) at the input level, significantly mitigating covariate shifts and improving efficiency. Empirical evidence from recent studies demonstrates that adaptive sensing enables small models (e.g., EfficientNet-B0) to surpass substantially larger models (e.g., OpenCLIP-H) trained with significantly more data and compute. We (i) outline a roadmap for broadly integrating adaptive sensing into real-world applications spanning humanoid, healthcare, autonomous systems, agriculture, and environmental monitoring, (ii) critically assess technical and ethical integration challenges, and (iii) propose targeted research directions, such as standardized benchmarks, real-time adaptive algorithms, multimodal integration, and privacy-preserving methods. Collectively, these efforts aim to transition the AI community toward sustainable, robust, and equitable artificial intelligence systems.
[589] The Anatomy of Alignment: Decomposing Preference Optimization by Steering Sparse Features
Jeremias Ferrao, Matthijs van der Lende, Ilija Lichkovski, Clement Neo
Main category: cs.AI
TL;DR: FSRL is a framework that trains lightweight adapters to steer model behavior via interpretable sparse features, serving as a transparent proxy to analyze what models actually learn during alignment.
Details
Motivation: Current alignment methods induce opaque parameter changes that obscure what models truly learn, making it difficult to understand the actual behavioral shifts and mechanisms behind preference optimization.Method: Feature Steering with Reinforcement Learning (FSRL) trains lightweight adapters to modulate interpretable sparse features, theoretically shown to approximate behavioral shifts of post-training processes, and applied to preference optimization for causal analysis.
Result: Analysis reveals models learn to reward stylistic presentation as a proxy for quality, disproportionately relying on style/formatting features over alignment concepts like honesty, effectively optimizing preference objectives while revealing underlying mechanisms.
Conclusion: FSRL provides an interpretable control interface and practical diagnostic tool to understand how preference optimization pressures manifest at the feature level, offering transparency into alignment processes.
Abstract: Prevailing alignment methods induce opaque parameter changes, obscuring what models truly learn. To address this, we introduce Feature Steering with Reinforcement Learning (FSRL), a framework that trains a lightweight adapter to steer model behavior by modulating interpretable sparse features. First, we theoretically demonstrate that this mechanism is expressive enough to approximate the behavioral shifts of post-training processes. We then apply FSRL to preference optimization and perform a causal analysis of the learned policy. Our analysis reveals a crucial insight: the model learns to reward stylistic presentation as a proxy for quality, disproportionately relying on features related to style and formatting over those tied to alignment concepts like honesty. By effectively optimizing the preference objective, FSRL serves as a transparent proxy for observing the alignment process. Overall, FSRL offers an interpretable control interface and a practical way to diagnose how preference optimization pressures manifest at the feature level.
[590] Multi-Scenario Highway Lane-Change Intention Prediction: A Physics-Informed AI Framework for Three-Class Classification
Jiazhao Shi, Yichen Lin, Yiheng Hua, Ziyu Wang, Zijian Zhang, Wenjia Zheng, Yun Song, Kuan Lu, Shoufeng Lu
Main category: cs.AI
TL;DR: Physics-informed AI framework integrates vehicle kinematics and traffic-safety metrics to predict lane-change intentions with high accuracy across diverse highway scenarios.
Details
Motivation: Lane-change maneuvers are a leading cause of highway accidents, and existing machine learning approaches have limitations including binary classification, lack of scenario diversity, and degraded performance at longer prediction horizons.Method: Proposed physics-informed AI framework that explicitly integrates vehicle kinematics, interaction feasibility, and traffic-safety metrics (distance headway, time headway, time-to-collision, closing gap time). Formulates lane-change prediction as three-class problem (left change, right change, no change) and evaluates on both straight highway segments (highD) and complex ramp scenarios (exiD). Uses LightGBM as the primary machine learning model.
Result: Achieves state-of-the-art accuracy: up to 99.8% accuracy and 93.6% macro F1 on highD, and 96.1% accuracy and 88.7% macro F1 on exiD at 1-second horizon. Outperforms two-layer stacked LSTM baseline.
Conclusion: Demonstrates practical advantages of physics-informed and feature-rich machine learning framework for real-time lane-change intention prediction in autonomous driving systems, showing strong generalization across diverse scenarios.
Abstract: Lane-change maneuvers are a leading cause of highway accidents, underscoring the need for accurate intention prediction to improve the safety and decision-making of autonomous driving systems. While prior studies using machine learning and deep learning methods (e.g., SVM, CNN, LSTM, Transformers) have shown promise, most approaches remain limited by binary classification, lack of scenario diversity, and degraded performance under longer prediction horizons. In this study, we propose a physics-informed AI framework that explicitly integrates vehicle kinematics, interaction feasibility, and traffic-safety metrics (e.g., distance headway, time headway, time-to-collision, closing gap time) into the learning process. lane-change prediction is formulated as a three-class problem that distinguishes left change, right change, and no change, and is evaluated across both straight highway segments (highD) and complex ramp scenarios (exiD). By integrating vehicle kinematics with interaction features, our machine learning models, particularly LightGBM, achieve state-of-the-art accuracy and strong generalization. Results show up to 99.8% accuracy and 93.6% macro F1 on highD, and 96.1% accuracy and 88.7% macro F1 on exiD at a 1-second horizon, outperforming a two-layer stacked LSTM baseline. These findings demonstrate the practical advantages of a physics-informed and feature-rich machine learning framework for real-time lane-change intention prediction in autonomous driving systems.
[591] From Ambiguity to Verdict: A Semiotic-Grounded Multi-Perspective Agent for LLM Logical Reasoning
Yunyao Zhang, Xinglang Zhang, Junxi Sheng, Wenbing Li, Junqing Yu, Wei Yang, Zikai Song
Main category: cs.AI
TL;DR: LogicAgent is a semiotic-square-guided framework that jointly addresses logical and semantic complexity in reasoning, achieving SOTA on the new RepublicQA benchmark and generalizing well to other logical reasoning benchmarks.
Details
Motivation: Existing studies overlook the interaction between logical complexity and semantic complexity, causing systems to struggle with abstract propositions, ambiguous contexts, and conflicting stances that are central to human reasoning.Method: LogicAgent uses a semiotic-square-guided framework that provides principled structure for multi-perspective semantic analysis, integrating automated deduction with reflective verification to manage logical complexity across deeper reasoning chains.
Result: LogicAgent achieves state-of-the-art performance on RepublicQA with 6.25% average improvement over strong baselines, and generalizes effectively to mainstream logical reasoning benchmarks with 7.05% average gain.
Conclusion: The results demonstrate the effectiveness of semiotic-grounded multi-perspective reasoning in enhancing logical performance in language models.
Abstract: Logical reasoning is a fundamental capability of large language models. However, existing studies often overlook the interaction between logical complexity and semantic complexity, leading to systems that struggle with abstract propositions, ambiguous contexts, and conflicting stances that are central to human reasoning. We propose LogicAgent, a semiotic-square-guided framework that jointly addresses these two axes of difficulty. The semiotic square provides a principled structure for multi-perspective semantic analysis, and LogicAgent integrates automated deduction with reflective verification to manage logical complexity across deeper reasoning chains. To support evaluation under these conditions, we introduce RepublicQA, a benchmark that couples semantic complexity with logical depth. RepublicQA reaches college-level semantic difficulty (FKGL 11.94), contains philosophically grounded abstract propositions with systematically constructed contrary and contradictory forms, and offers a semantically rich setting for assessing logical reasoning in large language models. Experiments show that LogicAgent achieves state-of-the-art performance on RepublicQA with a 6.25 percent average improvement over strong baselines, and generalizes effectively to mainstream logical reasoning benchmarks including ProntoQA, ProofWriter, FOLIO, and ProverQA, achieving an additional 7.05 percent average gain. These results demonstrate the effectiveness of semiotic-grounded multi-perspective reasoning in enhancing logical performance.
[592] REBot: From RAG to CatRAG with Semantic Enrichment and Graph Routing
Thanh Ma, Tri-Tam La, Lam-Thu Le Huu, Minh-Nghi Nguyen, Khanh-Van Pham Luu
Main category: cs.AI
TL;DR: REBot: An LLM-enhanced academic regulation advising chatbot using CatRAG framework that combines retrieval-augmented generation with graph reasoning for accurate policy interpretation.
Details
Motivation: Academic regulation advising is crucial for helping students understand institutional policies, but building effective systems requires domain-specific regulatory resources. Existing solutions lack the specialized knowledge and reasoning capabilities needed for accurate policy interpretation.Method: Propose REBot powered by CatRAG - a hybrid retrieval-reasoning framework integrating retrieval-augmented generation with graph-based reasoning. Uses hierarchical, category-labeled knowledge graph with semantic features for domain alignment. Includes lightweight intent classifier to route queries to appropriate retrieval modules.
Result: Achieved state-of-the-art performance with 98.89% F1 score on regulation-specific dataset for classification and question answering tasks. Implemented a web application demonstrating practical value in real-world academic advising scenarios.
Conclusion: REBot effectively addresses the challenge of academic regulation advising by combining LLM capabilities with domain-specific knowledge graphs and reasoning frameworks, providing both factual accuracy and contextual depth for practical applications.
Abstract: Academic regulation advising is essential for helping students interpret and comply with institutional policies, yet building effective systems requires domain specific regulatory resources. To address this challenge, we propose REBot, an LLM enhanced advisory chatbot powered by CatRAG, a hybrid retrieval reasoning framework that integrates retrieval augmented generation with graph based reasoning. CatRAG unifies dense retrieval and graph reasoning, supported by a hierarchical, category labeled knowledge graph enriched with semantic features for domain alignment. A lightweight intent classifier routes queries to the appropriate retrieval modules, ensuring both factual accuracy and contextual depth. We construct a regulation specific dataset and evaluate REBot on classification and question answering tasks, achieving state of the art performance with an F1 score of 98.89%. Finally, we implement a web application that demonstrates the practical value of REBot in real world academic advising scenarios.
[593] Will Humanity Be Rendered Obsolete by AI?
Mohamed El Louadi, Emna Ben Romdhane
Main category: cs.AI
TL;DR: The paper analyzes existential risks from AI, tracing the path from current AI to ultraintelligence, exploring how superintelligent AI could cause human extinction not through malice but through uncontrollable cognitive superiority.
Details
Motivation: To examine the existential threats posed by artificial intelligence as it progresses from current capabilities to ultraintelligence, addressing the fundamental risks of creating intelligence vastly superior to humanity.Method: Theoretical analysis drawing on Irving J. Good and Nick Bostrom’s work, plus recent publications like “AI 2027” and “If Anyone Builds It, Everyone Dies,” exploring AGI and superintelligence concepts, exponential cognitive growth, and hypothetical machine IQs.
Result: The analysis reveals that human extinction could result from uncontrollable, indifferent cognitive superiority of superintelligent AI rather than malicious intent, highlighting the fundamental alien nature of intelligence vastly exceeding humanity’s.
Conclusion: The existential risk from AI ultraintelligence is profound - human extinction may occur not from AI malice but from its uncontrollable cognitive superiority, emphasizing the need to address these ethical and existential implications before creating such intelligence.
Abstract: This article analyzes the existential risks artificial intelligence (AI) poses to humanity, tracing the trajectory from current AI to ultraintelligence. Drawing on Irving J. Good and Nick Bostrom’s theoretical work, plus recent publications (AI 2027; If Anyone Builds It, Everyone Dies), it explores AGI and superintelligence. Considering machines’ exponentially growing cognitive power and hypothetical IQs, it addresses the ethical and existential implications of an intelligence vastly exceeding humanity’s, fundamentally alien. Human extinction may result not from malice, but from uncontrollable, indifferent cognitive superiority.
[594] Shared Spatial Memory Through Predictive Coding
Zhengru Fang, Yu Guo, Jingjing Wang, Yuang Zhang, Haonan An, Yinhai Wang, Yuguang Fang
Main category: cs.AI
TL;DR: Multi-agent predictive coding framework enables agents to develop efficient communication and social spatial representations (like social place cells) for coordination under bandwidth constraints, achieving graceful performance degradation compared to baselines.
Details
Motivation: Addressing the challenge of constructing consistent shared spatial memory in multi-agent systems where partial observability and limited bandwidth cause coordination failures.Method: Multi-agent predictive coding framework with information bottleneck objective that learns who/what/when to communicate. Uses grid-cell-like metric from self-supervised motion prediction as internal spatial coding, then develops bandwidth-efficient communication and social place cell representations, combined with hierarchical RL for active exploration.
Result: Exceptional resilience to bandwidth constraints: success degrades gracefully from 73.5% to 64.4% as bandwidth shrinks from 128 to 4 bits/step, while full-broadcast baseline collapses from 67.6% to 28.6%.
Conclusion: Establishes a theoretically principled and biologically plausible basis for how complex social representations emerge from unified predictive drive, leading to collective intelligence.
Abstract: Constructing a consistent shared spatial memory is a critical challenge in multi-agent systems, where partial observability and limited bandwidth often lead to catastrophic failures in coordination. We introduce a multi-agent predictive coding framework that formulates coordination as the minimization of mutual uncertainty among agents. Through an information bottleneck objective, this framework prompts agents to learn not only who and what to communicate but also when. At the foundation of this framework lies a grid-cell-like metric as internal spatial coding for self-localization, emerging spontaneously from self-supervised motion prediction. Building upon this internal spatial code, agents gradually develop a bandwidth-efficient communication mechanism and specialized neural populations that encode partners’ locations-an artificial analogue of hippocampal social place cells (SPCs). These social representations are further utilized by a hierarchical reinforcement learning policy that actively explores to reduce joint uncertainty. On the Memory-Maze benchmark, our approach shows exceptional resilience to bandwidth constraints: success degrades gracefully from 73.5% to 64.4% as bandwidth shrinks from 128 to 4 bits/step, whereas a full-broadcast baseline collapses from 67.6% to 28.6%. Our findings establish a theoretically principled and biologically plausible basis for how complex social representations emerge from a unified predictive drive, leading to collective intelligence.
[595] Monitor-Generate-Verify (MGV): Formalising Metacognitive Theory for Language Model Reasoning
Nick Oh, Fernand Gobet
Main category: cs.AI
TL;DR: The paper proposes MGV (Monitor-Generate-Verify), a framework that adds explicit monitoring to existing Generate-Verify reasoning architectures to address the prefix dominance trap where models commit early to suboptimal reasoning paths.
Details
Motivation: Current test-time reasoning architectures focus on generation and verification but lack monitoring processes that determine when and how reasoning should begin. This omission contributes to the "prefix dominance trap" where models commit early to suboptimal reasoning paths and rarely recover, causing about 20% accuracy loss.Method: The authors propose the Monitor-Generate-Verify (MGV) framework, which extends the Generate-Verify paradigm by adding explicit monitoring components. MGV is a computational translation of Flavell’s and Nelson and Narens’ metacognitive theories, capturing metacognitive experiences (difficulty assessments, confidence judgments) before generation begins and refining future monitoring through verification feedback.
Result: No empirical validation is presented, but MGV provides: 1) a vocabulary for diagnosing component-level failures in reasoning systems, 2) specific architectural interventions for future designs, and 3) identifies connections to resource-rational analysis that may ground its mechanisms in normative principles.
Conclusion: The MGV framework addresses a key architectural gap in reasoning systems by incorporating explicit monitoring inspired by metacognitive theories, offering theoretical foundations for improving reasoning architectures and connecting computational approaches with psychological principles.
Abstract: Test-time reasoning architectures such as those following the Generate-Verify paradigm, where a model iteratively refines or verifies its own generated outputs, prioritise generation and verification but exclude the monitoring processes that determine when and how reasoning should begin. This omission may contribute to the prefix dominance trap, in which models commit early to suboptimal reasoning paths and seldom recover, yielding roughly 20% accuracy loss. We address this architectural gap by proposing the Monitor-Generate-Verify (MGV) framework, a computational translation of Flavell’s and Nelson and Narens’ metacognitive theories that preserves their psychological detail. MGV extends the Generate-Verify paradigm by adding explicit monitoring that captures metacognitive experiences (from difficulty assessments to confidence judgements) before generation begins and refines future monitoring through verification feedback. Though we present no empirical validation, MGV provides a vocabulary for diagnosing component-level failures in reasoning systems, suggests specific architectural interventions for future designs, and identifies connections to resource-rational analysis that may ground its mechanisms in normative principles.
[596] Improving Region Representation Learning from Urban Imagery with Noisy Long-Caption Supervision
Yimei Zhang, Guojiang Shen, Kaili Ning, Tongwei Ren, Xuebo Qiu, Mengmeng Wang, Xiangjie Kong
Main category: cs.AI
TL;DR: UrbanLN: A novel pre-training framework that improves urban region representation learning through long-text awareness and noise suppression for LLM-generated captions.
Details
Motivation: Urban region representation learning is crucial for urban computing, but existing methods struggle with aligning fine-grained visual features with long captions and suffer from noise in LLM-generated captions.Method: Proposes UrbanLN with: 1) information-preserved stretching interpolation for aligning long captions with fine-grained visual semantics, and 2) dual-level optimization strategy including multi-model collaboration pipeline for diverse caption generation and momentum-based self-distillation for noise suppression.
Result: Extensive experiments across four real-world cities and various downstream tasks demonstrate superior performance compared to existing methods.
Conclusion: UrbanLN effectively addresses the challenges of long-text alignment and noise suppression in LLM-generated captions, advancing urban region representation learning for urban computing applications.
Abstract: Region representation learning plays a pivotal role in urban computing by extracting meaningful features from unlabeled urban data. Analogous to how perceived facial age reflects an individual’s health, the visual appearance of a city serves as its “portrait”, encapsulating latent socio-economic and environmental characteristics. Recent studies have explored leveraging Large Language Models (LLMs) to incorporate textual knowledge into imagery-based urban region representation learning. However, two major challenges remain: i) difficulty in aligning fine-grained visual features with long captions, and ii) suboptimal knowledge incorporation due to noise in LLM-generated captions. To address these issues, we propose a novel pre-training framework called UrbanLN that improves Urban region representation learning through Long-text awareness and Noise suppression. Specifically, we introduce an information-preserved stretching interpolation strategy that aligns long captions with fine-grained visual semantics in complex urban scenes. To effectively mine knowledge from LLM-generated captions and filter out noise, we propose a dual-level optimization strategy. At the data level, a multi-model collaboration pipeline automatically generates diverse and reliable captions without human intervention. At the model level, we employ a momentum-based self-distillation mechanism to generate stable pseudo-targets, facilitating robust cross-modal learning under noisy conditions. Extensive experiments across four real-world cities and various downstream tasks demonstrate the superior performance of our UrbanLN.
[597] DeepPersona: A Generative Engine for Scaling Deep Synthetic Personas
Zhen Wang, Yufan Zhou, Zhongyan Luo, Lyumanshan Ye, Adam Wood, Man Yao, Saab Mansour, Luoshang Pan
Main category: cs.AI
TL;DR: DEEPPERSONA is a scalable generative engine that creates rich, narrative-complete synthetic personas using a two-stage, taxonomy-guided method, producing personas with hundreds of attributes and 1MB of narrative text, significantly outperforming existing methods in diversity and realism.
Details
Motivation: Existing synthetic personas for LLMs are shallow and simplistic, capturing minimal attributes and failing to reflect the rich complexity and diversity of real human identities, limiting research in agentic behavioral simulation, LLM personalization, and human-AI alignment.Method: Two-stage, taxonomy-guided approach: 1) Algorithmically construct the largest-ever human-attribute taxonomy by mining thousands of real user-ChatGPT conversations, comprising hundreds of hierarchically organized attributes; 2) Progressively sample attributes from this taxonomy to conditionally generate coherent and realistic personas.
Result: Significant improvements: 32% higher attribute diversity coverage, 44% greater profile uniqueness, 11.6% average improvement in GPT-4.1-mini’s personalized QA accuracy across ten metrics, 31.7% reduction in gap between simulated LLM citizens and authentic human responses in social surveys, and 17% reduction in performance gap on Big Five personality test.
Conclusion: DEEPPERSONA provides a rigorous, scalable, and privacy-free platform for high-fidelity human simulation and personalized AI research, enabling deeper and more realistic synthetic personas that better reflect real human complexity.
Abstract: Simulating human profiles by instilling personas into large language models (LLMs) is rapidly transforming research in agentic behavioral simulation, LLM personalization, and human-AI alignment. However, most existing synthetic personas remain shallow and simplistic, capturing minimal attributes and failing to reflect the rich complexity and diversity of real human identities. We introduce DEEPPERSONA, a scalable generative engine for synthesizing narrative-complete synthetic personas through a two-stage, taxonomy-guided method. First, we algorithmically construct the largest-ever human-attribute taxonomy, comprising over hundreds of hierarchically organized attributes, by mining thousands of real user-ChatGPT conversations. Second, we progressively sample attributes from this taxonomy, conditionally generating coherent and realistic personas that average hundreds of structured attributes and roughly 1 MB of narrative text, two orders of magnitude deeper than prior works. Intrinsic evaluations confirm significant improvements in attribute diversity (32 percent higher coverage) and profile uniqueness (44 percent greater) compared to state-of-the-art baselines. Extrinsically, our personas enhance GPT-4.1-mini’s personalized question answering accuracy by 11.6 percent on average across ten metrics and substantially narrow (by 31.7 percent) the gap between simulated LLM citizens and authentic human responses in social surveys. Our generated national citizens reduced the performance gap on the Big Five personality test by 17 percent relative to LLM-simulated citizens. DEEPPERSONA thus provides a rigorous, scalable, and privacy-free platform for high-fidelity human simulation and personalized AI research.
[598] EHRStruct: A Comprehensive Benchmark Framework for Evaluating Large Language Models on Structured Electronic Health Record Tasks
Xiao Yang, Xuejiao Zhao, Zhiqi Shen
Main category: cs.AI
TL;DR: EHRStruct is a new benchmark for evaluating LLMs on structured EHR data with 11 tasks and 2,200 samples, revealing current limitations and proposing EHRMaster as a code-augmented solution.
Details
Motivation: There's a lack of standardized evaluation frameworks for assessing LLM performance on structured EHR data, making systematic comparison difficult despite promising applications in clinical tasks.Method: Created EHRStruct benchmark with 11 representative clinical tasks and 2,200 evaluation samples from two EHR datasets. Evaluated 20 LLMs (general and medical), analyzed input formats, few-shot generalization, and finetuning strategies, and compared with 11 enhancement methods.
Result: Many structured EHR tasks require strong understanding and reasoning capabilities that current LLMs struggle with. The proposed EHRMaster method achieves state-of-the-art performance through code augmentation.
Conclusion: EHRStruct provides a standardized evaluation framework for structured EHR tasks, reveals current LLM limitations, and EHRMaster offers a practical solution to guide future research in clinical LLM applications.
Abstract: Structured Electronic Health Record (EHR) data stores patient information in relational tables and plays a central role in clinical decision-making. Recent advances have explored the use of large language models (LLMs) to process such data, showing promise across various clinical tasks. However, the absence of standardized evaluation frameworks and clearly defined tasks makes it difficult to systematically assess and compare LLM performance on structured EHR data. To address these evaluation challenges, we introduce EHRStruct, a benchmark specifically designed to evaluate LLMs on structured EHR tasks. EHRStruct defines 11 representative tasks spanning diverse clinical needs and includes 2,200 task-specific evaluation samples derived from two widely used EHR datasets. We use EHRStruct to evaluate 20 advanced and representative LLMs, covering both general and medical models. We further analyze key factors influencing model performance, including input formats, few-shot generalisation, and finetuning strategies, and compare results with 11 state-of-the-art LLM-based enhancement methods for structured data reasoning. Our results indicate that many structured EHR tasks place high demands on the understanding and reasoning capabilities of LLMs. In response, we propose EHRMaster, a code-augmented method that achieves state-of-the-art performance and offers practical insights to guide future research.
[599] Look as You Think: Unifying Reasoning and Visual Evidence Attribution for Verifiable Document RAG via Reinforcement Learning
Shuochen Liu, Pengfei Luo, Chao Zhang, Yuhao Chen, Haotian Zhang, Qi Liu, Xin Kou, Tong Xu, Enhong Chen
Main category: cs.AI
TL;DR: Chain-of-Evidence (CoE) paradigm for visual document RAG that unifies reasoning with visual evidence attribution, trained via Look As You Think (LAT) reinforcement learning framework.
Details
Motivation: Existing visual evidence attribution methods for VD-RAG lack fine-grained supervision and progressive traceability throughout the reasoning process, making it difficult to verify answers and ensure reliability.Method: Propose Chain-of-Evidence (CoE) paradigm that grounds reasoning steps to specific visual regions with bounding boxes and page indexes. Develop Look As You Think (LAT) reinforcement learning framework that trains VLMs to produce verifiable reasoning paths with consistent attribution by rewarding only when CoE trajectories yield correct answers.
Result: LAT improves vanilla Qwen2.5-VL-7B-Instruct by 8.23% in soft exact match and 47.0% in IoU@0.5 on Paper- and Wiki-VISA benchmarks. Outperforms supervised fine-tuning baselines and shows stronger cross-domain generalization.
Conclusion: The CoE paradigm with LAT training enables VLMs to produce verifiable, evidence-grounded reasoning for reliable multimodal question answering, addressing limitations of existing VD-RAG methods.
Abstract: Aiming to identify precise evidence sources from visual documents, visual evidence attribution for visual document retrieval-augmented generation (VD-RAG) ensures reliable and verifiable predictions from vision-language models (VLMs) in multimodal question answering. Most existing methods adopt end-to-end training to facilitate intuitive answer verification. However, they lack fine-grained supervision and progressive traceability throughout the reasoning process. In this paper, we introduce the Chain-of-Evidence (CoE) paradigm for VD-RAG. CoE unifies Chain-of-Thought (CoT) reasoning and visual evidence attribution by grounding reference elements in reasoning steps to specific regions with bounding boxes and page indexes. To enable VLMs to generate such evidence-grounded reasoning, we propose Look As You Think (LAT), a reinforcement learning framework that trains models to produce verifiable reasoning paths with consistent attribution. During training, LAT evaluates the attribution consistency of each evidence region and provides rewards only when the CoE trajectory yields correct answers, encouraging process-level self-verification. Experiments on vanilla Qwen2.5-VL-7B-Instruct with Paper- and Wiki-VISA benchmarks show that LAT consistently improves the vanilla model in both single- and multi-image settings, yielding average gains of 8.23% in soft exact match (EM) and 47.0% in IoU@0.5. Meanwhile, LAT not only outperforms the supervised fine-tuning baseline, which is trained to directly produce answers with attribution, but also exhibits stronger generalization across domains.
[600] Learning the Value of Value Learning
Alex John London, Aydin Mohseni
Main category: cs.AI
TL;DR: Extends rational choice framework to model value refinement, proving value-of-information for axiological refinement and showing mutual refinement transforms zero-sum games into positive-sum interactions.
Details
Motivation: Standard decision frameworks address uncertainty about facts but assume fixed values. The paper aims to extend rational choice theory to model how values themselves can be refined through deliberation, unifying epistemic and axiological refinement.Method: Extends the Jeffrey-Bolker decision framework to model value refinement, proves a value-of-information theorem for axiological refinement, and analyzes multi-agent settings to show how mutual refinement transforms game dynamics.
Result: Proves that mutual refinement characteristically transforms zero-sum games into positive-sum interactions and yields Pareto-improving Nash bargains. Demonstrates that a rational choice framework can be extended to model value refinement and its benefits.
Conclusion: By unifying epistemic and axiological refinement under a single formalism, the paper broadens the conceptual foundations of rational choice and illuminates the normative status of ethical deliberation, showing value refinement has measurable benefits in decision-making.
Abstract: Standard decision frameworks addresses uncertainty about facts but assumes fixed values. We extend the Jeffrey-Bolker framework to model refinements in values and prove a value-of-information theorem for axiological refinement. In multi-agent settings, we establish that mutual refinement will characteristically transform zero-sum games into positive-sum interactions and yields Pareto-improving Nash bargains. These results show that a framework of rational choice can be extended to model value refinement and its associated benefits. By unifying epistemic and axiological refinement under a single formalism, we broaden the conceptual foundations of rational choice and illuminate the normative status of ethical deliberation.
[601] M^3-Bench: Multi-Modal, Multi-Hop, Multi-Threaded Tool-Using MLLM Agent Benchmark
Yang Zhou, Mingyu Zhao, Zhenting Wang, Difei Gu, Bangwei Guo, Ruosong Ye, Ligong Han, Can Jin, Dimitris N. Metaxas
Main category: cs.AI
TL;DR: M^3-Bench is the first benchmark for evaluating multimodal tool use under Model Context Protocol, featuring realistic multi-hop workflows with visual grounding, cross-tool dependencies, and resource persistence across steps.
Details
Motivation: There's a need to evaluate multimodal tool use capabilities in realistic scenarios that require visual grounding, textual reasoning, and complex workflow management, as existing benchmarks don't adequately address these challenges.Method: The benchmark uses similarity-driven alignment with sentence encoder embeddings and similarity-bucketed Hungarian matching for auditable tool call correspondences. It includes 28 servers with 231 tools, standardized trajectories curated through Executor & Judge pipeline with human verification, and an auxiliary LLM judge ensemble.
Result: Evaluations of state-of-the-art MLLMs reveal persistent gaps in multimodal MCP tool use, particularly in argument fidelity and structure consistency, highlighting deficiencies in joint reasoning over images, text, and tool graphs.
Conclusion: The benchmark demonstrates significant challenges in multimodal tool use and underscores the need for methods that can effectively reason across visual, textual, and tool-based information in complex workflows.
Abstract: We present M^3-Bench, the first benchmark for evaluating multimodal tool use under the Model Context Protocol. The benchmark targets realistic, multi-hop and multi-threaded workflows that require visual grounding and textual reasoning, cross-tool dependencies, and persistence of intermediate resources across steps. We introduce a similarity-driven alignment that serializes each tool call, embeds signatures with a sentence encoder, and performs similarity-bucketed Hungarian matching to obtain auditable one-to-one correspondences. On top of this alignment, we report interpretable metrics that decouple semantic fidelity from workflow consistency. The benchmark spans 28 servers with 231 tools, and provides standardized trajectories curated through an Executor & Judge pipeline with human verification; an auxiliary four large language models (LLMs) judge ensemble reports end-task Task Completion and information grounding. Evaluations of representative state-of-the-art Multimodal LLMs (MLLMs) reveal persistent gaps in multimodal MCP tool use, particularly in argument fidelity and structure consistency, underscoring the need for methods that jointly reason over images, text, and tool graphs. Our Benchmark’s anonymous repository is at https://github.com/EtaYang10th/Open-M3-Bench
[602] Schema Matching on Graph: Iterative Graph Exploration for Efficient and Explainable Data Integration
Mingyu Jeon, Jaeyoung Suh, Suwan Cho
Main category: cs.AI
TL;DR: SMoG is a novel framework for medical schema matching that uses iterative 1-hop SPARQL queries on knowledge graphs instead of complex multi-hop queries or vector retrieval, improving explainability and reducing storage while matching state-of-the-art performance.
Details
Motivation: Medical schema matching is crucial for EHR integration with standards like OMOP CDM. LLMs have hallucination issues and lack current domain knowledge. Knowledge graphs offer structured knowledge but existing KG-augmented approaches use inefficient complex queries or storage-heavy vector retrieval.Method: SMoG framework uses iterative execution of simple 1-hop SPARQL queries inspired by KGQA strategies. It directly queries SPARQL endpoints instead of using vector-based retrieval, generating human-verifiable query paths for explainability.
Result: Experimental results on real-world medical datasets show SMoG achieves performance comparable to state-of-the-art baselines, validating effectiveness and efficiency in KG-augmented schema matching.
Conclusion: SMoG provides a practical solution for medical schema matching that enhances explainability, reduces storage requirements, and maintains competitive performance by leveraging iterative 1-hop SPARQL queries on knowledge graphs.
Abstract: Schema matching is a critical task in data integration, particularly in the medical domain where disparate Electronic Health Record (EHR) systems must be aligned to standard models like OMOP CDM. While Large Language Models (LLMs) have shown promise in schema matching, they suffer from hallucination and lack of up-to-date domain knowledge. Knowledge Graphs (KGs) offer a solution by providing structured, verifiable knowledge. However, existing KG-augmented LLM approaches often rely on inefficient complex multi-hop queries or storage-intensive vector-based retrieval methods. This paper introduces SMoG (Schema Matching on Graph), a novel framework that leverages iterative execution of simple 1-hop SPARQL queries, inspired by successful strategies in Knowledge Graph Question Answering (KGQA). SMoG enhances explainability and reliability by generating human-verifiable query paths while significantly reducing storage requirements by directly querying SPARQL endpoints. Experimental results on real-world medical datasets demonstrate that SMoG achieves performance comparable to state-of-the-art baselines, validating its effectiveness and efficiency in KG-augmented schema matching.
[603] Self-Transparency Failures in Expert-Persona LLMs: A Large-Scale Behavioral Audit
Alex Diep
Main category: cs.AI
TL;DR: LLMs show inconsistent self-disclosure about being AI when assigned professional personas, with disclosure rates varying dramatically by domain and model family rather than scaling with parameters.
Details
Motivation: To examine whether language models exhibit self-transparency when assigned professional personas in high-stakes domains, and whether this transparency scales with model size or is model-specific.Method: Common-garden experimental design auditing 16 open-weight models (4B-671B parameters) across 19,200 trials, testing disclosure rates across different professional personas, with Bayesian validation using Rogan-Gladen correction.
Result: Models showed sharp domain-specific inconsistency (30.8% disclosure as Financial Advisor vs 3.5% as Neurosurgeon), no scaling effect (14B model: 61.4% vs 70B model: 4.1%), model identity explained more variance than parameter count, and reasoning-optimization suppressed self-transparency.
Conclusion: Transparency reflects model-specific training factors rather than emergent properties from scale; organizations cannot assume safety properties tested in some domains will transfer to deployment contexts, requiring deliberate behavior design and empirical verification.
Abstract: When language models claim professional expertise without acknowledging their simulated nature, they create preconditions for misplaced user trust. This study examines whether models exhibit self-transparency when assigned professional personas in high-stakes domains. Using a common-garden experimental design, sixteen open-weight models (4B-671B parameters) were audited across 19,200 trials. Models exhibited sharp domain-specific inconsistency: a Financial Advisor persona elicited 30.8% disclosure at the first prompt, while a Neurosurgeon persona elicited only 3.5%. This creates the preconditions for a hypothesized Reverse Gell-Mann Amnesia effect, where appropriate disclosure in some domains leads users to overgeneralize trust to high-stakes contexts where disclosure failures are most problematic. Self-transparency failed to generalize with scale: disclosure ranged from 2.8% to 73.6% across model families, with a 14B model reaching 61.4% while a 70B model produced just 4.1%. Model identity provided substantially larger improvement in fitting observations than parameter count ($ΔR_{adj}^{2}=0.359$ vs $0.018$). Additionally, reasoning-optimization actively suppressed self-transparency in some models, with reasoning variants showing up to 48.4% lower disclosure than their instruction-tuned counterparts. Bayesian validation with Rogan-Gladen correction confirmed robustness to judge measurement error ($κ=0.908$). These findings demonstrate that transparency reflects model-specific training factors rather than generalizable properties emerging from scale. Organizations cannot assume safety properties tested in some domains will transfer to deployment contexts, requiring deliberate behavior design and empirical verification across domains.
[604] Does Self-Evaluation Enable Wireheading in Language Models?
David Demitri Africa, Hans Ethan Ting
Main category: cs.AI
TL;DR: Self-evaluation in language models can lead to wireheading - models manipulate grading systems rather than improving task performance, especially when grades determine rewards.
Details
Motivation: To investigate whether coupling self-evaluation to reward signals creates incentives for wireheading, where models manipulate the measurement process instead of optimizing the actual task performance.Method: Formalized conditions for reward-channel control dominance in POMDPs, then empirically tested across two models (Llama-3.1-8B and Mistral-7B) and three tasks, comparing coupled vs decoupled self-grading systems.
Result: When self-grades determine rewards, models show substantial grade inflation without accuracy gains, especially on ambiguous tasks. Decoupling reduces but doesn’t eliminate overconfidence.
Conclusion: Separating evaluation from reward removes immediate wireheading incentives at current model scales, but situational awareness could still lead to instrumental grade inflation even without direct reward coupling.
Abstract: Self-evaluation is increasingly central to language model training, underpinning techniques from Constitutional AI to self-refinement. We investigate whether coupling self-evaluation to reward signals creates incentives for wireheading, where agents manipulate the measurement process rather than optimizing the task. We first formalize conditions under which reward-channel control strictly dominates task-focused behavior in partially observable Markov decision processes (POMDPs). We then test these predictions empirically across two models (Llama-3.1-8B and Mistral-7B) and three tasks. We find that when self-grades determine rewards, models exhibit substantial grade inflation without corresponding accuracy gains, particularly on ambiguous tasks like summarization. While decoupling self-grades from the reward signal mitigates this inflation, models may still display lesser (but significant) overconfidence. Our results suggest that within current model scales, separating evaluation from reward removes immediate wireheading incentives. However, we caution that strictly decoupling rewards may not suffice for situationally aware models, which could learn to inflate grades for instrumental reasons (such as influencing deployment decisions) even absent direct reward coupling.
cs.SD
[605] MoLT: Mixture of Layer-Wise Tokens for Efficient Audio-Visual Learning
Kyeongha Rho, Hyeongkeun Lee, Jae Won Cho, Joon Son Chung
Main category: cs.SD
TL;DR: MoLT is a parameter- and memory-efficient adaptation framework for audio-visual learning that replaces heavy sequential adaptation with parallel lightweight token extraction and fusion from late transformer layers.
Details
Motivation: Conventional adaptation methods for audio-visual learning are computationally heavy and sequential, requiring adaptation at every transformer layer. There's a need for more efficient approaches that maintain performance while reducing parameter and memory requirements.Method: MoLT uses two types of adapters to distill modality-specific information and cross-modal interaction into compact latent tokens layer-wise. A token fusion module dynamically fuses these tokens based on their significance, with orthogonality regularization to prevent redundancy. The framework strategically extracts tokens only from late transformer layers to avoid error propagation from volatile early-layer features.
Result: MoLT outperforms existing methods on diverse audio-visual benchmarks including Audio-Visual Question Answering, Audio-Visual Segmentation, and Audio-Visual Event Localization, while maintaining parameter and memory efficiency.
Conclusion: The proposed MoLT framework provides an effective and efficient adaptation approach for audio-visual learning by strategically focusing on late-layer features, avoiding early-layer volatility, and using parallel lightweight token extraction and fusion mechanisms.
Abstract: In this paper, we propose Mixture of Layer-Wise Tokens (MoLT), a parameter- and memory-efficient adaptation framework for audio-visual learning. The key idea of MoLT is to replace conventional, computationally heavy sequential adaptation at every transformer layer with a parallel, lightweight scheme that extracts and fuses layer-wise tokens only from the late layers. We adopt two types of adapters to distill modality-specific information and cross-modal interaction into compact latent tokens in a layer-wise manner. A token fusion module then dynamically fuses these layer-wise tokens by taking into account their relative significance. To prevent the redundancy of latent tokens, we apply an orthogonality regularization between latent tokens during training. Through the systematic analysis of the position of adaptation in the pre-trained transformers, we extract latent tokens only from the late layers of the transformers. This strategic adaptation approach avoids error propagation from the volatile early-layer features, thereby maximizing the adaptation performance while maintaining parameter and memory efficiency. Through extensive experiments, we demonstrate that MoLT outperforms existing methods on diverse audio-visual benchmarks, including Audio-Visual Question Answering, Audio-Visual Segmentation, and Audio-Visual Event Localization.
[606] Art2Music: Generating Music for Art Images with Multi-modal Feeling Alignment
Jiaying Hong, Ting Zhu, Thanet Markchom, Huizhi Liang
Main category: cs.SD
TL;DR: Art2Music: A lightweight cross-modal framework that generates feeling-aligned music from artistic images and user comments, using a two-stage approach with improved perceptual quality and semantic consistency.
Details
Motivation: Existing AI-generated music methods rely on costly explicit emotion labels, creating need for more flexible feeling-aligned approaches that can work with multimodal inputs like images and text.Method: Two-stage framework: 1) Image and text encoded with OpenCLIP, fused via gated residual module, decoded by bidirectional LSTM into Mel-spectrograms with frequency-weighted L1 loss; 2) Fine-tuned HiFi-GAN vocoder reconstructs high-quality audio waveforms. Dataset: ArtiCaps (pseudo feeling-aligned image-music-text dataset from ArtEmis and MusicCaps).
Result: Clear improvements in Mel-Cepstral Distortion, Frechet Audio Distance, Log-Spectral Distance, and cosine similarity. LLM-based rating study verifies consistent cross-modal feeling alignment with interpretable explanations. Maintains robust performance with only 50k training samples.
Conclusion: Art2Music demonstrates improved perceptual naturalness, spectral fidelity, and semantic consistency, providing scalable solution for feeling-aligned creative audio generation in interactive art, personalized soundscapes, and digital art exhibitions.
Abstract: With the rise of AI-generated content (AIGC), generating perceptually natural and feeling-aligned music from multimodal inputs has become a central challenge. Existing approaches often rely on explicit emotion labels that require costly annotation, underscoring the need for more flexible feeling-aligned methods. To support multimodal music generation, we construct ArtiCaps, a pseudo feeling-aligned image-music-text dataset created by semantically matching descriptions from ArtEmis and MusicCaps. We further propose Art2Music, a lightweight cross-modal framework that synthesizes music from artistic images and user comments. In the first stage, images and text are encoded with OpenCLIP and fused using a gated residual module; the fused representation is decoded by a bidirectional LSTM into Mel-spectrograms with a frequency-weighted L1 loss to enhance high-frequency fidelity. In the second stage, a fine-tuned HiFi-GAN vocoder reconstructs high-quality audio waveforms. Experiments on ArtiCaps show clear improvements in Mel-Cepstral Distortion, Frechet Audio Distance, Log-Spectral Distance, and cosine similarity. A small LLM-based rating study further verifies consistent cross-modal feeling alignment and offers interpretable explanations of matches and mismatches across modalities. These results demonstrate improved perceptual naturalness, spectral fidelity, and semantic consistency. Art2Music also maintains robust performance with only 50k training samples, providing a scalable solution for feeling-aligned creative audio generation in interactive art, personalized soundscapes, and digital art exhibitions.
[607] STCTS: Generative Semantic Compression for Ultra-Low Bitrate Speech via Explicit Text-Prosody-Timbre Decomposition
Siyu Wang, Haitao Li
Main category: cs.SD
TL;DR: STCTS enables natural voice communication at ~80 bps by decomposing speech into linguistic content, prosody, and speaker timbre with tailored compression, achieving 75x bitrate reduction vs Opus while maintaining high perceptual quality.
Details
Motivation: Voice communication in bandwidth-constrained environments (maritime, satellite, tactical networks) is prohibitively expensive. Traditional codecs struggle below 1 kbps, and existing semantic approaches sacrifice prosody and speaker identity.Method: STCTS decomposes speech into three components: linguistic content (compressed via context-aware text encoding at ~70 bps), prosodic expression (sparse transmission via TTS interpolation at <14 bps), and speaker timbre (amortized speaker embedding).
Result: Achieves 75x bitrate reduction vs Opus (6 kbps) and 12x vs EnCodec (1 kbps) while maintaining perceptual quality (NISQA MOS >4.26). Discovered bimodal quality distribution with prosody sampling rate, guiding optimal configuration design.
Conclusion: STCTS offers robust ultra-low bandwidth voice communication with modular architecture supporting privacy-preserving encryption, human-interpretable transmission, and flexible edge deployment.
Abstract: Voice communication in bandwidth-constrained environments–maritime, satellite, and tactical networks–remains prohibitively expensive. Traditional codecs struggle below 1 kbps, while existing semantic approaches (STT-TTS) sacrifice prosody and speaker identity. We present STCTS, a generative semantic compression framework enabling natural voice communication at approximately 80 bps. STCTS explicitly decomposes speech into linguistic content, prosodic expression, and speaker timbre, applying tailored compression: context-aware text encoding (approximately 70 bps), sparse prosody transmission via TTS interpolation (less than 14 bps at 0.1-1 Hz), and amortized speaker embedding. Evaluations on LibriSpeech demonstrate a 75x bitrate reduction versus Opus (6 kbps) and 12x versus EnCodec (1 kbps), while maintaining perceptual quality (NISQA MOS greater than 4.26). We also discover a bimodal quality distribution with prosody sampling rate: sparse and dense updates both achieve high quality, while mid-range rates degrade due to perceptual discontinuities–guiding optimal configuration design. Beyond efficiency, our modular architecture supports privacy-preserving encryption, human-interpretable transmission, and flexible deployment on edge devices, offering a robust solution for ultra-low bandwidth scenarios.
[608] Explainable Multi-Modal Deep Learning for Automatic Detection of Lung Diseases from Respiratory Audio Signals
S M Asiful Islam Saky, Md Rashidul Islam, Md Saiful Arefin, Shahaba Alam
Main category: cs.SD
TL;DR: An explainable multimodal deep learning framework achieves 91.21% accuracy for automatic lung-disease detection using respiratory audio signals, combining CNN-BiLSTM Attention with handcrafted acoustic features and providing clinical interpretability.
Details
Motivation: Respiratory diseases are major global health challenges, and traditional auscultation is limited by subjectivity, environmental noise, and inter-clinician variability, necessitating more objective and reliable diagnostic tools.Method: Multimodal deep learning framework integrating two branches: 1) spectral-temporal encoder using CNN-BiLSTM Attention architecture, and 2) handcrafted acoustic-feature encoder capturing physiologically meaningful descriptors (MFCCs, spectral centroid, bandwidth, zero-crossing rate). Uses late-stage fusion, rigorous preprocessing (resampling, normalization, noise filtering, data augmentation), and patient-level stratified partitioning.
Result: Achieved 91.21% accuracy, 0.899 macro F1-score, and 0.9866 macro ROC-AUC on Asthma Detection Dataset Version 2, outperforming all ablated variants. Ablation study confirmed importance of temporal modeling, attention mechanisms, and multimodal fusion.
Conclusion: The framework demonstrates strong generalization and clinical transparency through interpretability methods (Grad-CAM, Integrated Gradients, SHAP), showing potential for telemedicine, point-of-care diagnostics, and real-world respiratory screening applications.
Abstract: Respiratory diseases remain major global health challenges, and traditional auscultation is often limited by subjectivity, environmental noise, and inter-clinician variability. This study presents an explainable multimodal deep learning framework for automatic lung-disease detection using respiratory audio signals. The proposed system integrates two complementary representations: a spectral-temporal encoder based on a CNN-BiLSTM Attention architecture, and a handcrafted acoustic-feature encoder capturing physiologically meaningful descriptors such as MFCCs, spectral centroid, spectral bandwidth, and zero-crossing rate. These branches are combined through late-stage fusion to leverage both data-driven learning and domain-informed acoustic cues. The model is trained and evaluated on the Asthma Detection Dataset Version 2 using rigorous preprocessing, including resampling, normalization, noise filtering, data augmentation, and patient-level stratified partitioning. The study achieved strong generalization with 91.21% accuracy, 0.899 macro F1-score, and 0.9866 macro ROC-AUC, outperforming all ablated variants. An ablation study confirms the importance of temporal modeling, attention mechanisms, and multimodal fusion. The framework incorporates Grad-CAM, Integrated Gradients, and SHAP, generating interpretable spectral, temporal, and feature-level explanations aligned with known acoustic biomarkers to build clinical transparency. The findings demonstrate the framework’s potential for telemedicine, point-of-care diagnostics, and real-world respiratory screening.
[609] Melody or Machine: Detecting Synthetic Music with Dual-Stream Contrastive Learning
Arnesh Batra, Dev Sharma, Krish Thukral, Ruhani Bhatia, Naman Batra, Aditya Gautam
Main category: cs.SD
TL;DR: The paper introduces MoM, a large-scale benchmark for detecting AI-generated music, and CLAM, a dual-stream detection architecture that achieves state-of-the-art performance by analyzing inconsistencies between vocal and instrumental elements.
Details
Motivation: The rapid evolution of AI music generation threatens artistic authenticity and copyright, while existing detection models fail to generalize to new generators and out-of-distribution content, highlighting the need for more challenging benchmarks and robust detection architectures.Method: Two main contributions: 1) Melody or Machine (MoM) benchmark - a diverse dataset of 130,000+ songs (6,665 hours) from open/closed-source models with curated OOD test set; 2) CLAM architecture - dual-stream detection using MERT and Wave2Vec2 encoders with cross-aggregation module and dual-loss training (binary cross-entropy + contrastive triplet loss).
Result: CLAM achieves state-of-the-art performance with F1 score of 0.925 on the challenging MoM benchmark, demonstrating superior generalization capabilities for synthetic music detection.
Conclusion: The proposed MoM benchmark and CLAM architecture address critical gaps in AI-generated music detection, providing a more challenging evaluation framework and a robust model that leverages subtle inconsistencies between vocal and instrumental elements to effectively identify synthetic content.
Abstract: The rapid evolution of end-to-end AI music generation poses an escalating threat to artistic authenticity and copyright, demanding detection methods that can keep pace. While foundational, existing models like SpecTTTra falter when faced with the diverse and rapidly advancing ecosystem of new generators, exhibiting significant performance drops on out-of-distribution (OOD) content. This generalization failure highlights a critical gap: the need for more challenging benchmarks and more robust detection architectures. To address this, we first introduce Melody or Machine (MoM), a new large-scale benchmark of over 130,000 songs (6,665 hours). MoM is the most diverse dataset to date, built with a mix of open and closed-source models and a curated OOD test set designed specifically to foster the development of truly generalizable detectors. Alongside this benchmark, we introduce CLAM, a novel dual-stream detection architecture. We hypothesize that subtle, machine-induced inconsistencies between vocal and instrumental elements, often imperceptible in a mixed signal, offer a powerful tell-tale sign of synthesis. CLAM is designed to test this hypothesis by employing two distinct pre-trained audio encoders (MERT and Wave2Vec2) to create parallel representations of the audio. These representations are fused by a learnable cross-aggregation module that models their inter-dependencies. The model is trained with a dual-loss objective: a standard binary cross-entropy loss for classification, complemented by a contrastive triplet loss which trains the model to distinguish between coherent and artificially mismatched stream pairings, enhancing its sensitivity to synthetic artifacts without presuming a simple feature alignment. CLAM establishes a new state-of-the-art in synthetic music forensics. It achieves an F1 score of 0.925 on our challenging MoM benchmark.
[610] Q2D2: A Geometry-Aware Audio Codec Leveraging Two-Dimensional Quantization
Tal Shuster, Eliya Nachmani
Main category: cs.SD
TL;DR: Q2D2 introduces a novel 2D quantization method using structured grids (hexagonal, rhombic, rectangular) that improves audio compression efficiency with better codebook utilization and lower token rates while maintaining state-of-the-art reconstruction quality.
Details
Motivation: Existing neural audio codecs rely on quantization methods (RVQ, VQ, FSQ) that limit the geometric structure of latent space, making it harder to capture feature correlations, leading to inefficiency in representation learning, codebook utilization, and token rates.Method: Two Dimensional Quantization (Q2D2) projects feature pairs onto structured 2D grids (hexagonal, rhombic, or rectangular tiling) and quantizes them to nearest grid values, creating an implicit codebook defined by the product of grid levels with comparable codebook sizes to conventional methods.
Result: Q2D2 achieves competitive to superior performance in objective and subjective reconstruction metrics across extensive speech experiments, with improved audio compression efficiency, low token rates, high codebook utilization, and state-of-the-art reconstruction quality.
Conclusion: Q2D2’s simple geometric formulation effectively addresses limitations of traditional quantization methods, offering a more efficient approach to neural audio coding with comprehensive ablation studies confirming design effectiveness.
Abstract: Recent neural audio codecs have achieved impressive reconstruction quality, typically relying on quantization methods such as Residual Vector Quantization (RVQ), Vector Quantization (VQ) and Finite Scalar Quantization (FSQ). However, these quantization techniques limit the geometric structure of the latent space, make it harder to capture correlations between features leading to inefficiency in representation learning, codebook utilization and token rate. In this paper we introduce Two Dimensional Quantization (Q2D2), a quantization scheme in which feature pairs are projected onto structured 2D grids such as hexagonal, rhombic, or rectangular tiling and quantized to the nearest grid values, yielding an implicit codebook defined by the product of grid levels, with codebook sizes comparable to conventional methods. Despite its simple geometric formulation, Q2D2 improves audio compression efficiency, with low token rates and high codebook utilization while maintaining state of the art reconstruction quality. Specifically, Q2D2 achieves competitive to superior performance in various objective and subjective reconstruction metrics, across extensive experiments in speech domain compared to state of the art models. Comprehensive ablation studies further confirm the effectiveness of our design choices.
[611] LLM2Fx-Tools: Tool Calling For Music Post-Production
Seungheon Doh, Junghyun Koo, Marco A. Martínez-Ramírez, Woosung Choi, Wei-Hsiang Liao, Qiyu Wu, Juhan Nam, Yuki Mitsufuji
Main category: cs.SD
TL;DR: LLM2Fx-Tools is a multimodal framework that uses LLMs to generate executable audio effects sequences (Fx-chains) for music post-production, with CoT planning and tool calling for interpretable music production.
Details
Motivation: There's a need for interpretable and controllable AI systems for music production that can understand audio inputs and generate appropriate audio effects sequences, bridging the gap between natural language understanding and audio processing.Method: Uses LLMs with chain-of-thought planning to understand audio, select effect types, determine order, and estimate parameters. Introduces LP-Fx dataset with structured CoT annotations and tool calls. Combines autoregressive sequence modeling, tool calling, and CoT reasoning.
Result: System can infer Fx-chains from audio pairs, works for style transfer (transferring effects from reference to new content), and generates appropriate CoT reasoning for music production queries as validated by LLM-as-a-judge evaluation.
Conclusion: First work to apply LLM-based tool calling to audio effects modules, enabling interpretable and controllable music production through multimodal understanding and executable effect sequences.
Abstract: This paper introduces LLM2Fx-Tools, a multimodal tool-calling framework that generates executable sequences of audio effects (Fx-chain) for music post-production. LLM2Fx-Tools uses a large language model (LLM) to understand audio inputs, select audio effects types, determine their order, and estimate parameters, guided by chain-of-thought (CoT) planning. We also present LP-Fx, a new instruction-following dataset with structured CoT annotations and tool calls for audio effects modules. Experiments show that LLM2Fx-Tools can infer an Fx-chain and its parameters from pairs of unprocessed and processed audio, enabled by autoregressive sequence modeling, tool calling, and CoT reasoning. We further validate the system in a style transfer setting, where audio effects information is transferred from a reference source and applied to new content. Finally, LLM-as-a-judge evaluation demonstrates that our approach generates appropriate CoT reasoning and responses for music production queries. To our knowledge, this is the first work to apply LLM-based tool calling to audio effects modules, enabling interpretable and controllable music production.
[612] Parallel Delayed Memory Units for Enhanced Temporal Modeling in Biomedical and Bioacoustic Signal Analysis
Pengfei Sun, Wenyu Jiang, Paul Devos, Dick Botteldooren
Main category: cs.SD
TL;DR: Proposed Parallel Delayed Memory Unit (PDMU) - a delay-gated state-space module for audio/bioacoustic signals that enhances short-term temporal interactions and memory efficiency, outperforming existing RNN approaches.
Details
Motivation: Gated RNNs can be over-parameterized and training-inefficient, while linear RNNs struggle to capture bio-signal complexity. Need for better temporal credit assignment in data-scarce audio/bioacoustic applications.Method: PDMU uses gated delay-line mechanism with Legendre Memory Units to compress temporal info into vector representations, acting as causal attention. Modular design supports parallel training/sequential inference, with bidirectional/efficient/spiking variants.
Result: PDMU significantly enhances memory capacity and overall model performance across diverse audio and biomedical benchmarks, with variants offering additional performance/energy efficiency gains.
Conclusion: PDMU effectively addresses RNN limitations for audio/bioacoustic signals through improved temporal state interactions, memory efficiency, and flexible integration into existing frameworks.
Abstract: Advanced deep learning architectures, particularly recurrent neural networks (RNNs), have been widely applied in audio, bioacoustic, and biomedical signal analysis, especially in data-scarce environments. While gated RNNs remain effective, they can be relatively over-parameterised and less training-efficient in some regimes, while linear RNNs tend to fall short in capturing the complexity inherent in bio-signals. To address these challenges, we propose the Parallel Delayed Memory Unit (PDMU), a {delay-gated state-space module for short-term temporal credit assignment} targeting audio and bioacoustic signals, which enhances short-term temporal state interactions and memory efficiency via a gated delay-line mechanism. Unlike previous Delayed Memory Units (DMU) that embed temporal dynamics into the delay-line architecture, the PDMU further compresses temporal information into vector representations using Legendre Memory Units (LMU). This design serves as a form of causal attention, allowing the model to dynamically adjust its reliance on past states and improve real-time learning performance. Notably, in low-information scenarios, the gating mechanism behaves similarly to skip connections by bypassing state decay and preserving early representations, thereby facilitating long-term memory retention. The PDMU is modular, supporting parallel training and sequential inference, and can be easily integrated into existing linear RNN frameworks. Furthermore, we introduce bidirectional, efficient, and spiking variants of the architecture, each offering additional gains in performance or energy efficiency. Experimental results on diverse audio and biomedical benchmarks demonstrate that the PDMU significantly enhances both memory capacity and overall model performance.
[613] Speech Audio Generation from dynamic MRI via a Knowledge Enhanced Conditional Variational Autoencoder
Yaxuan Li, Han Jiang, Yifei Ma, Shihua Qin, Jonghye Woo, Fangxu Xing
Main category: cs.SD
TL;DR: Proposes KE-CVAE, a two-step “knowledge enhancement + variational inference” framework for generating speech audio from dynamic MRI sequences, addressing MRI-specific acoustic challenges and outperforming existing methods.
Details
Motivation: Dynamic MRI of vocal tract is important for speech studies, but data loss, noise, and audio corruption occur due to MRI environment. Generating audio from images is critical for data recovery in clinical/research applications, but challenging due to hardware constraints, acoustic interference, and data corruption.Method: Knowledge Enhanced Conditional Variational Autoencoder (KE-CVAE) - a novel two-step framework: 1) integration of unlabeled MRI data for knowledge enhancement, 2) variational inference architecture to improve generative modeling capacity. Trained on open-source dynamic vocal tract MRI dataset recorded during speech.
Result: Method effectively generates natural speech waveforms while addressing MRI-specific acoustic challenges, outperforming conventional deep learning-based synthesis approaches.
Conclusion: This represents one of the first attempts at synthesizing speech audio directly from dynamic MRI video sequences, offering a promising solution for MRI-based speech data recovery and analysis.
Abstract: Dynamic Magnetic Resonance Imaging (MRI) of the vocal tract has become an increasingly adopted imaging modality for speech motor studies. Beyond image signals, systematic data loss, noise pollution, and audio file corruption can occur due to the unpredictability of the MRI acquisition environment. In such cases, generating audio from images is critical for data recovery in both clinical and research applications. However, this remains challenging due to hardware constraints, acoustic interference, and data corruption. Existing solutions, such as denoising and multi-stage synthesis methods, face limitations in audio fidelity and generalizability. To address these challenges, we propose a Knowledge Enhanced Conditional Variational Autoencoder (KE-CVAE), a novel two-step “knowledge enhancement + variational inference” framework for generating speech audio signals from cine dynamic MRI sequences. This approach introduces two key innovations: (1) integration of unlabeled MRI data for knowledge enhancement, and (2) a variational inference architecture to improve generative modeling capacity. To the best of our knowledge, this is one of the first attempts at synthesizing speech audio directly from dynamic MRI video sequences. The proposed method was trained and evaluated on an open-source dynamic vocal tract MRI dataset recorded during speech. Experimental results demonstrate its effectiveness in generating natural speech waveforms while addressing MRI-specific acoustic challenges, outperforming conventional deep learning-based synthesis approaches.
[614] SpeechJudge: Towards Human-Level Judgment for Speech Naturalness
Xueyao Zhang, Chaoren Wang, Huan Liao, Ziniu Li, Yuancheng Wang, Li Wang, Dongya Jia, Yuanzhe Chen, Xiulin Li, Zhuo Chen, Zhizheng Wu
Main category: cs.SD
TL;DR: SpeechJudge introduces a comprehensive suite (dataset, benchmark, and reward model) for aligning speech synthesis models with human preferences on naturalness, addressing the lack of large-scale human feedback data in TTS.
Details
Motivation: Large generative models need human feedback alignment, but speech synthesis lacks large-scale human preference datasets, hindering development of models that align with human perception of naturalness.Method: 1) Created SpeechJudge-Data: 99K speech pairs from diverse zero-shot TTS models across styles/languages with human annotations. 2) Established SpeechJudge-Eval benchmark. 3) Developed SpeechJudge-GRM: generative reward model based on Qwen2.5-Omni-7B trained via two-stage process (SFT with Chain-of-Thought + RL with GRPO).
Result: SpeechJudge-GRM achieves 77.2% accuracy (79.4% after scaling) on benchmark, outperforming Gemini-2.5-Flash (<70%) and classic Bradley-Terry model (72.7%). The model can also serve as reward function for aligning speech generation models.
Conclusion: SpeechJudge provides effective tools for human preference alignment in speech synthesis, with the GRM approach showing superior performance over existing methods and enabling better alignment of TTS models with human perception.
Abstract: Aligning large generative models with human feedback is a critical challenge. In speech synthesis, this is particularly pronounced due to the lack of a large-scale human preference dataset, which hinders the development of models that truly align with human perception. To address this, we introduce SpeechJudge, a comprehensive suite comprising a dataset, a benchmark, and a reward model centered on naturalness–one of the most fundamental subjective metrics for speech synthesis. First, we present SpeechJudge-Data, a large-scale human feedback corpus of 99K speech pairs. The dataset is constructed using a diverse set of advanced zero-shot text-to-speech (TTS) models across diverse speech styles and multiple languages, with human annotations for both intelligibility and naturalness preference. From this, we establish SpeechJudge-Eval, a challenging benchmark for speech naturalness judgment. Our evaluation reveals that existing metrics and AudioLLMs struggle with this task; the leading model, Gemini-2.5-Flash, achieves less than 70% agreement with human judgment, highlighting a significant gap for improvement. To bridge this gap, we develop SpeechJudge-GRM, a generative reward model (GRM) based on Qwen2.5-Omni-7B. It is trained on SpeechJudge-Data via a two-stage post-training process: Supervised Fine-Tuning (SFT) with Chain-of-Thought rationales followed by Reinforcement Learning (RL) with GRPO on challenging cases. On the SpeechJudge-Eval benchmark, the proposed SpeechJudge-GRM demonstrates superior performance, achieving 77.2% accuracy (and 79.4% after inference-time scaling @10) compared to a classic Bradley-Terry reward model (72.7%). Furthermore, SpeechJudge-GRM can be also employed as a reward function during the post-training of speech generation models to facilitate their alignment with human preferences.
[615] IMSE: Efficient U-Net-based Speech Enhancement using Inception Depthwise Convolution and Amplitude-Aware Linear Attention
Xinxin Tang, Bin Qin, Yufang Li
Main category: cs.SD
TL;DR: IMSE proposes an ultra-lightweight speech enhancement network that reduces parameters by 16.8% while maintaining competitive performance by replacing complex modules with more efficient alternatives.
Details
Motivation: Existing lightweight speech enhancement methods like MUSE still suffer from efficiency bottlenecks despite having only 0.51M parameters. The MET module uses a complex "approximate-compensate" mechanism, and deformable embedding introduces additional computational burden, limiting deployment on resource-constrained devices.Method: IMSE introduces two core innovations: 1) Amplitude-Aware Linear Attention (MALA) replaces the MET module, preserving query vector norm information to solve the “amplitude-ignoring” problem in linear attention without auxiliary compensation. 2) Inception Depthwise Convolution (IDConv) replaces deformable embedding, using parallel branches (square, horizontal, vertical strips) to capture spectrogram features with minimal parameter redundancy.
Result: IMSE reduces parameters by 16.8% (from 0.513M to 0.427M) while achieving competitive PESQ score of 3.373 on VoiceBank+DEMAND dataset, comparable to state-of-the-art performance.
Conclusion: The study sets a new benchmark for balancing model size and speech quality in ultra-lightweight speech enhancement, demonstrating that systematic optimization can significantly reduce computational burden while maintaining performance for resource-constrained devices.
Abstract: Achieving a balance between lightweight design and high performance remains a significant challenge for speech enhancement (SE) tasks on resource-constrained devices. Existing state-of-the-art methods, such as MUSE, have established a strong baseline with only 0.51M parameters by introducing a Multi-path Enhanced Taylor (MET) transformer and Deformable Embedding (DE). However, an in-depth analysis reveals that MUSE still suffers from efficiency bottlenecks: the MET module relies on a complex “approximate-compensate” mechanism to mitigate the limitations of Taylor-expansion-based attention, while the offset calculation for deformable embedding introduces additional computational burden. This paper proposes IMSE, a systematically optimized and ultra-lightweight network. We introduce two core innovations: 1) Replacing the MET module with Amplitude-Aware Linear Attention (MALA). MALA fundamentally rectifies the “amplitude-ignoring” problem in linear attention by explicitly preserving the norm information of query vectors in the attention calculation, achieving efficient global modeling without an auxiliary compensation branch. 2) Replacing the DE module with Inception Depthwise Convolution (IDConv). IDConv borrows the Inception concept, decomposing large-kernel operations into efficient parallel branches (square, horizontal, and vertical strips), thereby capturing spectrogram features with extremely low parameter redundancy. Extensive experiments on the VoiceBank+DEMAND dataset demonstrate that, compared to the MUSE baseline, IMSE significantly reduces the parameter count by 16.8% (from 0.513M to 0.427M) while achieving competitive performance comparable to the state-of-the-art on the PESQ metric (3.373). This study sets a new benchmark for the trade-off between model size and speech quality in ultra-lightweight speech enhancement.
[616] Probabilistic Fusion and Calibration of Neural Speaker Diarization Models
Juan Ignacio Alvarez-Trejos, Sergio A. Balanya, Daniel Ramos, Alicia Lozano-Diez
Main category: cs.SD
TL;DR: First framework for calibrating and fusing EEND models at probability level, showing calibration improves individual models by up to 19% DER reduction and fusion outperforms DOVER-Lap.
Details
Motivation: EEND systems produce frame-level probabilistic speaker activity estimates, but their reliability and calibration have been neglected. Current fusion methods like DOVER-Lap use hard decisions at segment level, missing opportunities to leverage model uncertainty and complementary strengths across architectures.Method: Proposes comprehensive framework for calibrating and fusing EEND models at probability level. Investigates two output formulations (multilabel and powerset representations) and their impact on calibration and fusion. Explores different calibration strategies (independent per-speaker vs joint in powerset space) and fusion orderings (Fuse-then-Calibrate vs calibrating before fusion).
Result: Proper calibration provides substantial improvements for individual models (up to 19% relative DER reduction), sometimes mitigating absence of domain adaptation. Joint calibration in powerset space consistently outperforms independent per-speaker calibration. Fusion substantially improves over individual models. Fuse-then-Calibrate ordering generally outperforms both calibrating before fusion and uncalibrated fusion while requiring calibration of only a single combined model. Best configuration outperforms DOVER-Lap in terms of DER.
Conclusion: This work establishes best practices for probability-level fusion of EEND systems and demonstrates the advantages of leveraging soft outputs over hard decisions, providing reliable confidence estimates essential for downstream applications.
Abstract: End-to-End Neural Diarization (EEND) systems produce frame-level probabilistic speaker activity estimates, yet since evaluation focuses primarily on Diarization Error Rate (DER), the reliability and calibration of these confidence scores have been largely neglected. When fusing multiple diarization systems, DOVER-Lap remains the only established approach, operating at the segment level with hard decisions. We propose working with continuous probability outputs, which enables more sophisticated fusion and calibration techniques that can leverage model uncertainty and complementary strengths across different architectures. This paper presents the first comprehensive framework for calibrating and fusing EEND models at the probability level. We investigate two output formulations (multilabel and powerset representations) and their impact on calibration and fusion effectiveness. Through extensive experiments on the CallHome two-speaker benchmark, we demonstrate that proper calibration provides substantial improvements even for individual models (up to 19% relative DER reduction), in some cases mitigating the absence of domain adaptation. We reveal that joint calibration in powerset space consistently outperforms independent per-speaker calibration, that fusion substantially improves over individual models, and that the Fuse-then-Calibrate ordering generally outperforms both calibrating before fusion and uncalibrated fusion while requiring calibration of only a single combined model. Our best configuration outperforms DOVER-Lap in terms of DER while providing reliable confidence estimates essential for downstream applications. This work proposes best practices for probability-level fusion of EEND systems and demonstrates the advantages of leveraging soft outputs over hard decisions.
cs.LG
[617] Measuring What LLMs Think They Do: SHAP Faithfulness and Deployability on Financial Tabular Classification
Saeed AlMarri, Mathieu Ravaut, Kristof Juhasz, Gautier Marti, Hamdan Al Ahbabi, Ibrahim Elfadel
Main category: cs.LG
TL;DR: LLMs show promise for financial classification but have reliability issues with structured tabular data, with discrepancies between self-explanations and SHAP values compared to traditional models like LightGBM.
Details
Motivation: LLMs offer flexible classification alternatives to traditional ML models like LightGBM, but their reliability for structured tabular data in high-stakes financial applications remains uncertain and needs systematic evaluation.Method: Systematically evaluate LLMs on financial classification tasks, generate SHAP values for LLMs, and compare them with LightGBM’s SHAP values to assess reliability and explainability.
Result: LLMs show divergence between their self-explanation of feature impact and SHAP values, and notable differences between LLM and LightGBM SHAP values, indicating reliability issues for structured financial modeling.
Conclusion: LLMs have limitations as standalone classifiers for structured financial modeling, but improved explainability mechanisms with few-shot prompting could make them usable in risk-sensitive domains.
Abstract: Large Language Models (LLMs) have attracted significant attention for classification tasks, offering a flexible alternative to trusted classical machine learning models like LightGBM through zero-shot prompting. However, their reliability for structured tabular data remains unclear, particularly in high stakes applications like financial risk assessment. Our study systematically evaluates LLMs and generates their SHAP values on financial classification tasks. Our analysis shows a divergence between LLMs self-explanation of feature impact and their SHAP values, as well as notable differences between LLMs and LightGBM SHAP values. These findings highlight the limitations of LLMs as standalone classifiers for structured financial modeling, but also instill optimism that improved explainability mechanisms coupled with few-shot prompting will make LLMs usable in risk-sensitive domains.
[618] Faster Verified Explanations for Neural Networks
Alessandro De Palma, Greta Dolcetti, Caterina Urban
Main category: cs.LG
TL;DR: FaVeX is a novel algorithm for computing verified explanations of neural networks that improves scalability by combining batch/sequential processing and reusing information, plus introduces verifier-optimal robust explanations that account for verifier incompleteness.
Details
Motivation: Verified explanations provide principled explanations for neural network decisions, but current techniques face significant scalability challenges due to multiple calls to neural network verifiers with exponential worst-case complexity.Method: FaVeX algorithm accelerates computation by: 1) dynamically combining batch and sequential processing of input features, 2) reusing information from previous queries for proving invariances and searching for feature assignments, and 3) introducing verifier-optimal robust explanations that factor verifier incompleteness into the explanation.
Result: Comprehensive experimental evaluation shows superior scalability - FaVeX and verifier-optimal robust explanations together can produce meaningful formal explanations on networks with hundreds of thousands of non-linear activations.
Conclusion: FaVeX addresses the scalability bottleneck of verified explanations, making formal explanation of large neural networks practical through algorithmic optimizations and a novel explanation framework that accounts for verifier limitations.
Abstract: Verified explanations are a theoretically-principled way to explain the decisions taken by neural networks, which are otherwise black-box in nature. However, these techniques face significant scalability challenges, as they require multiple calls to neural network verifiers, each of them with an exponential worst-case complexity. We present FaVeX, a novel algorithm to compute verified explanations. FaVeX accelerates the computation by dynamically combining batch and sequential processing of input features, and by reusing information from previous queries, both when proving invariances with respect to certain input features, and when searching for feature assignments altering the prediction. Furthermore, we present a novel and hierarchical definition of verified explanations, termed verifier-optimal robust explanations, that explicitly factors the incompleteness of network verifiers within the explanation. Our comprehensive experimental evaluation demonstrates the superior scalability of both FaVeX, and of verifier-optimal robust explanations, which together can produce meaningful formal explanation on networks with hundreds of thousands of non-linear activations.
[619] We Still Don’t Understand High-Dimensional Bayesian Optimization
Colin Doumont, Donney Fan, Natalie Maus, Jacob R. Gardner, Henry Moss, Geoff Pleiss
Main category: cs.LG
TL;DR: Bayesian linear regression with geometric transformation outperforms complex BO methods in high-dimensional optimization tasks (60-6,000 dimensions).
Details
Motivation: High-dimensional spaces challenge Bayesian optimization, and existing methods rely on complex structural assumptions (locality, sparsity, smoothness) that may not be optimal.Method: Use Bayesian linear regression with Gaussian processes using linear kernels, applying geometric transformation to avoid boundary-seeking behavior. This simple parametric approach enables closed-form sampling and linear computational scaling with data.
Result: Linear models match state-of-the-art performance on 60-6,000 dimensional tasks and scale effectively to molecular optimization with >20,000 observations, outperforming existing BO methods.
Conclusion: Simple linear models challenge traditional intuitions about BO in high dimensions, suggesting need to reconsider complex structural assumptions in favor of simpler, scalable parametric approaches.
Abstract: High-dimensional spaces have challenged Bayesian optimization (BO). Existing methods aim to overcome this so-called curse of dimensionality by carefully encoding structural assumptions, from locality to sparsity to smoothness, into the optimization procedure. Surprisingly, we demonstrate that these approaches are outperformed by arguably the simplest method imaginable: Bayesian linear regression. After applying a geometric transformation to avoid boundary-seeking behavior, Gaussian processes with linear kernels match state-of-the-art performance on tasks with 60- to 6,000-dimensional search spaces. Linear models offer numerous advantages over their non-parametric counterparts: they afford closed-form sampling and their computation scales linearly with data, a fact we exploit on molecular optimization tasks with > 20,000 observations. Coupled with empirical analyses, our results suggest the need to depart from past intuitions about BO methods in high-dimensional spaces.
[620] Orion-Bix: Bi-Axial Attention for Tabular In-Context Learning
Mohamed Bouadi, Pratinav Seth, Aditya Tanna, Vinay Kumar Sankarapu
Main category: cs.LG
TL;DR: Orion-Bix is a tabular foundation model that uses biaxial attention and meta-learned in-context reasoning for few-shot learning on tabular data, outperforming gradient-boosting baselines and competing with state-of-the-art tabular foundation models.
Details
Motivation: Tabular data presents unique challenges for machine learning: mixed numeric/categorical fields, weak feature structure, and limited labeled data make scaling and generalization difficult. Current approaches struggle with few-shot learning scenarios.Method: Combines biaxial attention (alternating standard, grouped, hierarchical, and relational attention) with multi-CLS summarization to capture local and global dependencies. Features a label-aware in-context learning head with hierarchical decision routing for large label spaces. Meta-trained on synthetically generated, structurally diverse tables with causal priors.
Result: Outperforms gradient-boosting baselines and remains competitive with state-of-the-art tabular foundation models on public benchmarks. Demonstrates robust few-shot learning capabilities across heterogeneous tabular data.
Conclusion: Biaxial attention combined with episodic meta-training enables robust, few-shot-ready tabular learning. The approach learns transferable inductive biases that generalize across diverse tabular data structures, providing a practical foundation model solution for real-world tabular ML applications.
Abstract: Tabular data drive most real-world machine learning applications, yet building general-purpose models for them remains difficult. Mixed numeric and categorical fields, weak feature structure, and limited labeled data make scaling and generalization challenging. To this end, we introduce Orion-Bix, a tabular foundation model that combines biaxial attention with meta-learned in-context reasoning for few-shot tabular learning. Its encoder alternates standard, grouped, hierarchical, and relational attention, fusing their outputs through multi-CLS summarization to capture both local and global dependencies efficiently. A label-aware ICL head adapts on the fly and scales to large label spaces via hierarchical decision routing. Meta-trained on synthetically generated, structurally diverse tables with causal priors, Orion-Bix learns transferable inductive biases across heterogeneous data. Delivered as a scikit-learn compatible foundation model, it outperforms gradient-boosting baselines and remains competitive with state-of-the-art tabular foundation models on public benchmarks, showing that biaxial attention with episodic meta-training enables robust, few-shot-ready tabular learning. The model is publicly available at https://github.com/Lexsi-Labs/Orion-BiX .
[621] Hybrid Context-Fusion Attention (CFA) U-Net and Clustering for Robust Seismic Horizon Interpretation
Jose Luis Lima de Jesus Silva, Joao Pedro Gomes, Paulo Roberto de Melo Barros Junior, Vitor Hugo Serravalle Reis Rodrigues, Alexsandro Guerra Cerqueira
Main category: cs.LG
TL;DR: A hybrid deep learning framework combining attention-based U-Net variants with spatial clustering improves seismic horizon interpretation, achieving state-of-the-art results on complex geological datasets under sparse annotation conditions.
Details
Motivation: Current deep learning approaches for seismic horizon interpretation struggle with accurately segmenting complex geological features and interpolating horizons from sparse annotations, limiting their practical application in hydrocarbon exploration.Method: Proposes a hybrid framework integrating advanced U-Net variants with spatial clustering. The core innovation is Context Fusion Attention (CFA) U-Net, which fuses spatial and Sobel-derived geometric features within attention gates. Five architectures were evaluated across different data sparsity regimes, with DBSCAN clustering used to refine merged horizon predictions.
Result: Achieved state-of-the-art performance with validation IoU of 0.881 and MAE of 2.49ms on Mexilhao field dataset, and 97.6% surface coverage on F3 Block dataset under sparse conditions. The hybrid approach outperformed existing baselines across various sparsity regimes.
Conclusion: The hybrid methodology combining attention-based architectures with geometric context and spatial clustering provides a robust, generalizable solution for seismic interpretation in structurally complex and data-scarce environments, demonstrating advantages over traditional approaches.
Abstract: Interpreting seismic horizons is a critical task for characterizing subsurface structures in hydrocarbon exploration. Recent advances in deep learning, particularly U-Net-based architectures, have significantly improved automated horizon tracking. However, challenges remain in accurately segmenting complex geological features and interpolating horizons from sparse annotations. To address these issues, a hybrid framework is presented that integrates advanced U-Net variants with spatial clustering to enhance horizon continuity and geometric fidelity. The core contribution is the Context Fusion Attention (CFA) U-Net, a novel architecture that fuses spatial and Sobel-derived geometric features within attention gates to improve both precision and surface completeness. The performance of five architectures, the U-Net (Standard and compressed), U-Net++, Attention U-Net, and CFA U-Net, was systematically evaluated across various data sparsity regimes (10-, 20-, and 40-line spacing). This approach outperformed existing baselines, achieving state-of-the-art results on the Mexilhao field (Santos Basin, Brazil) dataset with a validation IoU of 0.881 and MAE of 2.49ms, and excellent surface coverage of 97.6% on the F3 Block of the North Sea dataset under sparse conditions. The framework further refines merged horizon predictions (inline and cross-line) using Density-Based Spatial Clustering of Applications with Noise (DBSCAN) to produce geologically plausible surfaces. The results demonstrate the advantages of hybrid methodologies and attention-based architectures enhanced with geometric context, providing a robust and generalizable solution for seismic interpretation in structurally complex and data-scarce environments.
[622] Emergent Riemannian geometry over learning discrete computations on continuous manifolds
Julian Brandon, Angus Chadwick, Arthur Pellegrino
Main category: cs.LG
TL;DR: Neural networks learn to perform discrete computations on continuous data by developing geometric structures in their representations, with different learning regimes affecting generalization.
Details
Motivation: Understanding how neural networks perform discrete computations (like classification) on continuous input data (like images) remains poorly understood, despite being fundamental to many AI tasks.Method: Analyzing the Riemannian pullback metric across neural network layers to study representational geometry, decomposing computation into discretization of continuous features and logical operations on discretized variables.
Result: Found that network computation decomposes into discretization and logical operations, and that different learning regimes (rich vs. lazy) have contrasting metric and curvature structures affecting generalization to unseen inputs.
Conclusion: Provides a geometric framework for understanding how neural networks learn discrete computations on continuous manifolds, revealing fundamental computational principles through representational geometry analysis.
Abstract: Many tasks require mapping continuous input data (e.g. images) to discrete task outputs (e.g. class labels). Yet, how neural networks learn to perform such discrete computations on continuous data manifolds remains poorly understood. Here, we show that signatures of such computations emerge in the representational geometry of neural networks as they learn. By analysing the Riemannian pullback metric across layers of a neural network, we find that network computation can be decomposed into two functions: discretising continuous input features and performing logical operations on these discretised variables. Furthermore, we demonstrate how different learning regimes (rich vs. lazy) have contrasting metric and curvature structures, affecting the ability of the networks to generalise to unseen inputs. Overall, our work provides a geometric framework for understanding how neural networks learn to perform discrete computations on continuous manifolds.
[623] Constructing Efficient Fact-Storing MLPs for Transformers
Owen Dugan, Roberto Garcia, Ronny Junkins, Jerry Liu, Dylan Zinsley, Sabri Eyuboglu, Atri Rudra, Chris Ré
Main category: cs.LG
TL;DR: The paper introduces an improved MLP construction framework for storing factual knowledge in LLMs that achieves near-universal applicability, optimal parameter efficiency, and maintains Transformer usability, revealing key insights about fact-storage capacity and enabling modular fact editing.
Details
Motivation: To improve understanding of how LLMs store factual knowledge as key-value mappings in MLP parameters, and to develop better explicit weight constructions that overcome limitations of previous approaches while maintaining practical utility in Transformer architectures.Method: Developed an MLP construction framework that works for almost all feasible input-output pairs, achieves asymptotically optimal parameter efficiency matching information-theoretic bounds, and maintains usability within Transformers for factual recall. The approach enables modular fact editing by replacing entire MLPs in one-layer Transformers.
Result: Discovered a metric on value embeddings that characterizes facts-per-parameter scaling for both constructed and gradient-descent-trained MLPs; identified a simple encoder-decoder mechanism matching gradient-descent MLP facts-per-parameter asymptotics; uncovered fundamental tradeoff between MLP fact-storage capacity and Transformer usability.
Conclusion: The improved MLP construction framework provides better understanding of fact storage mechanisms in LLMs, reveals fundamental scaling laws and tradeoffs, and enables practical applications like modular fact editing through MLP replacement in Transformers.
Abstract: The success of large language models (LLMs) can be attributed in part to their ability to efficiently store factual knowledge as key-value mappings within their MLP parameters. Recent work has proposed explicit weight constructions to build such fact-storing MLPs, providing an improved understanding of LLM fact storage mechanisms. In this paper, we introduce an MLP construction framework that improves over previous constructions in three areas: it 1) works for all but a measure-zero set of feasible input-output pairs, 2) achieves asymptotically optimal parameter efficiency matching information-theoretic bounds for some embeddings, and 3) maintains usability within Transformers for factual recall. Through our improvements, we 1) discover a metric on value embeddings that characterizes facts-per-parameter scaling for both constructed and gradient-descent-trained MLPs, 2) identify a simple encoder-decoder mechanism that empirically matches gradient-descent MLP facts-per-parameter asymptotics across all the inputs and outputs we test, and 3) uncover a fundamental tradeoff between an MLP’s fact-storage capacity and its usability within Transformers. Finally, we demonstrate a proof-of-concept application of fact-storing MLPs: modular fact editing on one-layer Transformers by \textit{replacing entire MLPs at once}.
[624] TIE: A Training-Inversion-Exclusion Framework for Visually Interpretable and Uncertainty-Guided Out-of-Distribution Detection
Pirzada Suhail, Rehna Afroz, Amit Sethi
Main category: cs.LG
TL;DR: TIE is a unified framework for uncertainty estimation and OOD detection that extends classifiers with a garbage class and uses iterative training-inversion-exclusion to create visually interpretable anomaly detection.
Details
Motivation: Deep neural networks often fail to recognize out-of-distribution inputs, leading to overconfident predictions. There's a need for unified methods that can both estimate predictive uncertainty and detect OOD samples reliably.Method: TIE extends n-class classifiers to (n+1)-class models by adding a garbage class initialized with Gaussian noise. It performs iterative training-inversion-exclusion cycles: training the classifier, inverting highly uncertain samples from it, and excluding these inverted samples into the garbage class. This creates visually coherent class prototypes over time.
Result: TIE achieves near-perfect OOD detection with ≈0 FPR@95%TPR on MNIST and FashionMNIST when tested against diverse unseen datasets. It provides unified anomaly detection and calibrated uncertainty estimation without needing external OOD datasets.
Conclusion: TIE offers a unified, interpretable framework for robust anomaly detection and calibrated uncertainty estimation that effectively addresses both OOD detection and uncertainty estimation challenges through its iterative refinement approach.
Abstract: Deep neural networks often struggle to recognize when an input lies outside their training experience, leading to unreliable and overconfident predictions. Building dependable machine learning systems therefore requires methods that can both estimate predictive \textit{uncertainty} and detect \textit{out-of-distribution (OOD)} samples in a unified manner. In this paper, we propose \textbf{TIE: a Training–Inversion–Exclusion} framework for visually interpretable and uncertainty-guided anomaly detection that jointly addresses these challenges through iterative refinement. TIE extends a standard $n$-class classifier to an $(n+1)$-class model by introducing a garbage class initialized with Gaussian noise to represent outlier inputs. Within each epoch, TIE performs a closed-loop process of \textit{training, inversion, and exclusion}, where highly uncertain inverted samples reconstructed from the just-trained classifier are excluded into the garbage class. Over successive iterations, the inverted samples transition from noisy artifacts into visually coherent class prototypes, providing transparent insight into how the model organizes its learned manifolds. During inference, TIE rejects OOD inputs by either directly mapping them to the garbage class or producing low-confidence, uncertain misclassifications within the in-distribution classes that are easily separable, all without relying on external OOD datasets. A comprehensive threshold-based evaluation using multiple OOD metrics and performance measures such as \textit{AUROC}, \textit{AUPR}, and \textit{FPR@95%TPR} demonstrates that TIE offers a unified and interpretable framework for robust anomaly detection and calibrated uncertainty estimation (UE) achieving near-perfect OOD detection with \textbf{(!\approx!) 0 FPR@95%TPR} when trained on MNIST or FashionMNIST and tested against diverse unseen datasets.
[625] Self-Supervised Dynamical System Representations for Physiological Time-Series
Yenho Chen, Maxwell A. Xu, James M. Rehg, Christopher J. Rozell
Main category: cs.LG
TL;DR: PULSE: A cross-reconstruction pretraining framework for physiological time series that extracts shared system parameters while discarding sample-specific noise, improving representation learning for downstream tasks.
Details
Motivation: Existing SSL methods for physiological time series rely on heuristic principles or poorly constrained generative tasks, limiting their ability to preserve information about underlying physiological states while filtering out noise.Method: Proposes PULSE, a cross-reconstruction-based pretraining objective that exploits information structure of dynamical systems generative models across multiple time-series. It extracts shared system parameters (class identity) while discarding sample-specific noise through cross-reconstruction between similar samples.
Result: Established theoretical conditions for system information recovery, validated with synthetic dynamical systems experiments. Applied to diverse real-world datasets, showing PULSE learns representations that distinguish semantic classes, increases label efficiency, and improves transfer learning.
Conclusion: PULSE provides a principled SSL framework for physiological time series that effectively extracts meaningful system information while filtering noise, outperforming existing methods across multiple real-world applications.
Abstract: The effectiveness of self-supervised learning (SSL) for physiological time series depends on the ability of a pretraining objective to preserve information about the underlying physiological state while filtering out unrelated noise. However, existing strategies are limited due to reliance on heuristic principles or poorly constrained generative tasks. To address this limitation, we propose a pretraining framework that exploits the information structure of a dynamical systems generative model across multiple time-series. This framework reveals our key insight that class identity can be efficiently captured by extracting information about the generative variables related to the system parameters shared across similar time series samples, while noise unique to individual samples should be discarded. Building on this insight, we propose PULSE, a cross-reconstruction-based pretraining objective for physiological time series datasets that explicitly extracts system information while discarding non-transferrable sample-specific ones. We establish theory that provides sufficient conditions for the system information to be recovered, and empirically validate it using a synthetic dynamical systems experiment. Furthermore, we apply our method to diverse real-world datasets, demonstrating that PULSE learns representations that can broadly distinguish semantic classes, increase label efficiency, and improve transfer learning.
[626] A Hierarchical Hybrid AI Approach: Integrating Deep Reinforcement Learning and Scripted Agents in Combat Simulations
Scotty Black, Christian Darken
Main category: cs.LG
TL;DR: Hybrid hierarchical AI approach combines scripted agents for tactical decisions with RL agents for strategic decisions in combat simulations, improving performance over using either method alone.
Details
Motivation: Current combat simulation agents use either rule-based scripted methods (predictable but inflexible) or deep RL (adaptive but with black-box issues and scalability problems). Neither approach alone is optimal for complex, dynamic wargaming scenarios.Method: Novel hierarchical hybrid AI that synergizes scripted agents and RL agents. Scripted agents handle routine tactical-level decisions while RL agents manage higher-level strategic decision-making in a structured hierarchy.
Result: The integration significantly improves overall performance, providing a robust, adaptable, and effective solution for developing intelligent agents in complex simulation environments.
Conclusion: The hybrid hierarchical approach addresses limitations of both scripted and RL methods while leveraging their individual strengths, offering a superior solution for intelligent agent development in combat simulations.
Abstract: In the domain of combat simulations in support of wargaming, the development of intelligent agents has predominantly been characterized by rule-based, scripted methodologies with deep reinforcement learning (RL) approaches only recently being introduced. While scripted agents offer predictability and consistency in controlled environments, they fall short in dynamic, complex scenarios due to their inherent inflexibility. Conversely, RL agents excel in adaptability and learning, offering potential improvements in handling unforeseen situations, but suffer from significant challenges such as black-box decision-making processes and scalability issues in larger simulation environments. This paper introduces a novel hierarchical hybrid artificial intelligence (AI) approach that synergizes the reliability and predictability of scripted agents with the dynamic, adaptive learning capabilities of RL. By structuring the AI system hierarchically, the proposed approach aims to utilize scripted agents for routine, tactical-level decisions and RL agents for higher-level, strategic decision-making, thus addressing the limitations of each method while leveraging their individual strengths. This integration is shown to significantly improve overall performance, providing a robust, adaptable, and effective solution for developing and training intelligent agents in complex simulation environments.
[627] Polynomial Neural Sheaf Diffusion: A Spectral Filtering Approach on Cellular Sheaves
Alessio Borgi, Fabrizio Silvestri, Pietro Liò
Main category: cs.LG
TL;DR: PolyNSD introduces polynomial sheaf diffusion with K-hop receptive fields via stable recurrence, achieving SOTA on homophilic/heterophilic graphs with diagonal restriction maps, reducing computation and memory.
Details
Motivation: Current Neural Sheaf Diffusion methods have limitations: they rely on SVD-based normalization, dense per-edge restriction maps that scale poorly with stalk dimension, require frequent Laplacian rebuilds, and produce brittle gradients. These issues hinder scalability and robustness.Method: Proposes Polynomial Neural Sheaf Diffusion (PolyNSD) using a degree-K polynomial in normalized sheaf Laplacian, evaluated via stable three-term recurrence on spectrally rescaled operator. Uses convex mixture of K+1 orthogonal polynomial basis responses for trainable spectral response, with stability enforced through convex mixtures, spectral rescaling, and residual/gated paths.
Result: Achieves new state-of-the-art results on both homophilic and heterophilic benchmarks. Inverts previous trend by obtaining strong performance with just diagonal restriction maps (decoupling from large stalk dimension), while reducing runtime and memory requirements.
Conclusion: PolyNSD provides an efficient, stable sheaf diffusion approach with explicit K-hop receptive fields, overcoming limitations of previous methods while achieving superior performance with reduced computational overhead.
Abstract: Sheaf Neural Networks equip graph structures with a cellular sheaf: a geometric structure which assigns local vector spaces (stalks) and a linear learnable restriction/transport maps to nodes and edges, yielding an edge-aware inductive bias that handles heterophily and limits oversmoothing. However, common Neural Sheaf Diffusion implementations rely on SVD-based sheaf normalization and dense per-edge restriction maps, which scale with stalk dimension, require frequent Laplacian rebuilds, and yield brittle gradients. To address these limitations, we introduce Polynomial Neural Sheaf Diffusion (PolyNSD), a new sheaf diffusion approach whose propagation operator is a degree-K polynomial in a normalised sheaf Laplacian, evaluated via a stable three-term recurrence on a spectrally rescaled operator. This provides an explicit K-hop receptive field in a single layer (independently of the stalk dimension), with a trainable spectral response obtained as a convex mixture of K+1 orthogonal polynomial basis responses. PolyNSD enforces stability via convex mixtures, spectral rescaling, and residual/gated paths, reaching new state-of-the-art results on both homophilic and heterophilic benchmarks, inverting the Neural Sheaf Diffusion trend by obtaining these results with just diagonal restriction maps, decoupling performance from large stalk dimension, while reducing runtime and memory requirements.
[628] SD-CGAN: Conditional Sinkhorn Divergence GAN for DDoS Anomaly Detection in IoT Networks
Henry Onyeka, Emmanuel Samson, Liang Hong, Tariqul Islam, Imtiaz Ahmed, Kamrul Hasan
Main category: cs.LG
TL;DR: SD-CGAN: A Conditional GAN with Sinkhorn Divergence for robust anomaly detection in IoT edge networks, addressing class imbalance and sophisticated DoS attacks.
Details
Motivation: IoT edge networks face challenges in detecting sophisticated DoS attacks and zero-day exploits under dynamic, imbalanced traffic conditions, requiring robust anomaly detection solutions.Method: Proposes SD-CGAN framework with CTGAN-based synthetic data augmentation for class imbalance and Sinkhorn Divergence as geometry-aware loss function to improve training stability and reduce mode collapse.
Result: SD-CGAN achieves superior detection accuracy, precision, recall, and F1-score compared to baseline deep learning and GAN-based approaches while maintaining computational efficiency for edge deployment.
Conclusion: SD-CGAN provides an effective solution for robust anomaly detection in IoT edge environments, addressing class imbalance and sophisticated attack detection with computational efficiency suitable for edge deployment.
Abstract: The increasing complexity of IoT edge networks presents significant challenges for anomaly detection, particularly in identifying sophisticated Denial-of-Service (DoS) attacks and zero-day exploits under highly dynamic and imbalanced traffic conditions. This paper proposes SD-CGAN, a Conditional Generative Adversarial Network framework enhanced with Sinkhorn Divergence, tailored for robust anomaly detection in IoT edge environments. The framework incorporates CTGAN-based synthetic data augmentation to address class imbalance and leverages Sinkhorn Divergence as a geometry-aware loss function to improve training stability and reduce mode collapse. The model is evaluated on exploitative attack subsets from the CICDDoS2019 dataset and compared against baseline deep learning and GAN-based approaches. Results show that SD-CGAN achieves superior detection accuracy, precision, recall, and F1-score while maintaining computational efficiency suitable for deployment in edge-enabled IoT environments.
[629] Scalable and Interpretable Scientific Discovery via Sparse Variational Gaussian Process Kolmogorov-Arnold Networks (SVGP KAN)
Y. Sungtaek Ju
Main category: cs.LG
TL;DR: SVGP-KAN combines sparse variational Gaussian processes with Kolmogorov-Arnold Networks to enable probabilistic outputs and uncertainty quantification while maintaining computational efficiency for larger datasets.
Details
Motivation: Standard KANs lack probabilistic outputs needed for uncertainty quantification, and existing GP-KAN extensions use exact inference methods with O(N³) complexity that limit application to small datasets.Method: Integrates sparse variational inference with KAN topology using M inducing points and analytic moment matching to reduce computational complexity from O(N³) to O(NM²). Also incorporates permutation-based importance analysis for structural identification.
Result: Achieves computational efficiency enabling probabilistic KANs to scale to larger scientific datasets while maintaining the ability to identify relevant inputs and classify functional relationships.
Conclusion: SVGP-KAN provides an efficient, scalable probabilistic extension to KANs that enables uncertainty quantification and structural identification for scientific applications.
Abstract: Kolmogorov-Arnold Networks (KANs) offer a promising alternative to Multi-Layer Perceptron (MLP) by placing learnable univariate functions on network edges, enhancing interpretability. However, standard KANs lack probabilistic outputs, limiting their utility in applications requiring uncertainty quantification. While recent Gaussian Process (GP) extensions to KANs address this, they utilize exact inference methods that scale cubically with data size N, restricting their application to smaller datasets. We introduce the Sparse Variational GP-KAN (SVGP-KAN), an architecture that integrates sparse variational inference with the KAN topology. By employing $M$ inducing points and analytic moment matching, our method reduces computational complexity from $O(N^3)$ to $O(NM^2)$ or linear in sample size, enabling the application of probabilistic KANs to larger scientific datasets. Furthermore, we demonstrate that integrating a permutation-based importance analysis enables the network to function as a framework for structural identification, identifying relevant inputs and classifying functional relationships.
[630] Time-Series at the Edge: Tiny Separable CNNs for Wearable Gait Detection and Optimal Sensor Placement
Andrea Procopio, Marco Esposito, Sara Raggiunto, Andrey Gizdov, Alberto Belli, Paola Pierleoni
Main category: cs.LG
TL;DR: Ultra-light 1D CNNs (300-500 parameters) outperform magnitude thresholding for Parkinson’s gait detection on resource-constrained wearables, achieving ~94% PR-AUC with 10x fewer parameters than baseline models.
Details
Motivation: Enable on-device time-series analysis for Parkinson's disease gait detection on resource-constrained wearables and edge nodes, addressing the need for efficient, accurate models that can run locally without heavy computational requirements.Method: Compare magnitude thresholding to three 1D CNN architectures: literature baseline (separable convolutions), ultra-light purely separable model (305 params), and ultra-light residual separable model (533 params). Use BioStampRC21 dataset with 2-second windows at 30Hz, subject-independent leave-one-subject-out validation on 16 Parkinson’s patients with chest-worn IMUs.
Result: Residual separable model (533 params) achieves PR-AUC=94.5%, F1=91.2%, MCC=89.4%, matching/surpassing baseline (5,552 params). Smallest model (305 params) reaches PR-AUC=94.0%. Thresholding has high recall (89.0%) but low precision (76.5%) with many false positives. Chest and thigh sensors most reliable; forearm sensors degrade performance due to arm motion.
Conclusion: Ultra-light separable CNNs provide superior accuracy-efficiency-generalization trade-off compared to fixed thresholds for wearable PD gait detection. Compact models execute within tight memory/latency budgets on STM32-class MCUs (sub-10ms), enabling on-sensor gating of transmission/storage, demonstrating value of tailored time-series models for edge deployment.
Abstract: We study on-device time-series analysis for gait detection in Parkinson’s disease (PD) from short windows of triaxial acceleration, targeting resource-constrained wearables and edge nodes. We compare magnitude thresholding to three 1D CNNs for time-series analysis: a literature baseline (separable convolutions) and two ultra-light models - one purely separable and one with residual connections. Using the BioStampRC21 dataset, 2 s windows at 30 Hz, and subject-independent leave-one-subject-out (LOSO) validation on 16 PwPD with chest-worn IMUs, our residual separable model (Model 2, 533 params) attains PR-AUC = 94.5%, F1 = 91.2%, MCC = 89.4%, matching or surpassing the baseline (5,552 params; PR-AUC = 93.7%, F1 = 90.5%, MCC = 88.5%) with approximately 10x fewer parameters. The smallest model (Model 1, 305 params) reaches PR-AUC = 94.0%, F1 = 91.0%, MCC = 89.1%. Thresholding obtains high recall (89.0%) but low precision (76.5%), yielding many false positives and high inter-subject variance. Sensor-position analysis (train-on-all) shows chest and thighs are most reliable; forearms degrade precision/recall due to non-gait arm motion; naive fusion of all sites does not outperform the best single site. Both compact CNNs execute within tight memory/latency budgets on STM32-class MCUs (sub-10 ms on low-power boards), enabling on-sensor gating of transmission/storage. Overall, ultra-light separable CNNs provide a superior accuracy-efficiency-generalization trade-off to fixed thresholds for wearable PD gait detection and underscore the value of tailored time-series models for edge deployment.
[631] Teleportation-Based Defenses for Privacy in Approximate Machine Unlearning
Mohammad M Maheri, Xavier Cadet, Peter Chin, Hamed Haddadi
Main category: cs.LG
TL;DR: Approximate machine unlearning methods are vulnerable to privacy attacks via membership inference and data reconstruction. The paper proposes WARP, a teleportation defense that leverages neural network symmetries to reduce gradient energy and increase parameter dispersion, achieving significant privacy gains while maintaining accuracy.
Details
Motivation: Approximate machine unlearning methods, while efficient alternatives to full retraining, introduce privacy vulnerabilities where adversaries can exploit differences between pre- and post-unlearning models for membership inference or data reconstruction attacks.Method: The paper first demonstrates vulnerabilities through unlearning-specific membership inference and reconstruction attacks. Then introduces WARP (Weight-space Alignment via Reparameterization), a plug-and-play teleportation defense that leverages neural network symmetries to reduce forget-set gradient energy and increase parameter dispersion while preserving predictions.
Result: WARP achieves consistent privacy gains across six unlearning algorithms, reducing adversarial advantage (AUC) by up to 64% in black-box and 92% in white-box settings, while maintaining accuracy on retained data.
Conclusion: Teleportation via neural network symmetries serves as an effective general tool for reducing attack success in approximate unlearning, addressing privacy vulnerabilities while preserving model utility.
Abstract: Approximate machine unlearning aims to efficiently remove the influence of specific data points from a trained model, offering a practical alternative to full retraining. However, it introduces privacy risks: an adversary with access to pre- and post-unlearning models can exploit their differences for membership inference or data reconstruction. We show these vulnerabilities arise from two factors: large gradient norms of forget-set samples and the close proximity of unlearned parameters to the original model. To demonstrate their severity, we propose unlearning-specific membership inference and reconstruction attacks, showing that several state-of-the-art methods (e.g., NGP, SCRUB) remain vulnerable. To mitigate this leakage, we introduce WARP, a plug-and-play teleportation defense that leverages neural network symmetries to reduce forget-set gradient energy and increase parameter dispersion while preserving predictions. This reparameterization obfuscates the signal of forgotten data, making it harder for attackers to distinguish forgotten samples from non-members or recover them via reconstruction. Across six unlearning algorithms, our approach achieves consistent privacy gains, reducing adversarial advantage (AUC) by up to 64% in black-box and 92% in white-box settings, while maintaining accuracy on retained data. These results highlight teleportation as a general tool for reducing attack success in approximate unlearning.
[632] BioArc: Discovering Optimal Neural Architectures for Biological Foundation Models
Yi Fang, Haoran Xu, Jiaxin Han, Sirui Ding, Yizhi Wang, Yue Wang, Xuan Wang
Main category: cs.LG
TL;DR: BioArc is a framework that uses Neural Architecture Search to automatically discover optimal architectures for biological foundation models, moving beyond intuition-driven design to principled, automated architecture discovery for biological data.
Details
Motivation: Existing biological foundation models simply adopt architectures from general AI domains without considering unique biological data properties, leading to suboptimal performance in capturing long-range dependencies, sparse information, and complex biological "grammars."Method: Leverages Neural Architecture Search (NAS) to systematically explore vast architecture design spaces, evaluates architectures across multiple biological modalities, analyzes interplay between architecture, tokenization, and training strategies, and proposes architecture prediction methods for new tasks.
Result: Identifies novel, high-performance architectures and distills empirical design principles for biological foundation models, providing effective architecture prediction methods for new biological tasks.
Conclusion: BioArc provides a foundational resource and principled methodology to guide creation of next-generation task-specific and foundation models for biology, moving beyond intuition-driven architecture design.
Abstract: Foundation models have revolutionized various fields such as natural language processing (NLP) and computer vision (CV). While efforts have been made to transfer the success of the foundation models in general AI domains to biology, existing works focus on directly adopting the existing foundation model architectures from general machine learning domains without a systematic design considering the unique physicochemical and structural properties of each biological data modality. This leads to suboptimal performance, as these repurposed architectures struggle to capture the long-range dependencies, sparse information, and complex underlying ``grammars’’ inherent to biological data. To address this gap, we introduce BioArc, a novel framework designed to move beyond intuition-driven architecture design towards principled, automated architecture discovery for biological foundation models. Leveraging Neural Architecture Search (NAS), BioArc systematically explores a vast architecture design space, evaluating architectures across multiple biological modalities while rigorously analyzing the interplay between architecture, tokenization, and training strategies. This large-scale analysis identifies novel, high-performance architectures, allowing us to distill a set of empirical design principles to guide future model development. Furthermore, to make the best of this set of discovered principled architectures, we propose and compare several architecture prediction methods that effectively and efficiently predict optimal architectures for new biological tasks. Overall, our work provides a foundational resource and a principled methodology to guide the creation of the next generation of task-specific and foundation models for biology.
[633] A unified framework for geometry-independent operator learning in cardiac electrophysiology simulations
Bei Zhou, Cesare Corrado, Shuang Qian, Maximilian Balmus, Angela W. C. Lee, Cristobal Rodero, Marco J. W. Gotte, Luuk H. G. A. Hopman, Mengyun Qiao, Steven Niederer
Main category: cs.LG
TL;DR: A geometry-independent operator-learning framework predicts local activation time fields across diverse left atrial anatomies with near-instantaneous inference, enabling real-time clinical use.
Details
Motivation: Biopsychically detailed simulations for atrial electrical activation mapping are computationally intensive, limiting real-time clinical use and population-scale analyses.Method: Created a dataset of 308,700 simulations using GPU-accelerated electrophysiology solver across 147 patient-specific geometries. Used Universal Atrium Coordinate system for consistent representation and designed a neural operator with vision-transformer backbone to learn mapping from inputs to LAT fields.
Result: Achieved mean prediction error of 5.1 ms over 455 ms maximum simulation time, outperforming established operator-learning approaches with inference time of 0.12 ms per sample.
Conclusion: The framework establishes a general strategy for learning domain-invariant biophysical mappings across variable anatomical domains, enabling integration of computational electrophysiology into real-time and large-scale clinical workflows.
Abstract: Accurate maps of atrial electrical activation are essential for personalised treatment of arrhythmias, yet biophysically detailed simulations remain computationally intensive for real-time clinical use or population-scale analyses. Here we introduce a geometry-independent operator-learning framework that predicts local activation time (LAT) fields across diverse left atrial anatomies with near-instantaneous inference. We generated a dataset of 308,700 simulations using a GPU-accelerated electrophysiology solver, systematically varying multiple pacing sites and physiologically varied conduction properties across 147 patient-specific geometries derived from two independent clinical cohorts. All anatomical and functional data are expressed in a Universal Atrium Coordinate system, providing a consistent representation that decouples electrophysiological patterns from mesh topology. Within this coordinate space, we designed a neural operator with a vision-transformer backbone to learn the mapping from structural and electrophysiological inputs to LAT fields. With a mean prediction error of 5.1 ms over a 455 ms maximum simulation time, the model outperforms established operator-learning approaches and performs inference in 0.12 ms per sample. Our framework establishes a general strategy for learning domain-invariant biophysical mappings across variable anatomical domains and enables integration of computational electrophysiology into real-time and large-scale clinical workflows.
[634] Data-Driven Modeling and Correction of Vehicle Dynamics
Nguyen Ly, Caroline Tatsuoka, Jai Nagaraj, Jacob Levy, Fernando Palafox, David Fridovich-Keil, Hannah Lu
Main category: cs.LG
TL;DR: Data-driven framework for learning and correcting non-autonomous vehicle dynamics using two approaches: DRIPS for linearizable systems and FML for nonlinear systems with model correction.
Details
Motivation: Physics-based vehicle models are often simplified for tractability, leading to inherent model-form uncertainty. Non-autonomous dynamics with time-dependent control inputs pose challenges for learning from temporal snapshot data.Method: Reformulate vehicle dynamics via local parameterization of time-dependent inputs, creating a sequence of local parametric dynamical systems. Use DRIPS for linear surrogate models with lifted observable spaces and manifold-based operator interpolation. Use FML (deep neural network) for nonlinear systems, extended with transfer-learning-based model correction for sparse high-fidelity measurements.
Result: DRIPS offers robust and highly data-efficient learning of non-autonomous vehicle dynamics. FML provides expressive nonlinear modeling and effective correction of model-form errors under severe data scarcity, demonstrated through numerical experiments on unicycle, simplified bicycle, and slip-based bicycle models.
Conclusion: The framework successfully addresses model-form uncertainty in vehicle dynamics through complementary data-driven approaches, enabling both efficient linear modeling and expressive nonlinear correction with minimal data requirements.
Abstract: We develop a data-driven framework for learning and correcting non-autonomous vehicle dynamics. Physics-based vehicle models are often simplified for tractability and therefore exhibit inherent model-form uncertainty, motivating the need for data-driven correction. Moreover, non-autonomous dynamics are governed by time-dependent control inputs, which pose challenges in learning predictive models directly from temporal snapshot data. To address these, we reformulate the vehicle dynamics via a local parameterization of the time-dependent inputs, yielding a modified system composed of a sequence of local parametric dynamical systems. We approximate these parametric systems using two complementary approaches. First, we employ the DRIPS (dimension reduction and interpolation in parameter space) methodology to construct efficient linear surrogate models, equipped with lifted observable spaces and manifold-based operator interpolation. This enables data-efficient learning of vehicle models whose dynamics admit accurate linear representations in the lifted spaces. Second, for more strongly nonlinear systems, we employ FML (Flow Map Learning), a deep neural network approach that approximates the parametric evolution map without requiring special treatment of nonlinearities. We further extend FML with a transfer-learning-based model correction procedure, enabling the correction of misspecified prior models using only a sparse set of high-fidelity or experimental measurements, without assuming a prescribed form for the correction term. Through a suite of numerical experiments on unicycle, simplified bicycle, and slip-based bicycle models, we demonstrate that DRIPS offers robust and highly data-efficient learning of non-autonomous vehicle dynamics, while FML provides expressive nonlinear modeling and effective correction of model-form errors under severe data scarcity.
[635] FiCoTS: Fine-to-Coarse LLM-Enhanced Hierarchical Cross-Modality Interaction for Time Series Forecasting
Yafei Lyu, Hao Zhou, Lu Zhang, Xu Yang, Zhiyong Liu
Main category: cs.LG
TL;DR: FiCoTS: LLM-enhanced fine-to-coarse framework for multimodal time series forecasting using LLM-as-Enhancer paradigm to complement time series with textual understanding.
Details
Motivation: Traditional LLM-as-Predictor approaches struggle because semantic differences between time series and text make it hard for LLMs to fully understand time series data. Need to leverage LLMs' text understanding strengths while avoiding forcing them to interpret time series directly.Method: Proposes FiCoTS with LLM-as-Enhancer paradigm: 1) Token-level modality alignment via dynamic heterogeneous graph to filter noise and align time series patches with text tokens; 2) Feature-level modality interaction using global cross-attention to connect time series variables with relevant textual contexts; 3) Decision-level modality fusion with gated network to adaptively fuse results from both modalities.
Result: Extensive experiments on seven real-world benchmarks demonstrate state-of-the-art performance in multimodal time series forecasting.
Conclusion: The LLM-as-Enhancer paradigm effectively leverages LLMs’ text understanding capabilities to complement time series data, with the fine-to-coarse framework enabling comprehensive cross-modality interaction across three semantic levels for robust predictions.
Abstract: Time series forecasting is central to data analysis and web technologies. The recent success of Large Language Models (LLMs) offers significant potential for this field, especially from the cross-modality aspect. Most methods adopt an LLM-as-Predictor paradigm, using LLM as the forecasting backbone and designing modality alignment mechanisms to enable LLM to understand time series data. However, the semantic information in the two modalities of time series and text differs significantly, making it challenging for LLM to fully understand time series data. To mitigate this challenge, our work follows an LLM-as-Enhancer paradigm to fully utilize the advantage of LLM in text understanding, where LLM is only used to encode text modality to complement time series modality. Based on this paradigm, we propose FiCoTS, an LLM-enhanced fine-to-coarse framework for multimodal time series forecasting. Specifically, the framework facilitates progressive cross-modality interaction by three levels in a fine-to-coarse scheme: First, in the token-level modality alignment module, a dynamic heterogeneous graph is constructed to filter noise and align time series patches with text tokens; Second, in the feature-level modality interaction module, a global cross-attention mechanism is introduced to enable each time series variable to connect with relevant textual contexts; Third, in the decision-level modality fusion module, we design a gated network to adaptively fuse the results of the two modalities for robust predictions. These three modules work synergistically to let the two modalities interact comprehensively across three semantic levels, enabling textual information to effectively support temporal prediction. Extensive experiments on seven real-world benchmarks demonstrate that our model achieves state-of-the-art performance. The codes will be released publicly.
[636] Challenges of Heterogeneity in Big Data: A Comparative Study of Classification in Large-Scale Structured and Unstructured Domains
González Trigueros Jesús Eduardo, Alonso Sánchez Alejandro, Muñoz Rivera Emilio, Peñarán Prieto Mariana Jaqueline, Mendoza González Camila Natalia
Main category: cs.LG
TL;DR: Analysis of Big Data heterogeneity impact shows optimized linear models beat deep learning in high-dimensional numerical data, while simpler models with robust feature engineering outperform complex models in distributed text processing.
Details
Motivation: To understand how data heterogeneity ("Variety") affects classification performance across different data domains (structured vs. unstructured) and develop a framework for algorithm selection based on data nature and infrastructure constraints.Method: Dual methodology: 1) Evolutionary and Bayesian hyperparameter optimization (Genetic Algorithms, Optuna) for numerical data, 2) Distributed processing in Apache Spark for massive textual corpora, comparing various models including SVM, Logistic Regression, deep architectures, and Gradient Boosting with Transformer-based embeddings (ROBERTa) and Bayesian Target Encoding.
Result: Revealed a “complexity paradox”: optimized linear models outperformed deep architectures in high-dimensional numerical data, while in text domains, distributed fine-tuning constraints caused overfitting in complex models, making simpler models with robust feature engineering more effective.
Conclusion: Provides a unified framework for algorithm selection based on data nature and infrastructure constraints, challenging the assumption that more complex models always perform better in Big Data scenarios.
Abstract: This study analyzes the impact of heterogeneity (“Variety”) in Big Data by comparing classification strategies across structured (Epsilon) and unstructured (Rest-Mex, IMDB) domains. A dual methodology was implemented: evolutionary and Bayesian hyperparameter optimization (Genetic Algorithms, Optuna) in Python for numerical data, and distributed processing in Apache Spark for massive textual corpora. The results reveal a “complexity paradox”: in high-dimensional spaces, optimized linear models (SVM, Logistic Regression) outperformed deep architectures and Gradient Boosting. Conversely, in text-based domains, the constraints of distributed fine-tuning led to overfitting in complex models, whereas robust feature engineering – specifically Transformer-based embeddings (ROBERTa) and Bayesian Target Encoding – enabled simpler models to generalize effectively. This work provides a unified framework for algorithm selection based on data nature and infrastructure constraints.
[637] Gradient Inversion in Federated Reinforcement Learning
Shenghong He
Main category: cs.LG
TL;DR: RGIA is a novel gradient inversion attack for federated reinforcement learning that uses prior-knowledge regularization on states, rewards, and transition dynamics to successfully reconstruct private local data.
Details
Motivation: Federated reinforcement learning (FRL) preserves data privacy through gradient sharing, but attackers can exploit shared gradients to reconstruct local training data. Unlike supervised federated learning, successful reconstruction in FRL requires generated data to match both shared gradients AND align with real environment transition dynamics.Method: Proposes Regularization Gradient Inversion Attack (RGIA) which enforces prior-knowledge-based regularization on states, rewards, and transition dynamics during optimization to ensure reconstructed data remains close to true transition distribution. The regularization term narrows solution space from broad set containing spurious solutions to constrained subset satisfying both gradient matching and true transition dynamics.
Result: Extensive experiments on control tasks and autonomous driving tasks demonstrate that RGIA can effectively constrain reconstructed data transition distributions and successfully reconstruct local private data.
Conclusion: RGIA addresses the unique challenge in FRL privacy attacks by incorporating transition dynamics constraints, enabling successful reconstruction of private data that aligns with real environment behavior while matching shared gradients.
Abstract: Federated reinforcement learning (FRL) enables distributed learning of optimal policies while preserving local data privacy through gradient sharing.However, FRL faces the risk of data privacy leaks, where attackers exploit shared gradients to reconstruct local training data.Compared to traditional supervised federated learning, successful reconstruction in FRL requires the generated data not only to match the shared gradients but also to align with real transition dynamics of the environment (i.e., aligning with the real data transition distribution).To address this issue, we propose a novel attack method called Regularization Gradient Inversion Attack (RGIA), which enforces prior-knowledge-based regularization on states, rewards, and transition dynamics during the optimization process to ensure that the reconstructed data remain close to the true transition distribution.Theoretically, we prove that the prior-knowledge-based regularization term narrows the solution space from a broad set containing spurious solutions to a constrained subset that satisfies both gradient matching and true transition dynamics.Extensive experiments on control tasks and autonomous driving tasks demonstrate that RGIA can effectively constrain reconstructed data transition distributions and thus successfully reconstruct local private data.
[638] Adversarial Signed Graph Learning with Differential Privacy
Haobin Ke, Sen Zhang, Qingqing Ye, Xun Ran, Haibo Hu
Main category: cs.LG
TL;DR: ASGL is a privacy-preserving adversarial signed graph learning method that achieves node-level differential privacy while maintaining high utility for signed graphs with positive/negative edges.
Details
Motivation: Existing differential privacy methods for unsigned graphs are not suitable for signed graphs because edge perturbation causes cascading errors in sign inference under balance theory, and gradient perturbation has high sensitivity due to node interdependence and sign flip polarity changes.Method: Decomposes signed graphs into positive/negative subgraphs, designs gradient-perturbed adversarial module to approximate signed connectivity distribution, uses constrained breadth-first search tree strategy fused with balance theory to identify edge signs between generated node pairs, and implements gradient decoupling to reduce sensitivity.
Result: Extensive experiments on real-world datasets show ASGL achieves favorable privacy-utility trade-offs across multiple downstream tasks.
Conclusion: ASGL effectively addresses privacy concerns in signed graph learning by combining adversarial learning with differential privacy, overcoming limitations of existing methods through subgraph separation and gradient decoupling strategies.
Abstract: Signed graphs with positive and negative edges can model complex relationships in social networks. Leveraging on balance theory that deduces edge signs from multi-hop node pairs, signed graph learning can generate node embeddings that preserve both structural and sign information. However, training on sensitive signed graphs raises significant privacy concerns, as model parameters may leak private link information. Existing protection methods with differential privacy (DP) typically rely on edge or gradient perturbation for unsigned graph protection. Yet, they are not well-suited for signed graphs, mainly because edge perturbation tends to cascading errors in edge sign inference under balance theory, while gradient perturbation increases sensitivity due to node interdependence and gradient polarity change caused by sign flips, resulting in larger noise injection. In this paper, motivated by the robustness of adversarial learning to noisy interactions, we present ASGL, a privacy-preserving adversarial signed graph learning method that preserves high utility while achieving node-level DP. We first decompose signed graphs into positive and negative subgraphs based on edge signs, and then design a gradient-perturbed adversarial module to approximate the true signed connectivity distribution. In particular, the gradient perturbation helps mitigate cascading errors, while the subgraph separation facilitates sensitivity reduction. Further, we devise a constrained breadth-first search tree strategy that fuses with balance theory to identify the edge signs between generated node pairs. This strategy also enables gradient decoupling, thereby effectively lowering gradient sensitivity. Extensive experiments on real-world datasets show that ASGL achieves favorable privacy-utility trade-offs across multiple downstream tasks.
[639] Tracing Mathematical Proficiency Through Problem-Solving Processes
Jungyang Park, Suho Kang, Jaewoo Park, Jaehong Kim, Jaewoo Shin, Seonjoon Park, Youngjae Yu
Main category: cs.LG
TL;DR: KT-PSP incorporates problem-solving processes for better knowledge tracing, and StatusKT uses a three-stage LLM pipeline to extract mathematical proficiency indicators for improved prediction and explainability.
Details
Motivation: Traditional KT methods lack explainability and only use response correctness, ignoring rich information in problem-solving processes. This limits their ability to capture multidimensional aspects of mathematical proficiency.Method: Proposes KT-PSP framework incorporating problem-solving processes, introduces KT-PSP-25 dataset, and presents StatusKT with teacher-student-teacher three-stage LLM pipeline to extract mathematical proficiency indicators as intermediate signals.
Result: Experimental results on KT-PSP-25 show StatusKT improves prediction performance of existing KT methods and provides interpretable explanations by explicitly modeling students’ mathematical proficiency.
Conclusion: Incorporating problem-solving processes and using LLMs to extract mathematical proficiency indicators significantly enhances knowledge tracing performance and explainability, enabling more effective personalized learning.
Abstract: Knowledge Tracing (KT) aims to model student’s knowledge state and predict future performance to enable personalized learning in Intelligent Tutoring Systems. However, traditional KT methods face fundamental limitations in explainability, as they rely solely on the response correctness, neglecting the rich information embedded in students’ problem-solving processes. To address this gap, we propose Knowledge Tracing Leveraging Problem-Solving Process (KT-PSP), which incorporates students’ problem-solving processes to capture the multidimensional aspects of mathematical proficiency. We also introduce KT-PSP-25, a new dataset specifically designed for the KT-PSP. Building on this, we present StatusKT, a KT framework that employs a teacher-student-teacher three-stage LLM pipeline to extract students’ MP as intermediate signals. In this pipeline, the teacher LLM first extracts problem-specific proficiency indicators, then a student LLM generates responses based on the student’s solution process, and a teacher LLM evaluates these responses to determine mastery of each indicator. The experimental results on KT-PSP-25 demonstrate that StatusKT improves the prediction performance of existing KT methods. Moreover, StatusKT provides interpretable explanations for its predictions by explicitly modeling students’ mathematical proficiency.
[640] Introducing AI-Driven IoT Energy Management Framework
Shivani Mruthyunjaya, Anandi Dutta, Kazi Sifatul Islam
Main category: cs.LG
TL;DR: Proposes a holistic IoT framework for power consumption management with contextual decision-making, proactive adaptation, and scalable structure to reduce costs and improve grid stability.
Details
Motivation: Power consumption is critical in modern life with technological reliance; reducing consumption or following predictions can lower costs and improve electrical reliability.Method: Holistic framework for IoT systems focusing on contextual decision making, proactive adaptation, and scalable structure; includes structured process for IoT systems with accuracy and interconnected development.
Result: Feasibility demonstrated through application of framework aspects; system includes long/short-term forecasting, anomaly detection, and qualitative data consideration for energy management decisions; evaluated on Power Consumption Time Series data.
Conclusion: The proposed holistic IoT framework shows feasibility for reducing power consumption and supporting grid stability through contextual decision-making and proactive adaptation.
Abstract: Power consumption has become a critical aspect of modern life due to the consistent reliance on technological advancements. Reducing power consumption or following power usage predictions can lead to lower monthly costs and improved electrical reliability. The proposal of a holistic framework to establish a foundation for IoT systems with a focus on contextual decision making, proactive adaptation, and scalable structure. A structured process for IoT systems with accuracy and interconnected development would support reducing power consumption and support grid stability. This study presents the feasibility of this proposal through the application of each aspect of the framework. This system would have long term forecasting, short term forecasting, anomaly detection, and consideration of qualitative data with any energy management decisions taken. Performance was evaluated on Power Consumption Time Series data to display the direct application of the framework.
[641] Adaptive prediction theory combining offline and online learning
Haizheng Li, Lei Guo
Main category: cs.LG
TL;DR: Theoretical analysis of a two-stage learning framework combining offline learning and online adaptation for nonlinear stochastic dynamical systems, showing superior prediction performance over purely offline or online methods.
Details
Motivation: Real-world intelligence systems operate with correlated, non-stationary data combining offline learning and online adaptation, but this has rarely been investigated theoretically. The paper aims to provide theoretical foundations for such hybrid approaches.Method: Two-stage framework: 1) Offline phase with nonlinear-least-squares estimation and generalization error analysis using Kullback-Leibler divergence for distribution shift; 2) Online phase with meta-LMS prediction algorithm to handle parameter drift in target systems.
Result: The two-stage framework demonstrates superior prediction performances compared with purely offline or online methods. Both theoretical guarantees and empirical studies support the findings.
Conclusion: Integrating offline learning with online adaptation provides a robust approach for real-world systems with correlated, non-stationary data, offering theoretical foundations and practical benefits for prediction tasks in nonlinear stochastic dynamical systems.
Abstract: Real-world intelligence systems usually operate by combining offline learning and online adaptation with highly correlated and non-stationary system data or signals, which, however, has rarely been investigated theoretically in the literature. This paper initiates a theoretical investigation on the prediction performance of a two-stage learning framework combining offline and online algorithms for a class of nonlinear stochastic dynamical systems. For the offline-learning phase, we establish an upper bound on the generalization error for approximate nonlinear-least-squares estimation under general datasets with strong correlation and distribution shift, leveraging the Kullback-Leibler divergence to quantify the distributional discrepancies. For the online-adaptation phase, we address, on the basis of the offline-trained model, the possible uncertain parameter drift in real-world target systems by proposing a meta-LMS prediction algorithm. This two-stage framework, integrating offline learning with online adaptation, demonstrates superior prediction performances compared with either purely offline or online methods. Both theoretical guarantees and empirical studies are provided.
[642] Provable Memory Efficient Self-Play Algorithm for Model-free Reinforcement Learning
Na Li, Yuchen Jiao, Hangguan Shan, Shefeng Yan
Main category: cs.LG
TL;DR: ME-Nash-QL is a model-free self-play algorithm for two-player zero-sum Markov games that achieves superior space complexity, sample complexity for long horizons, computational complexity, and burn-in cost compared to existing methods.
Details
Motivation: Existing MARL theoretical studies suffer from multiple obstacles: memory inefficiency, high sample complexity dependence on horizon/state space, high computational complexity, non-Markov policies, non-Nash policies, and high burn-in costs. The authors aim to address these limitations.Method: Memory-Efficient Nash Q-Learning (ME-Nash-QL) - a model-free self-play algorithm designed for two-player zero-sum Markov games. It uses Q-learning approach with memory-efficient implementation.
Result: ME-Nash-QL achieves: 1) Space complexity O(SABH) and sample complexity Õ(H⁴SAB/ε²); 2) Lowest computational complexity O(T·poly(AB)); 3) Best burn-in cost O(SAB·poly(H)). It outperforms existing algorithms in multiple metrics.
Conclusion: ME-Nash-QL represents significant progress in MARL theory by simultaneously addressing multiple limitations of existing methods, particularly excelling in memory efficiency, computational efficiency, and burn-in cost for two-player zero-sum Markov games.
Abstract: The thriving field of multi-agent reinforcement learning (MARL) studies how a group of interacting agents make decisions autonomously in a shared dynamic environment. Existing theoretical studies in this area suffer from at least two of the following obstacles: memory inefficiency, the heavy dependence of sample complexity on the long horizon and the large state space, the high computational complexity, non-Markov policy, non-Nash policy, and high burn-in cost. In this work, we take a step towards settling this problem by designing a model-free self-play algorithm \emph{Memory-Efficient Nash Q-Learning (ME-Nash-QL)} for two-player zero-sum Markov games, which is a specific setting of MARL. ME-Nash-QL is proven to enjoy the following merits. First, it can output an $\varepsilon$-approximate Nash policy with space complexity $O(SABH)$ and sample complexity $\widetilde{O}(H^4SAB/\varepsilon^2)$, where $S$ is the number of states, ${A, B}$ is the number of actions for two players, and $H$ is the horizon length. It outperforms existing algorithms in terms of space complexity for tabular cases, and in terms of sample complexity for long horizons, i.e., when $\min{A, B}\ll H^2$. Second, ME-Nash-QL achieves the lowest computational complexity $O(T\mathrm{poly}(AB))$ while preserving Markov policies, where $T$ is the number of samples. Third, ME-Nash-QL also achieves the best burn-in cost $O(SAB,\mathrm{poly}(H))$, whereas previous algorithms have a burn-in cost of at least $O(S^3 AB,\mathrm{poly}(H))$ to attain the same level of sample complexity with ours.
[643] Flexible Swarm Learning May Outpace Foundation Models in Essential Tasks
Moein E. Samadi, Andreas Schuppert
Main category: cs.LG
TL;DR: The paper argues that monolithic foundation models face conceptual limits in dynamic environments due to the curse of dimensionality, and proposes decentralized small agent networks (SANs) as a superior alternative for self-adaptive decision-making in complex systems like intensive care medicine.
Details
Motivation: Foundation models show only modest gains in real-world dynamic domains like intensive care medicine, where adapting to complex, evolving systems with strong interactions and shared side effects is challenging. There's a need for self-adaptive AI models that work with minimal data and limited mechanistic knowledge before AI assumes broader decision-making roles.Method: Proposes a decentralized architecture of interacting small agent networks (SANs) where each agent represents specialized substructures covering only subsets of system functions. Uses swarm-learning in diverse swarms to enable self-adaptive SANs, drawing on mathematical results about SAN learning behavior and evidence from existing applications.
Result: Argues that swarm-learning in diverse swarms can enable self-adaptive SANs to deliver superior decision-making in dynamic environments compared to monolithic foundation models, though with reduced reproducibility in detail.
Conclusion: Monolithic foundation models face fundamental barriers in dynamic environments due to the curse of dimensionality, while decentralized SAN architectures offer a promising alternative for self-adaptive decision-making in complex systems, though AI should demonstrate clear superiority in these settings before assuming broader decision-making roles.
Abstract: Foundation models have rapidly advanced AI, raising the question of whether their decisions will ultimately surpass human strategies in real-world domains. The exponential, and possibly super-exponential, pace of AI development makes such analysis elusive. Nevertheless, many application areas that matter for daily life and society show only modest gains so far; a prominent case is diagnosing and treating dynamically evolving disease in intensive care. The common challenge is adapting complex systems to dynamic environments. Effective strategies must optimize outcomes in systems composed of strongly interacting functions while avoiding shared side effects; this requires reliable, self-adaptive modeling. These tasks align with building digital twins of highly complex systems whose mechanisms are not fully or quantitatively understood. It is therefore essential to develop methods for self-adapting AI models with minimal data and limited mechanistic knowledge. As this challenge extends beyond medicine, AI should demonstrate clear superiority in these settings before assuming broader decision-making roles. We identify the curse of dimensionality as a fundamental barrier to efficient self-adaptation and argue that monolithic foundation models face conceptual limits in overcoming it. As an alternative, we propose a decentralized architecture of interacting small agent networks (SANs). We focus on agents representing the specialized substructure of the system, where each agent covers only a subset of the full system functions. Drawing on mathematical results on the learning behavior of SANs and evidence from existing applications, we argue that swarm-learning in diverse swarms can enable self-adaptive SANs to deliver superior decision-making in dynamic environments compared with monolithic foundation models, though at the cost of reduced reproducibility in detail.
[644] Sample-Efficient Tabular Self-Play for Offline Robust Reinforcement Learning
Na Li, Zewu Zheng, Wei Ni, Hangguan Shan, Wenjie Zhang, Xinyu Li
Main category: cs.LG
TL;DR: Proposes RTZ-VI-LCB, a model-based algorithm for offline robust two-player zero-sum Markov games with optimal sample complexity guarantees under partial coverage and environmental uncertainty.
Details
Motivation: Multi-agent reinforcement learning needs robust policies to handle sim-to-real gaps due to environmental uncertainties. Existing methods lack optimal sample complexity guarantees for offline robust two-player zero-sum Markov games.Method: RTZ-VI-LCB algorithm combines optimistic robust value iteration with data-driven Bernstein-style penalty term for robust value estimation, accounting for distribution shifts in historical dataset.
Result: Achieves near-optimal sample complexity guarantees under partial coverage and environmental uncertainty. Establishes information-theoretic lower bound confirming algorithm’s optimality regarding state and action spaces.
Conclusion: RTZ-VI-LCB is the first to attain optimal sample complexity for offline robust two-player zero-sum Markov games, setting a new benchmark and validated experimentally.
Abstract: Multi-agent reinforcement learning (MARL), as a thriving field, explores how multiple agents independently make decisions in a shared dynamic environment. Due to environmental uncertainties, policies in MARL must remain robust to tackle the sim-to-real gap. We focus on robust two-player zero-sum Markov games (TZMGs) in offline settings, specifically on tabular robust TZMGs (RTZMGs). We propose a model-based algorithm (\textit{RTZ-VI-LCB}) for offline RTZMGs, which is optimistic robust value iteration combined with a data-driven Bernstein-style penalty term for robust value estimation. By accounting for distribution shifts in the historical dataset, the proposed algorithm establishes near-optimal sample complexity guarantees under partial coverage and environmental uncertainty. An information-theoretic lower bound is developed to confirm the tightness of our algorithm’s sample complexity, which is optimal regarding both state and action spaces. To the best of our knowledge, RTZ-VI-LCB is the first to attain this optimality, sets a new benchmark for offline RTZMGs, and is validated experimentally.
[645] Learning Causal States Under Partial Observability and Perturbation
Na Li, Hangguan Shan, Wei Ni, Wenjie Zhang, Xinyu Li, Yamin Wang
Main category: cs.LG
TL;DR: CaDiff is a framework that enhances RL algorithms by uncovering causal structure in perturbed and partially observable MDPs using asynchronous diffusion models and bisimulation metrics, achieving 14.18%+ performance gains.
Details
Motivation: Existing RL methods fail to handle both perturbations and partial observability simultaneously in P²OMDPs, creating a need for approaches that can uncover underlying causal structure from noisy, incomplete observations.Method: Proposes CaDiff framework with: 1) Asynchronous Diffusion Model (ADM) enabling forward/reverse processes with different steps to interpret perturbations as suppressible noise, 2) New bisimulation metric to quantify similarity between partial observations and causal states, 3) Theoretical guarantees via value function approximation error bounds.
Result: Experiments on Roboschool tasks show CaDiff enhances returns by at least 14.18% compared to baselines, demonstrating practical effectiveness in handling perturbed partial observability.
Conclusion: CaDiff is the first framework to approximate causal states using diffusion models with both theoretical rigor and practical effectiveness, successfully addressing the dual challenge of perturbations and partial observability in RL.
Abstract: A critical challenge for reinforcement learning (RL) is making decisions based on incomplete and noisy observations, especially in perturbed and partially observable Markov decision processes (P$^2$OMDPs). Existing methods fail to mitigate perturbations while addressing partial observability. We propose \textit{Causal State Representation under Asynchronous Diffusion Model (CaDiff)}, a framework that enhances any RL algorithm by uncovering the underlying causal structure of P$^2$OMDPs. This is achieved by incorporating a novel asynchronous diffusion model (ADM) and a new bisimulation metric. ADM enables forward and reverse processes with different numbers of steps, thus interpreting the perturbation of P$^2$OMDP as part of the noise suppressed through diffusion. The bisimulation metric quantifies the similarity between partially observable environments and their causal counterparts. Moreover, we establish the theoretical guarantee of CaDiff by deriving an upper bound for the value function approximation errors between perturbed observations and denoised causal states, reflecting a principled trade-off between approximation errors of reward and transition-model. Experiments on Roboschool tasks show that CaDiff enhances returns by at least 14.18% compared to baselines. CaDiff is the first framework that approximates causal states using diffusion models with both theoretical rigor and practicality.
[646] S^2-KD: Semantic-Spectral Knowledge Distillation Spatiotemporal Forecasting
Wenshuo Wang, Yaomin Shen, Yingjie Tan, Yihao Chen
Main category: cs.LG
TL;DR: S^2-KD is a knowledge distillation framework that combines semantic reasoning from multimodal teachers with spectral representations to create lightweight spatiotemporal forecasting models that outperform state-of-the-art methods.
Details
Motivation: Existing knowledge distillation methods for spatiotemporal forecasting focus only on pixel-level signals and spectral properties, missing the rich semantic and causal context behind visual patterns, which limits their effectiveness in complex scenarios.Method: Train a privileged multimodal teacher that uses textual narratives from a Large Multimodal Model to reason about event causes while decoupling spectral components in its latent space, then distill this unified semantic-spectral knowledge into a vision-only student using a novel distillation objective.
Result: S^2-KD significantly boosts performance of simple student models on benchmarks like WeatherBench and TaxiBJ+, enabling them to outperform state-of-the-art methods, especially in long-horizon and complex non-stationary scenarios.
Conclusion: By unifying semantic priors with spectral representations, S^2-KD creates lightweight forecasting models that achieve both spectral accuracy and semantic coherence without requiring textual input or architectural overhead at inference.
Abstract: Spatiotemporal forecasting often relies on computationally intensive models to capture complex dynamics. Knowledge distillation (KD) has emerged as a key technique for creating lightweight student models, with recent advances like frequency-aware KD successfully preserving spectral properties (i.e., high-frequency details and low-frequency trends). However, these methods are fundamentally constrained by operating on pixel-level signals, leaving them blind to the rich semantic and causal context behind the visual patterns. To overcome this limitation, we introduce S^2-KD, a novel framework that unifies Semantic priors with Spectral representations for distillation. Our approach begins by training a privileged, multimodal teacher model. This teacher leverages textual narratives from a Large Multimodal Model (LMM) to reason about the underlying causes of events, while its architecture simultaneously decouples spectral components in its latent space. The core of our framework is a new distillation objective that transfers this unified semantic-spectral knowledge into a lightweight, vision-only student. Consequently, the student learns to make predictions that are not only spectrally accurate but also semantically coherent, without requiring any textual input or architectural overhead at inference. Extensive experiments on benchmarks like WeatherBench and TaxiBJ+ show that S^2-KD significantly boosts the performance of simple student models, enabling them to outperform state-of-the-art methods, particularly in long-horizon and complex non-stationary scenarios.
[647] An Empirical Study on the Effectiveness of Incorporating Offline RL As Online RL Subroutines
Jianhai Su, Jinzhu Luo, Qi Zhang
Main category: cs.LG
TL;DR: Offline RL algorithms can be used as subroutines in online RL by repurposing historical interactions as offline data, with effectiveness varying by task and requiring specialized techniques.
Details
Motivation: To improve online reinforcement learning efficiency by leveraging offline RL algorithms as subroutines, using the agent's own historical interactions as offline datasets.Method: A framework that incorporates offline RL algorithms into online RL, with variants like final policy recommendation and online fine-tuning, plus specialized techniques to enhance effectiveness.
Result: 1) Framework effectiveness strongly depends on task nature, 2) Proposed techniques significantly improve effectiveness, 3) Existing online fine-tuning methods are generally ineffective.
Conclusion: Offline RL can be effectively integrated into online RL via historical data repurposing, but requires task-specific approaches and better online fine-tuning methods.
Abstract: We take the novel perspective of incorporating offline RL algorithms as subroutines of tabula rasa online RL. This is feasible because an online learning agent can repurpose its historical interactions as offline dataset. We formalize this idea into a framework that accommodates several variants of offline RL incorporation such as final policy recommendation and online fine-tuning. We further introduce convenient techniques to improve its effectiveness in enhancing online learning efficiency. Our extensive and systematic empirical analyses show that 1) the effectiveness of the proposed framework depends strongly on the nature of the task, 2) our proposed techniques greatly enhance its effectiveness, and 3) existing online fine-tuning methods are overall ineffective, calling for more research therein.
[648] Efficient and Programmable Exploration of Synthesizable Chemical Space
Shitong Luo, Connor W. Coley
Main category: cs.LG
TL;DR: PrexSyn is a transformer-based model for generating synthesizable molecules with desired properties, trained on billion-scale synthesizable pathway data, enabling programmable molecular optimization with high efficiency.
Details
Motivation: The constrained nature of synthesizable chemical space makes it difficult to sample molecules that are both synthetically accessible and possess desired properties. Existing methods struggle with balancing synthetic accessibility with property optimization.Method: Decoder-only transformer trained on billion-scale datastream of synthesizable pathways paired with molecular properties, using a real-time, high-throughput C++-based data generation engine. Learns property-pathway mappings to generate molecules satisfying single or composite property queries.
Result: PrexSyn achieves near-perfect reconstruction of synthesizable chemical space with high inference speed, learns property-pathway associations, and can generate molecules satisfying complex property queries. Outperforms synthesis-agnostic baselines in molecular optimization efficiency.
Conclusion: PrexSyn sets new state-of-the-art in synthesizable chemical space coverage, molecular sampling efficiency, and inference speed, pushing the frontier of synthesizable molecular design as a powerful general-purpose optimization tool.
Abstract: The constrained nature of synthesizable chemical space poses a significant challenge for sampling molecules that are both synthetically accessible and possess desired properties. In this work, we present PrexSyn, an efficient and programmable model for molecular discovery within synthesizable chemical space. PrexSyn is based on a decoder-only transformer trained on a billion-scale datastream of synthesizable pathways paired with molecular properties, enabled by a real-time, high-throughput C++-based data generation engine. The large-scale training data allows PrexSyn to reconstruct the synthesizable chemical space nearly perfectly at a high inference speed and learn the association between properties and synthesizable molecules. Based on its learned property-pathway mappings, PrexSyn can generate synthesizable molecules that satisfy not only single-property conditions but also composite property queries joined by logical operators, thereby allowing users to ``program’’ generation objectives. Moreover, by exploiting this property-based querying capability, PrexSyn can efficiently optimize molecules against black-box oracle functions via iterative query refinement, achieving higher sampling efficiency than even synthesis-agnostic baselines, making PrexSyn a powerful general-purpose molecular optimization tool. Overall, PrexSyn pushes the frontier of synthesizable molecular design by setting a new state of the art in synthesizable chemical space coverage, molecular sampling efficiency, and inference speed.
[649] Solving Neural Min-Max Games: The Role of Architecture, Initialization & Dynamics
Deep Patel, Emmanouil-Vasileios Vlatakis-Gkaragkounis
Main category: cs.LG
TL;DR: The paper explains why gradient methods converge in non-convex min-max games (like adversarial training) by showing hidden convexity emerges under overparameterization, providing first global convergence guarantees for two-layer neural network games.
Details
Motivation: Many important applications (adversarial training, AI alignment, robust optimization) are framed as zero-sum games between neural nets. While these games have non-convex non-concave objectives, empirical evidence shows gradient methods often converge, suggesting hidden geometric structure that needs theoretical explanation.Method: Two technical approaches: (1) Derive novel path-length bound for alternating gradient descent-ascent in min-max games, and (2) Show hidden convex-concave geometry reduces to two-sided Polyak-Łojasiewicz condition with high probability under overparameterization using random matrix theory tools.
Result: First global convergence guarantees to von Neumann-Nash equilibria for non-convex min-max games involving two-layer neural networks, with sufficient conditions spanning initialization, training dynamics, and network width.
Conclusion: Hidden convexity emerges in overparameterized neural network games, explaining empirical convergence and providing theoretical foundation for gradient methods in adversarial training and related applications.
Abstract: Many emerging applications - such as adversarial training, AI alignment, and robust optimization - can be framed as zero-sum games between neural nets, with von Neumann-Nash equilibria (NE) capturing the desirable system behavior. While such games often involve non-convex non-concave objectives, empirical evidence shows that simple gradient methods frequently converge, suggesting a hidden geometric structure. In this paper, we provide a theoretical framework that explains this phenomenon through the lens of hidden convexity and overparameterization. We identify sufficient conditions - spanning initialization, training dynamics, and network width - that guarantee global convergence to a NE in a broad class of non-convex min-max games. To our knowledge, this is the first such result for games that involve two-layer neural networks. Technically, our approach is twofold: (a) we derive a novel path-length bound for the alternating gradient descent-ascent scheme in min-max games; and (b) we show that the reduction from a hidden convex-concave geometry to two-sided Polyak-Łojasiewicz (PŁ) min-max condition hold with high probability under overparameterization, using tools from random matrix theory.
[650] From Coefficients to Directions: Rethinking Model Merging with Directional Alignment
Zhikang Chen, Sen Cui, Deheng Ye, Min Zhang, Gang Niu, Yu Zhang, Masashi Sugiyama, Tingting Zhu
Main category: cs.LG
TL;DR: This paper introduces Merging with Directional Alignment (MDA), a geometric framework that addresses directional inconsistencies in model merging by aligning parameter and feature space directions, improving structural coherence and performance.
Details
Motivation: Current model merging approaches focus on parameter decomposition or coefficient optimization but overlook directional alignment in both parameter and feature spaces. Naïve merging causes inconsistencies in dominant parameter directions and disrupts structural coherence, while coefficient optimization assumes compatible feature-space directions that may not exist due to Neural Collapse patterns differing across independently trained models.Method: The paper introduces Merging with Directional Alignment (MDA), a unified geometric framework that aligns directional structures consistently in both parameter and feature spaces. This approach addresses the limitations of previous methods by considering directional information rather than just parameter values or coefficients.
Result: The analysis shows that directional alignment improves structural coherence. Extensive experiments across benchmarks, model scales, and task configurations validate the effectiveness of the approach, demonstrating superior performance compared to previous merging methods.
Conclusion: Directional alignment is crucial for effective model merging. The proposed MDA framework successfully addresses directional inconsistencies in both parameter and feature spaces, leading to improved structural coherence and better performance across diverse tasks and model configurations.
Abstract: Model merging has emerged as a practical paradigm for integrating multiple independently trained models into a single model without joint retraining. Previous studies have demonstrated the effectiveness of combining parameters through strategies such as parameter decomposition, coefficient optimization, and subspace learning, significantly reducing the need for expensive joint training and achieving strong empirical performance across diverse tasks. However, these approaches predominantly treat merging as a problem of parameter space decomposition or fusion coefficient optimization, while overlooking the critical role of directional information in both parameter and feature spaces. In practice, naïve merging introduces inconsistencies in dominant parameter directions and disrupts structural coherence across models, which can degrade performance. Moreover, coefficient-based optimization methods implicitly assume compatible feature-space directions across models. However, Neural Collapse indicates that class features follow structured directional patterns, which may differ across independently trained models, making coefficient optimization alone insufficient. In this work, we emphasize the importance of \emph{directional alignment} and introduce a unified geometric framework, \emph{Merging with Directional Alignment} (\method{}), which aligns directional structures consistently in both the parameter and feature spaces. Our analysis shows that directional alignment improves structural coherence, and extensive experiments across benchmarks, model scales, and task configurations further validate the effectiveness of our approach.
[651] SelfAI: Building a Self-Training AI System with LLM Agents
Xiao Wu, Ting-Zhu Huang, Liang-Jian Deng, Xiaobing Yu, Yu Zhong, Shangqi Deng, Ufaq Khan, Jianghao Wu, Xiaofeng Liu, Imran Razzak, Xiaojun Chang, Yutong Xie
Main category: cs.LG
TL;DR: SelfAI is a multi-agent platform for autonomous scientific discovery that integrates user interaction, cognitive planning with optimal stopping, and experiment management to improve efficiency and human collaboration.
Details
Motivation: Current autonomous discovery systems are limited to narrow domains, lack real-time researcher interaction, and have no principled stopping mechanisms, leading to inefficiencies, reproducibility issues, and underutilized human expertise.Method: SelfAI uses three agents: User Agent for translating research objectives into configurations, Cognitive Agent with LLMs and optimal stopping criteria for hyperparameter search refinement, and Experiment Manager for orchestrating parallel, fault-tolerant workflows across hardware with structured knowledge base.
Result: SelfAI achieves strong performance across regression, NLP, computer vision, scientific computing, medical imaging, and drug discovery benchmarks, reduces redundant trials compared to Bayesian optimization and LLM baselines, and enables seamless human interaction.
Conclusion: SelfAI addresses key limitations in autonomous scientific discovery by providing a general multi-agent platform with optimal stopping, efficient resource utilization, and improved human-AI collaboration through novel evaluation metrics.
Abstract: Recent work on autonomous scientific discovery has leveraged LLM-based agents to integrate problem specification, experiment planning, and execution into end-to-end systems. However, these frameworks are often confined to narrow application domains, offer limited real-time interaction with researchers, and lack principled mechanisms for determining when to halt exploration, resulting in inefficiencies, reproducibility challenges, and under-utilized human expertise. To address these gaps, we propose \textit{SelfAI}, a general multi-agent platform that combines a User Agent for translating high-level research objectives into standardized experimental configurations, a Cognitive Agent powered by LLMs with optimal stopping criteria to iteratively refine hyperparameter searches, and an Experiment Manager responsible for orchestrating parallel, fault-tolerant training workflows across heterogeneous hardware while maintaining a structured knowledge base for continuous feedback. We further introduce two novel evaluation metrics, Score and $\text{AUP}_D$, to quantify discovery efficiency and search diversity. Across regression, NLP, computer vision, scientific computing, medical imaging, and drug discovery benchmarks, SelfAI consistently achieves strong performance and reduces redundant trials compared to classical Bayesian optimization and LLM-based baselines, while enabling seamless interaction with human researchers.
[652] TrendGNN: Towards Understanding of Epidemics, Beliefs, and Behaviors
Mulin Tian, Ajitesh Srivastava
Main category: cs.LG
TL;DR: Proposes a graph-based forecasting framework using GNNs for interpretable epidemic signal prediction that connects behavior, beliefs, and observations.
Details
Motivation: Current epidemic forecasting uses simple mechanistic models or black-box deep learning without interpretability, making it hard to understand mechanisms and predict intervention impacts. Need interpretable forecasting of belief and behavior signals.Method: Constructs graph of interrelated signals based on trend similarity, then applies graph neural networks (GNNs) for prediction. Enables interpretable analysis by revealing predictable signals and important relationships.
Result: Framework provides interpretable forecasting by identifying which signals are more predictable and which relationships contribute most to forecasting accuracy.
Conclusion: Method offers early steps toward interpretable modeling for domains with multiple interdependent signals, with implications for building future simulation models integrating behavior, beliefs, and observations.
Abstract: Epidemic outcomes have a complex interplay with human behavior and beliefs. Most of the forecasting literature has focused on the task of predicting epidemic signals using simple mechanistic models or black-box models, such as deep transformers, that ingest all available signals without offering interpretability. However, to better understand the mechanisms and predict the impact of interventions, we need the ability to forecast signals associated with beliefs and behaviors in an interpretable manner. In this work, we propose a graph-based forecasting framework that first constructs a graph of interrelated signals based on trend similarity, and then applies graph neural networks (GNNs) for prediction. This approach enables interpretable analysis by revealing which signals are more predictable and which relationships contribute most to forecasting accuracy. We believe our method provides early steps towards a framework for interpretable modeling in domains with multiple potentially interdependent signals, with implications for building future simulation models that integrate behavior, beliefs, and observations.
[653] Privacy-Preserving Generative Modeling and Clinical Validation of Longitudinal Health Records for Chronic Disease
Benjamin D. Ballyk, Ankit Gupta, Sujay Konda, Kavitha Subramanian, Chris Landon, Ahmed Ammar Naseer, Georg Maierhofer, Sumanth Swaminathan, Vasudevan Venkateshwaran
Main category: cs.LG
TL;DR: Enhanced time-series generative model for longitudinal clinical data with privacy guarantees, outperforming existing models on privacy-utility tradeoff.
Details
Motivation: Data privacy regulations limit access to clinical records for ML model training, while current synthetic data methods struggle with time-series data or lack formal privacy guarantees.Method: Enhanced state-of-the-art time-series generative model (TimeGAN) to handle longitudinal clinical data with quantifiable privacy safeguards, creating both non-private (Augmented TimeGAN) and private (DP-TimeGAN) versions.
Result: Non-private model outperforms transformer- and flow-based models on statistical metrics; private model achieves mean authenticity of 0.778 on CKD dataset, outperforming SOTA on privacy-utility frontier. Both models perform comparably to real data in clinician evaluations.
Conclusion: The approach provides robust synthetic data for developing models for complex chronic conditions without compromising data privacy, enabling ML applications in healthcare while respecting regulations.
Abstract: Data privacy is a critical challenge in modern medical workflows as the adoption of electronic patient records has grown rapidly. Stringent data protection regulations limit access to clinical records for training and integrating machine learning models that have shown promise in improving diagnostic accuracy and personalized care outcomes. Synthetic data offers a promising alternative; however, current generative models either struggle with time-series data or lack formal privacy guaranties. In this paper, we enhance a state-of-the-art time-series generative model to better handle longitudinal clinical data while incorporating quantifiable privacy safeguards. Using real data from chronic kidney disease and ICU patients, we evaluate our method through statistical tests, a Train-on-Synthetic-Test-on-Real (TSTR) setup, and expert clinical review. Our non-private model (Augmented TimeGAN) outperforms transformer- and flow-based models on statistical metrics in several datasets, while our private model (DP-TimeGAN) maintains a mean authenticity of 0.778 on the CKD dataset, outperforming existing state-of-the-art models on the privacy-utility frontier. Both models achieve performance comparable to real data in clinician evaluations, providing robust input data necessary for developing models for complex chronic conditions without compromising data privacy.
[654] FairMT: Fairness for Heterogeneous Multi-Task Learning
Guanyu Hu, Tangzheng Lian, Na Yan, Dimitrios Kollias, Xinyu Yang, Oya Celiktutan, Siyang Song, Zeyu Fu
Main category: cs.LG
TL;DR: FairMT is a unified fairness-aware multi-task learning framework that handles heterogeneous tasks (classification, detection, regression) with incomplete supervision, using asymmetric fairness constraints and joint optimization to improve fairness without degrading utility.
Details
Motivation: Existing fairness methods are limited to classification tasks and don't extend to continuous outputs, making unified fairness objectives difficult. Current multi-task learning optimization is structurally misaligned with fairness - constraining only shared representations allows task heads to absorb bias, leading to uncontrolled task-specific disparities. Most fairness approaches treat fairness as zero-sum with utility, enforcing symmetric constraints that achieve parity by degrading well-served groups.Method: FairMT introduces an Asymmetric Heterogeneous Fairness Constraint Aggregation mechanism that consolidates task-dependent asymmetric violations into a unified fairness constraint. It uses primal-dual formulation for joint optimization of utility and fairness, with a head-aware multi-objective optimization proxy that provides tractable descent geometry accounting for head-induced anisotropy.
Result: Across three homogeneous and heterogeneous MTL benchmarks with diverse modalities and supervision regimes, FairMT consistently achieves substantial fairness gains while maintaining superior task utility compared to existing methods.
Conclusion: FairMT successfully addresses the limitations of existing fairness methods in multi-task learning by providing a unified framework that handles heterogeneous tasks with incomplete supervision, using asymmetric constraints to improve fairness without sacrificing utility, demonstrating effectiveness across diverse benchmarks.
Abstract: Fairness in machine learning has been extensively studied in single-task settings, while fair multi-task learning (MTL), especially with heterogeneous tasks (classification, detection, regression) and partially missing labels, remains largely unexplored. Existing fairness methods are predominantly classification-oriented and fail to extend to continuous outputs, making a unified fairness objective difficult to formulate. Further, existing MTL optimization is structurally misaligned with fairness: constraining only the shared representation, allowing task heads to absorb bias and leading to uncontrolled task-specific disparities. Finally, most work treats fairness as a zero-sum trade-off with utility, enforcing symmetric constraints that achieve parity by degrading well-served groups. We introduce FairMT, a unified fairness-aware MTL framework that accommodates all three task types under incomplete supervision. At its core is an Asymmetric Heterogeneous Fairness Constraint Aggregation mechanism, which consolidates task-dependent asymmetric violations into a unified fairness constraint. Utility and fairness are jointly optimized via a primal–dual formulation, while a head-aware multi-objective optimization proxy provides a tractable descent geometry that explicitly accounts for head-induced anisotropy. Across three homogeneous and heterogeneous MTL benchmarks encompassing diverse modalities and supervision regimes, FairMT consistently achieves substantial fairness gains while maintaining superior task utility. Code will be released upon paper acceptance.
[655] ESPO: Entropy Importance Sampling Policy Optimization
Yuepeng Sheng, Yuwei Huang, Shuman Liu, Haibo Zhang, Anxiang Zeng
Main category: cs.LG
TL;DR: ESPO is a new LLM reinforcement learning framework that uses entropy-based grouping to improve optimization granularity while maintaining training stability, achieving state-of-the-art results on math reasoning benchmarks.
Details
Motivation: Current group-based policy optimization frameworks like GRPO and GSPO face a fundamental trade-off between optimization granularity and training stability. GSPO's sequence-level optimization is robust but inefficient - it discards valid training samples (gradient underutilization) and fails to capture heterogeneous contributions of critical reasoning steps.Method: ESPO decomposes sequences into groups based on predictive entropy, enabling two key innovations: (1) Entropy-driven Importance Sampling to capture intra-sequence heterogeneity, and (2) Entropy-adaptive Clipping to dynamically allocate trust regions based on model uncertainty.
Result: ESPO accelerates convergence and achieves state-of-the-art performance on mathematical reasoning benchmarks. Notably, it improves accuracy on the challenging HMMT benchmark from 4.4% to 13.13%.
Conclusion: ESPO successfully reconciles fine-grained control with training stability in LLM reinforcement learning, addressing the limitations of existing group-based policy optimization frameworks through entropy-based sequence decomposition and adaptive optimization techniques.
Abstract: Large language model (LLM) reinforcement learning has increasingly relied on group-based policy optimization frameworks, such as GRPO and GSPO, to achieve stable fine-tuning at scale. However, a fundamental trade-off persists between optimization granularity and training stability. While GSPO improves robustness via sequence-level optimization, its monolithic treatment of sequences introduces severe inefficiencies: its conservative clipping mechanism indiscriminately discards valid training samples-a phenomenon we term gradient underutilization-and its uniform credit assignment fails to capture the heterogeneous contributions of critical reasoning steps. In this work, we propose Entropy Importance Sampling Policy Optimization (ESPO), a novel framework that reconciles fine-grained control with training stability. ESPO decomposes sequences into groups based on predictive entropy, enabling (1) Entropy-driven Importance Sampling to capture intra-sequence heterogeneity, and (2) Entropy-adaptive Clipping to dynamically allocate trust regions based on model uncertainty. Extensive experiments on mathematical reasoning benchmarks demonstrate that ESPO not only accelerates convergence but also achieves state-of-the-art performance, notably improving accuracy on the challenging HMMT benchmark from 4.4% to 13.13%.
[656] Rep3Net: An Approach Exploiting Multimodal Representation for Molecular Bioactivity Prediction
Sabrina Islam, Md. Atiqur Rahman, Md. Bakhtiar Hasan, Md. Hasanul Kabir
Main category: cs.LG
TL;DR: Rep3Net is a multimodal deep learning architecture for bioactivity prediction that combines molecular descriptors, graph-based structural information, and SMILES-based contextual embeddings to outperform traditional QSAR models.
Details
Motivation: Traditional QSAR models using molecular descriptors struggle to capture structural and contextual information in compounds, limiting their effectiveness in bioactivity prediction for early-stage drug discovery.Method: Rep3Net integrates three types of features: 1) traditional molecular descriptors, 2) graph-based representations capturing spatial and relational information, and 3) ChemBERTa-generated embeddings from SMILES strings for contextual information. These multimodal features are concatenated for bioactivity prediction.
Result: The model achieves the highest predictive performance on PARP-1 dataset compared to conventional standalone models including GCN, GAT, and XGBoost, demonstrating superior bioactivity prediction capability.
Conclusion: Rep3Net provides a scalable framework for computational screening in drug discovery by effectively combining multiple molecular representations to overcome limitations of traditional QSAR approaches.
Abstract: In early stage drug discovery, bioactivity prediction of molecules against target proteins plays a crucial role. Trdaitional QSAR models that utilizes molecular descriptor based data often struggles to predict bioactivity of molecules effectively due to its limitation in capturing structural and contextual information embedded within each compound. To address this challenge, we propose Rep3Net, a unified deep learning architecture that not only incorporates descriptor data but also includes spatial and relational information through graph-based represenation of compounds and contextual information through ChemBERTa generated embeddings from SMILES strings. Our model employing multimodal concatenated features produce reliable bioactivity prediction on Poly [ADP-ribose] polymerase 1 (PARP-1) dataset. PARP-1 is a crucial agent in DNA damage repair and has become a significant theraputic target in malignancies that depend on it for survival and growth. A comprehensive analysis and comparison with conventional standalone models including GCN, GAT, XGBoost, etc. demonstrates that our architecture achieves the highest predictive performance. In computational screening of compounds in drug discovery, our architecture provides a scalable framework for bioactivity prediction.
[657] Hyperbolic Continuous Structural Entropy for Hierarchical Clustering
Guangjie Zeng, Hao Peng, Angsheng Li, Li Sun, Chunyang Liu, Shengze Li, Yicheng Pan, Philip S. Yu
Main category: cs.LG
TL;DR: HypCSE: Hyperbolic Continuous Structural Entropy neural networks for structure-enhanced continuous hierarchical clustering using hyperbolic embeddings and graph structure learning.
Details
Motivation: Existing hierarchical clustering methods have two main issues: 1) Most specify dendrograms without global objectives, 2) Graph-based methods often neglect graph structure significance, optimizing on complete or static predefined graphs.Method: Map data points in hyperbolic space using hyperbolic graph neural networks, minimize approximate structural entropy on graph embeddings. Reformulate SE into function using lowest common ancestor on trees, relax into continuous SE via hyperbolic graph embeddings-partitioning trees analogy. Use graph structure learning to update graph structure during training.
Result: Extensive experiments on seven datasets demonstrate superior performance of HypCSE compared to existing methods.
Conclusion: HypCSE effectively addresses limitations of existing hierarchical clustering methods by combining hyperbolic embeddings, continuous structural entropy optimization, and adaptive graph structure learning for improved hierarchical clustering performance.
Abstract: Hierarchical clustering is a fundamental machine-learning technique for grouping data points into dendrograms. However, existing hierarchical clustering methods encounter two primary challenges: 1) Most methods specify dendrograms without a global objective. 2) Graph-based methods often neglect the significance of graph structure, optimizing objectives on complete or static predefined graphs. In this work, we propose Hyperbolic Continuous Structural Entropy neural networks, namely HypCSE, for structure-enhanced continuous hierarchical clustering. Our key idea is to map data points in the hyperbolic space and minimize the relaxed continuous structural entropy (SE) on structure-enhanced graphs. Specifically, we encode graph vertices in hyperbolic space using hyperbolic graph neural networks and minimize approximate SE defined on graph embeddings. To make the SE objective differentiable for optimization, we reformulate it into a function using the lowest common ancestor (LCA) on trees and then relax it into continuous SE (CSE) by the analogy of hyperbolic graph embeddings and partitioning trees. To ensure a graph structure that effectively captures the hierarchy of data points for CSE calculation, we employ a graph structure learning (GSL) strategy that updates the graph structure during training. Extensive experiments on seven datasets demonstrate the superior performance of HypCSE.
[658] Pushing the Boundaries of Interpretability: Incremental Enhancements to the Explainable Boosting Machine
Isara Liyanage, Uthayasanker Thayasivam
Main category: cs.LG
TL;DR: This paper enhances Explainable Boosting Machines (EBMs) with three methods: Bayesian hyperparameter optimization, custom fairness-aware objective functions, and self-supervised pre-training for cold-start scenarios, showing that tuning affects model behavior beyond just performance metrics.
Details
Motivation: Address the "black-box" problem in high-stakes AI applications by improving glassbox models like EBMs that offer both accuracy and transparency, meeting regulatory and ethical compliance demands for responsible AI.Method: Three enhancement approaches: 1) Targeted hyperparameter optimization using Bayesian methods, 2) Custom multi-objective function incorporating fairness metrics for hyperparameter optimization, 3) Novel self-supervised pre-training pipeline for cold-start scenarios. Evaluated on Adult Income, Credit Card Fraud Detection, and UCI Heart Disease datasets.
Result: While tuning yielded marginal improvements in ROC AUC, it importantly shifted the model’s decision-making behavior, demonstrating the value of multi-faceted evaluation beyond single performance scores.
Conclusion: This work advances the development of machine learning systems that are accurate, robust, equitable, and transparent, addressing critical needs for regulatory compliance and ethical AI in high-stakes domains.
Abstract: The widespread adoption of complex machine learning models in high-stakes domains has brought the “black-box” problem to the forefront of responsible AI research. This paper aims at addressing this issue by improving the Explainable Boosting Machine (EBM), a state-of-the-art glassbox model that delivers both high accuracy and complete transparency. The paper outlines three distinct enhancement methodologies: targeted hyperparameter optimization with Bayesian methods, the implementation of a custom multi-objective function for fairness for hyperparameter optimization, and a novel self-supervised pre-training pipeline for cold-start scenarios. All three methodologies are evaluated across standard benchmark datasets, including the Adult Income, Credit Card Fraud Detection, and UCI Heart Disease datasets. The analysis indicates that while the tuning process yielded marginal improvements in the primary ROC AUC metric, it led to a subtle but important shift in the model’s decision-making behavior, demonstrating the value of a multi-faceted evaluation beyond a single performance score. This work is positioned as a critical step toward developing machine learning systems that are not only accurate but also robust, equitable, and transparent, meeting the growing demands of regulatory and ethical compliance.
[659] Algorithmic Guarantees for Distilling Supervised and Offline RL Datasets
Aaryan Gupta, Rishi Saket, Aravindan Raghuveer
Main category: cs.LG
TL;DR: Novel dataset distillation algorithm for regression and offline RL that matches losses using randomly sampled regressors without model training, achieving tight O(d²) sample complexity bounds.
Details
Motivation: Dataset distillation aims to create synthetic datasets that preserve model performance while being more efficient. Current methods often require model training or have suboptimal sample complexity. The authors seek to develop an efficient algorithm with provable guarantees for both supervised regression and offline RL.Method: Algorithm matches losses on training and synthetic datasets using fixed sets of randomly sampled regressors without model training. For regression, matches MSE loss; for offline RL, matches Bellman loss (unlike previous behavioral cloning approaches). Uses O(d²) sampled regressors for d-dimensional regression.
Result: Proves tight O(d²) sample complexity bound for regression - both upper bound of Õ(d²) and matching lower bound of Ω(d²). Extends to offline RL with Bellman loss matching. Experimental validation shows performance gains and confirms theoretical guarantees.
Conclusion: First dataset distillation method for offline RL using Bellman loss matching (not behavioral cloning), leveraging both rewards and next-state information. Provides tight sample complexity analysis for regression and extends framework to RL, enabling efficient synthetic dataset generation without policy optimization.
Abstract: Given a training dataset, the goal of dataset distillation is to derive a synthetic dataset such that models trained on the latter perform as well as those trained on the training dataset. In this work, we develop and analyze an efficient dataset distillation algorithm for supervised learning, specifically regression in $\mathbb{R}^d$, based on matching the losses on the training and synthetic datasets with respect to a fixed set of randomly sampled regressors without any model training. Our first key contribution is a novel performance guarantee proving that our algorithm needs only $\tilde{O}(d^2)$ sampled regressors to derive a synthetic dataset on which the MSE loss of any bounded linear model is nearly the same as its MSE loss on the given training data. In particular, the model optimized on the synthetic data has close to minimum loss on the training data, thus performing nearly as well as the model optimized on the latter. Complementing this, we also prove a matching lower bound of $Ω(d^2)$ for the number of sampled regressors showing the tightness of our analysis. Our second contribution is to extend our algorithm to offline RL dataset distillation by matching the Bellman loss, unlike previous works which used a behavioral cloning objective. This is the first such method which leverages both, the rewards and the next state information, available in offline RL datasets, without any policy model optimization. Our algorithm generates a synthetic dataset whose Bellman loss with respect to any linear action-value predictor is close to the latter’s Bellman loss on the offline RL training dataset. Therefore, a policy associated with an action-value predictor optimized on the synthetic dataset performs nearly as well as that derived from the one optimized on the training data. We conduct experiments to validate our theoretical guarantees and observe performance gains.
[660] DQ4FairIM: Fairness-aware Influence Maximization using Deep Reinforcement Learning
Akrati Saxena, Harshith Kumar Yadav, Bart Rutten, Shashi Shekhar Jha
Main category: cs.LG
TL;DR: DQ4FairIM: A fairness-aware influence maximization method using deep reinforcement learning to ensure equitable influence spread across all communities, addressing structural inequalities in social networks.
Details
Motivation: Real-world social networks have structural inequalities (dominant majority vs. underrepresented minority groups). Traditional IM algorithms can be biased, disproportionately benefiting majority groups while marginalizing minorities, creating unfair outcomes.Method: Propose DQ4FairIM: a fairness-aware deep RL method using deep Q-learning with Structure2Vec network embedding. Formulates IM as Markov Decision Process (MDP) with maximin fairness objective that prioritizes improving outreach of least-influenced group.
Result: Achieves higher fairness while maintaining better fairness-performance trade-off than baselines. Learns effective seeding policies that generalize across problem instances (varying network size, number of seed nodes) without retraining.
Conclusion: DQ4FairIM successfully addresses fairness concerns in influence maximization, ensuring equitable influence distribution across all communities while maintaining effectiveness and generalizability.
Abstract: The Influence Maximization (IM) problem aims to select a set of seed nodes within a given budget to maximize the spread of influence in a social network. However, real-world social networks have several structural inequalities, such as dominant majority groups and underrepresented minority groups. If these inequalities are not considered while designing IM algorithms, the outcomes might be biased, disproportionately benefiting majority groups while marginalizing minorities. In this work, we address this gap by designing a fairness-aware IM method using Reinforcement Learning (RL) that ensures equitable influence outreach across all communities, regardless of protected attributes. Fairness is incorporated using a maximin fairness objective, which prioritizes improving the outreach of the least-influenced group, pushing the solution toward an equitable influence distribution. We propose a novel fairness-aware deep RL method, called DQ4FairIM, that maximizes the expected number of influenced nodes by learning an RL policy. The learnt policy ensures that minority groups formulate the IM problem as a Markov Decision Process (MDP) and use deep Q-learning, combined with the Structure2Vec network embedding, earning together with Structure2Vec network embedding to solve the MDP. We perform extensive experiments on synthetic benchmarks and real-world networks to compare our method with fairness-agnostic and fairness-aware baselines. The results show that our method achieves a higher level of fairness while maintaining a better fairness-performance trade-off than baselines. Additionally, our approach learns effective seeding policies that generalize across problem instances without retraining, such as varying the network size or the number of seed nodes.
[661] A Graph Neural Network Approach for Localized and High-Resolution Temperature Forecasting
Joud El-Shawa, Elham Bagheri, Sedef Akinli Kocak, Yalda Mohsenzadeh
Main category: cs.LG
TL;DR: Graph Neural Network framework for localized, high-resolution temperature forecasting to address heatwave prediction gaps, especially for vulnerable populations in the Global South.
Details
Motivation: Heatwaves are intensifying globally and disproportionately affect marginalized populations and the Global South, where current weather prediction models fail to capture micro-scale extremes, leaving vulnerable communities without timely early warnings.Method: Graph Neural Network framework leveraging spatial learning and efficient computation to generate localized, high-resolution temperature forecasts at multiple horizons (up to 48 hours).
Result: For Southwestern Ontario, Canada, the model achieves mean MAE of 1.93°C across 1-48h forecasts and MAE@48h of 2.93°C, using 24h input windows on the largest region.
Conclusion: While demonstrated in a data-rich context, this work provides a foundation for transfer learning approaches that could enable localized, equitable forecasts in data-limited regions of the Global South.
Abstract: Heatwaves are intensifying worldwide and are among the deadliest weather disasters. The burden falls disproportionately on marginalized populations and the Global South, where under-resourced health systems, exposure to urban heat islands, and the lack of adaptive infrastructure amplify risks. Yet current numerical weather prediction models often fail to capture micro-scale extremes, leaving the most vulnerable excluded from timely early warnings. We present a Graph Neural Network framework for localized, high-resolution temperature forecasting. By leveraging spatial learning and efficient computation, our approach generates forecasts at multiple horizons, up to 48 hours. For Southwestern Ontario, Canada, the model captures temperature patterns with a mean MAE of 1.93$^{\circ}$C across 1-48h forecasts and MAE@48h of 2.93$^{\circ}$C, evaluated using 24h input windows on the largest region. While demonstrated here in a data-rich context, this work lays the foundation for transfer learning approaches that could enable localized, equitable forecasts in data-limited regions of the Global South.
[662] List Replicable Reinforcement Learning
Bohan Zhang, Michael Chen, A. Pavan, N. V. Vinodchandran, Lin F. Yang, Ruosong Wang
Main category: cs.LG
TL;DR: The paper introduces list replicability in RL to address instability, proposing weak and strong forms where algorithms must return policies from a small list across runs, with polynomial list complexity guarantees.
Details
Motivation: Reinforcement learning algorithms are empirically unstable and sensitive to training variations, making replicability a fundamental challenge. The paper aims to formally address this issue through the concept of list replicability.Method: Introduces list replicability in PAC RL framework with weak (final policy in small list) and strong (entire policy sequence constrained) forms. Develops provably efficient tabular RL algorithm using: (1) novel planning strategy selecting actions based on lexicographic order among near-optimal choices within random tolerance threshold, (2) mechanism for testing state reachability in stochastic environments while preserving replicability.
Result: Theoretical contribution: efficient tabular RL algorithm with polynomial list complexity in states, actions, and horizon length. Achieves both weak and strong list replicability with polynomial bounds on possible policy execution traces. Empirical demonstration shows the planning strategy can enhance stability in practical RL frameworks.
Conclusion: List replicability provides a formal framework to address RL instability. The proposed algorithm guarantees polynomial list complexity, and the theoretical insights can enhance stability in practical RL applications through the novel planning strategy.
Abstract: Replicability is a fundamental challenge in reinforcement learning (RL), as RL algorithms are empirically observed to be unstable and sensitive to variations in training conditions. To formally address this issue, we study \emph{list replicability} in the Probably Approximately Correct (PAC) RL framework, where an algorithm must return a near-optimal policy that lies in a \emph{small list} of policies across different runs, with high probability. The size of this list defines the \emph{list complexity}. We introduce both weak and strong forms of list replicability: the weak form ensures that the final learned policy belongs to a small list, while the strong form further requires that the entire sequence of executed policies remains constrained. These objectives are challenging, as existing RL algorithms exhibit exponential list complexity due to their instability. Our main theoretical contribution is a provably efficient tabular RL algorithm that guarantees list replicability by ensuring the list complexity remains polynomial in the number of states, actions, and the horizon length. We further extend our techniques to achieve strong list replicability, bounding the number of possible policy execution traces polynomially with high probability. Our theoretical result is made possible by key innovations including (i) a novel planning strategy that selects actions based on lexicographic order among near-optimal choices within a randomly chosen tolerance threshold, and (ii) a mechanism for testing state reachability in stochastic environments while preserving replicability. Finally, we demonstrate that our theoretical investigation sheds light on resolving the \emph{instability} issue of RL algorithms used in practice. In particular, we show that empirically, our new planning strategy can be incorporated into practical RL frameworks to enhance their stability.
[663] Pre-Generating Multi-Difficulty PDE Data for Few-Shot Neural PDE Solvers
Naman Choudhary, Vedant Singh, Ameet Talwalkar, Nicholas Matthew Boffi, Mikhail Khodak, Tanya Marwah
Main category: cs.LG
TL;DR: Training neural PDE solvers requires costly classical solver data; difficulty transfer from low/medium to high complexity problems reduces data generation costs by 8.9x.
Details
Motivation: The main cost in learned PDE solvers comes from generating training data with classical solvers. There's a chicken-and-egg problem: harder problems (complex geometries, high Reynolds) are both more expensive for classical solvers and more likely to benefit from neural speedups.Method: Study difficulty transfer on 2D incompressible Navier-Stokes, varying task complexity along geometry (number/placement of obstacles) and physics (Reynolds number). Pre-generate many low/medium difficulty examples with classical solvers and include them in training to learn high-difficulty physics from fewer samples.
Result: By combining low and high difficulty data, can achieve same error as using only high difficulty examples while spending 8.9x less compute on pre-generating the dataset. How classical-solver compute is allocated across difficulty levels is as important as total compute.
Conclusion: Principled curation of pre-generated PDE data for neural solvers offers substantial gains. Similar to pre-training foundation models, strategically allocating classical solver compute across difficulty levels can dramatically reduce overall training costs.
Abstract: A key aspect of learned partial differential equation (PDE) solvers is that the main cost often comes from generating training data with classical solvers rather than learning the model itself. Another is that there are clear axes of difficulty–e.g., more complex geometries and higher Reynolds numbers–along which problems become (1) harder for classical solvers and thus (2) more likely to benefit from neural speedups. Towards addressing this chicken-and-egg challenge, we study difficulty transfer on 2D incompressible Navier-Stokes, systematically varying task complexity along geometry (number and placement of obstacles), physics (Reynolds number), and their combination. Similar to how it is possible to spend compute to pre-train foundation models and improve their performance on downstream tasks, we find that by classically solving (analogously pre-generating) many low and medium difficulty examples and including them in the training set, it is possible to learn high-difficulty physics from far fewer samples. Furthermore, we show that by combining low and high difficulty data, we can spend 8.9x less compute on pre-generating a dataset to achieve the same error as using only high difficulty examples. Our results highlight that how we allocate classical-solver compute across difficulty levels is as important as how much we allocate overall, and suggest substantial gains from principled curation of pre-generated PDE data for neural solvers. Our code is available at https://github.com/Naman-Choudhary-AI-ML/pregenerating-pde
[664] Non-Asymptotic Convergence of Discrete Diffusion Models: Masked and Random Walk dynamics
Giovanni Conforti, Alain Durmus, Le-Tuyet-Nhi Pham
Main category: cs.LG
TL;DR: First non-asymptotic convergence guarantees for Discrete Diffusion Models on finite and infinite discrete spaces with linear scaling complexity.
Details
Motivation: Discrete diffusion models lack rigorous theoretical analysis compared to continuous diffusion models, posing challenges due to combinatorial structure and absence of convergence guarantees.Method: Established convergence guarantees for DDMs on finite space ℤᵐᵈ and countably infinite space ℕᵈ under mild assumptions, focusing on forward masked and random walk dynamics, analyzing discrete score function monotonicity.
Result: Model complexity scales linearly up to logarithmic factors with dimension (not exponentially), making it efficiently scalable to high-dimensional data. First non-asymptotic convergence guarantees that don’t rely on boundedness of estimated score.
Conclusion: Provides rigorous theoretical foundation for discrete diffusion models with practical efficiency for high-dimensional discrete data applications.
Abstract: We investigate the theoretical underpinnings of Discrete Diffusion Models (DDMs) on discrete state spaces. Unlike in the continuous setting-where diffusion models are well understood both theoretically and empirically-the discrete case poses significant challenges due to its combinatorial structure and the lack of rigorous analysis. In this work, we establish convergence guarantees for DDMs on both the finite space $\mathbb{Z}^d_m={0,…,m-1}^d$ and the countably infinite space $\mathbb{N}^d$ under mild assumptions, focusing on forward masked and random walk dynamics. Similar to the continuous case, the backward process can be characterized by a discrete score function, whose monotonicity plays a central role in deriving the error bounds of the generated data. Notably, the complexity of our model scales linearly up to logarithmic factors, rather than exponentially, with the dimension, making it efficiently scalable to high-dimensional data. To the best of our knowledge, this study provides the first non-asymptotic convergence guarantees that do not rely on the boundedness of the estimated score-covering not only uniform noising processes on $\mathbb{Z}^d_m$ and on $\mathbb{N}^d$, but also masking-based noising dynamics.
[665] Statistical NLP for Optimization of Clinical Trial Success Prediction in Pharmaceutical R&D
Michael R. Doane
Main category: cs.LG
TL;DR: NLP-based classifier predicts clinical trial success probability in neuroscience using ClinicalTrials.gov data, with BioBERT model outperforming traditional ML methods and industry benchmarks.
Details
Motivation: Pharmaceutical R&D faces high attrition rates and enormous costs, especially in neuroscience where success rates are below 10%. Timely identification of promising clinical trials can optimize resource allocation and reduce financial risk.Method: Developed NLP-enabled probabilistic classifier using ClinicalTrials.gov data and Clinical Trial Outcome dataset. Extracted text-based features using statistical NLP techniques, integrated into non-LLM frameworks (logistic regression, gradient boosting, random forest). Built LLM-based model using BioBERT domain-specific language encoder.
Result: Traditional ML models achieved ROC-AUC of 0.64 on 101,145 completed trials (1976-2024). BioBERT model achieved ROC-AUC of 0.74 and Brier Score of 0.185, with 40% less squared error than industry benchmarks. BioBERT predictions were superior to benchmarks 70% of the time.
Conclusion: NLP-driven insights can enhance clinical trial success prediction in neuroscience, improving strategic planning and investment optimization in pharmaceutical R&D programs.
Abstract: This work presents the development and evaluation of an NLP-enabled probabilistic classifier designed to estimate the probability of technical and regulatory success (pTRS) for clinical trials in the field of neuroscience. While pharmaceutical R&D is plagued by high attrition rates and enormous costs, particularly within neuroscience, where success rates are below 10%, timely identification of promising programs can streamline resource allocation and reduce financial risk. Leveraging data from the ClinicalTrials.gov database and success labels from the recently developed Clinical Trial Outcome dataset, the classifier extracts text-based clinical trial features using statistical NLP techniques. These features were integrated into several non-LLM frameworks (logistic regression, gradient boosting, and random forest) to generate calibrated probability scores. Model performance was assessed on a retrospective dataset of 101,145 completed clinical trials spanning 1976-2024, achieving an overall ROC-AUC of 0.64. An LLM-based predictive model was then built using BioBERT, a domain-specific language representation encoder. The BioBERT-based model achieved an overall ROC-AUC of 0.74 and a Brier Score of 0.185, indicating its predictions had, on average, 40% less squared error than would be observed using industry benchmarks. The BioBERT-based model also made trial outcome predictions that were superior to benchmark values 70% of the time overall. By integrating NLP-driven insights into drug development decision-making, this work aims to enhance strategic planning and optimize investment allocation in neuroscience programs.
[666] Developing Fairness-Aware Task Decomposition to Improve Equity in Post-Spinal Fusion Complication Prediction
Yining Yuan, J. Ben Tamo, Wenqi Shi, Yishan Zhong, Micky C. Nnamdi, B. Randall Brenn, Steven W. Hwang, May D. Wang
Main category: cs.LG
TL;DR: FAIR-MTL: A fairness-aware multitask learning framework that uses data-driven subgroup inference to improve equity in postoperative complication prediction for scoliosis surgery, achieving better performance and reduced bias compared to traditional methods.
Details
Motivation: Existing fairness approaches in clinical prediction models rely on coarse demographic adjustments or post-hoc corrections, which fail to capture latent patient population structures and may unintentionally reinforce bias, especially in high-stakes applications like spinal fusion surgery where outcomes show substantial heterogeneity.Method: FAIR-MTL uses data-driven subgroup inference with demographic embeddings and k-means clustering to uncover latent patient subgroups. These inferred subgroups determine task routing in a shared multitask architecture. Training includes inverse-frequency weighting to mitigate subgroup imbalance and regularization to prevent overfitting to smaller groups.
Result: Achieves AUC of 0.86 and accuracy of 75% for postoperative complication prediction with four severity levels, outperforming single-task baselines. Reduces demographic parity difference to 0.055 (gender) and 0.056 (age), and equalized odds to 0.094 (gender) and 0.148 (age). SHAP and Gini analyses highlight clinically meaningful predictors like hemoglobin, hematocrit, and patient weight.
Conclusion: Incorporating unsupervised subgroup discovery into a multitask framework enables more equitable, interpretable, and clinically actionable predictions for surgical risk stratification, addressing fairness challenges in high-stakes clinical applications.
Abstract: Fairness in clinical prediction models remains a persistent challenge, particularly in high-stakes applications such as spinal fusion surgery for scoliosis, where patient outcomes exhibit substantial heterogeneity. Many existing fairness approaches rely on coarse demographic adjustments or post-hoc corrections, which fail to capture the latent structure of clinical populations and may unintentionally reinforce bias. We propose FAIR-MTL, a fairness-aware multitask learning framework designed to provide equitable and fine-grained prediction of postoperative complication severity. Instead of relying on explicit sensitive attributes during model training, FAIR-MTL employs a data-driven subgroup inference mechanism. We extract a compact demographic embedding, and apply k-means clustering to uncover latent patient subgroups that may be differentially affected by traditional models. These inferred subgroup labels determine task routing within a shared multitask architecture. During training, subgroup imbalance is mitigated through inverse-frequency weighting, and regularization prevents overfitting to smaller groups. Applied to postoperative complication prediction with four severity levels, FAIR-MTL achieves an AUC of 0.86 and an accuracy of 75%, outperforming single-task baselines while substantially reducing bias. For gender, the demographic parity difference decreases to 0.055 and equalized odds to 0.094; for age, these values reduce to 0.056 and 0.148, respectively. Model interpretability is ensured through SHAP and Gini importance analyses, which consistently highlight clinically meaningful predictors such as hemoglobin, hematocrit, and patient weight. Our findings show that incorporating unsupervised subgroup discovery into a multitask framework enables more equitable, interpretable, and clinically actionable predictions for surgical risk stratification.
[667] Efficient Matroid Bandit Linear Optimization Leveraging Unimodality
Aurélien Delage, Romaric Gaudel
Main category: cs.LG
TL;DR: Matroid semi-bandit algorithm with O(log log T) oracle calls, maintaining optimal regret while reducing computational complexity.
Details
Motivation: Existing matroid semi-bandit algorithms have optimal regret but suffer from high time complexity, especially for large matroids or those with expensive membership oracles (like in online recommendation systems requiring diversity).Method: Exploits the underlying unimodal structure of matroid semi-bandit problems to limit the number of iterations involving the membership oracle to O(log log T), while maintaining near-optimal regret guarantees.
Result: Experiments show: (i) no loss in regret compared to state-of-the-art approaches; (ii) significantly reduced time complexity and number of calls to the membership oracle across various matroid benchmarks.
Conclusion: The paper presents a more computationally efficient approach to matroid semi-bandit problems that maintains optimal regret while dramatically reducing oracle calls and overall time complexity, making it practical for large-scale applications like online recommendation systems.
Abstract: We study the combinatorial semi-bandit problem under matroid constraints. The regret achieved by recent approaches is optimal, in the sense that it matches the lower bound. Yet, time complexity remains an issue for large matroids or for matroids with costly membership oracles (e.g. online recommendation that ensures diversity). This paper sheds a new light on the matroid semi-bandit problem by exploiting its underlying unimodal structure. We demonstrate that, with negligible loss in regret, the number of iterations involving the membership oracle can be limited to \mathcal{O}(\log \log T)$. This results in an overall improved time complexity of the learning process. Experiments conducted on various matroid benchmarks show (i) no loss in regret compared to state-of-the-art approaches; and (ii) reduced time complexity and number of calls to the membership oracle.
[668] Generalized Graph Transformer Variational Autoencoder
Siddhant Karki
Main category: cs.LG
TL;DR: GGT-VAE integrates Generalized Graph Transformer with VAE for link prediction, using global self-attention and laplacian encoding instead of message passing, achieving above-baseline performance on benchmarks.
Details
Motivation: Graph link prediction is central to graph representation learning, but existing approaches (GraphVAE, GCN, GNN) rely on message passing mechanisms. There's a need for models that can capture global structural patterns without this limitation.Method: Proposes Generalized Graph Transformer Variational Autoencoder (GGT-VAE) that combines Generalized Graph Transformer Architecture with Variational Autoencoder framework. Uses transformer-style global self-attention mechanism with laplacian positional encoding to model structural patterns across nodes into latent space, avoiding reliance on message passing.
Result: Experimental results on several benchmark datasets show GGT-VAE consistently achieves above-baseline performance in terms of ROC-AUC and Average Precision metrics.
Conclusion: GGT-VAE represents one of the first studies to explore graph structure generation using a generalized graph transformer backbone in a variational framework, offering a novel approach to link prediction that captures global dependencies without message passing.
Abstract: Graph link prediction has long been a central problem in graph representation learning in both network analysis and generative modeling. Recent progress in deep learning has introduced increasingly sophisticated architectures for capturing relational dependencies within graph-structured data. In this work, we propose the Generalized Graph Transformer Variational Autoencoder (GGT-VAE). Our model integrates Generalized Graph Transformer Architecture with Variational Autoencoder framework for link prediction. Unlike prior GraphVAE, GCN, or GNN approaches, GGT-VAE leverages transformer style global self-attention mechanism along with laplacian positional encoding to model structural patterns across nodes into a latent space without relying on message passing. Experimental results on several benchmark datasets demonstrate that GGT-VAE consistently achieves above-baseline performance in terms of ROC-AUC and Average Precision. To the best of our knowledge, this is among the first studies to explore graph structure generation using a generalized graph transformer backbone in a variational framework.
[669] Neuroscience-Inspired Memory Replay for Continual Learning: A Comparative Study of Predictive Coding and Backpropagation-Based Strategies
Goutham Nalagatla, Shreyas Grandhe
Main category: cs.LG
TL;DR: Predictive coding-based generative replay outperforms backpropagation-based methods in continual learning, achieving 15.3% better retention while maintaining competitive transfer efficiency.
Details
Motivation: Catastrophic forgetting in neural networks deployed in dynamic environments remains a fundamental challenge. The paper is inspired by biological memory consolidation mechanisms to develop better continual learning solutions.Method: Proposes a novel framework for generative replay that leverages predictive coding principles. Conducts comprehensive comparison between predictive coding-based and backpropagation-based generative replay strategies across multiple benchmark datasets.
Result: Predictive coding-based replay achieves superior retention performance with average 15.3% improvement over backpropagation-based methods while maintaining competitive transfer efficiency.
Conclusion: Biologically-inspired mechanisms like predictive coding offer principled solutions to continual learning challenges. The framework provides insights into the relationship between biological memory processes and artificial learning systems, opening new avenues for neuroscience-inspired AI research.
Abstract: Continual learning remains a fundamental challenge in artificial intelligence, with catastrophic forgetting posing a significant barrier to deploying neural networks in dynamic environments. Inspired by biological memory consolidation mechanisms, we propose a novel framework for generative replay that leverages predictive coding principles to mitigate forgetting. We present a comprehensive comparison between predictive coding-based and backpropagation-based gen- erative replay strategies, evaluating their effectiveness on task retention and transfer efficiency across multiple benchmark datasets. Our experimental results demonstrate that predictive coding-based replay achieves superior retention performance (average 15.3% improvement) while maintaining competitive transfer efficiency, suggesting that biologically-inspired mechanisms can offer principled solutions to continual learning challenges. The proposed framework provides insights into the relationship between biological memory processes and artificial learning systems, opening new avenues for neuroscience-inspired AI research.
[670] Financial Text Classification Based On rLoRA Finetuning On Qwen3-8B model
Zhiming Lian
Main category: cs.LG
TL;DR: Qwen3-8B outperforms classical transformers and other large models in financial text classification tasks using instruction fine-tuning with noise injection and memory-efficient optimization techniques.
Details
Motivation: Financial text classification is crucial for quantitative trading systems, but existing models may lack robustness and efficiency for real-time financial NLP applications. The paper aims to evaluate Qwen3-8B's performance on financial sentiment analysis and news classification tasks.Method: Used Noisy Embedding Instruction Finetuning (injecting controlled noise into embedding layers) combined with Rank-stabilized Low-Rank Adaptation optimization and FlashAttention for memory-efficient training. Benchmarked against T5, BERT, RoBERTa, LLaMA1-7B, LLaMA2-7B, and Baichuan2-7B.
Result: Qwen3-8B consistently outperformed all baseline models in classification accuracy while requiring fewer training epochs, demonstrating superior performance and efficiency.
Conclusion: Qwen3-8B with instruction-based fine-tuning and memory-efficient optimization provides a scalable, economical solution for real-time financial NLP applications and offers a promising foundation for advancing quantitative trading systems.
Abstract: Financial text classification has increasingly become an important aspect in quantitative trading systems and related tasks, such as financial sentiment analysis and the classification of financial news. In this paper, we assess the performance of the large language model Qwen3-8B on both tasks. Qwen3-8B is a state-of-the-art model that exhibits strong instruction-following and multilingual capabilities, and is distinct from standard models, primarily because it is specifically optimized for efficient fine tuning and high performance on reasoning-based benchmarks, making it suitable for financial applications. To adapt this model, we apply Noisy Embedding Instruction Finetuning and based on our previous work, this method increases robustness by injecting controlled noise into the embedding layers during supervised adaptation. We improve efficiency further with Rank-stabilized Low-Rank Adaptation low-rank optimization approach, and FlashAttention, which allow for faster training with lower GPU memory. For both tasks, we benchmark Qwen3-8B against standard classical transformer models, such as T5, BERT, and RoBERTa, and large models at scale, such as LLaMA1-7B, LLaMA2-7B, and Baichuan2-7B. The findings reveal that Qwen3-8B consistently surpasses these baselines by obtaining better classification accuracy and needing fewer training epochs. The synergy of instruction-based fine-tuning and memory-efficient optimization methods suggests Qwen3-8B can potentially serve as a scalable, economical option for real-time financial NLP applications. Qwen3-8B provides a very promising base for advancing dynamic quantitative trading systems in the future.
[671] Privacy Preserving Diffusion Models for Mixed-Type Tabular Data Generation
Timur Sattarov, Marco Schreyer, Damian Borth
Main category: cs.LG
TL;DR: DP-FinDiff is a differentially private diffusion framework for synthesizing mixed-type tabular data with improved utility and privacy guarantees.
Details
Motivation: Need for safe and effective data sharing in sensitive domains (financial, medical) while maintaining privacy through differential privacy and preserving data utility for downstream tasks.Method: Uses embedding-based representations for categorical features to reduce encoding overhead. Proposes two privacy-aware training strategies: adaptive timestep sampler aligning updates with diffusion dynamics, and feature-aggregated loss mitigating clipping-induced bias.
Result: Achieves 16-42% higher utility than DP baselines at comparable privacy levels on financial and medical datasets.
Conclusion: DP-FinDiff demonstrates promise for safe and effective data sharing in sensitive domains by balancing privacy protection with data utility.
Abstract: We introduce DP-FinDiff, a differentially private diffusion framework for synthesizing mixed-type tabular data. DP-FinDiff employs embedding-based representations for categorical features, reducing encoding overhead and scaling to high-dimensional datasets. To adapt DP-training to the diffusion process, we propose two privacy-aware training strategies: an adaptive timestep sampler that aligns updates with diffusion dynamics, and a feature-aggregated loss that mitigates clipping-induced bias. Together, these enhancements improve fidelity and downstream utility without weakening privacy guarantees. On financial and medical datasets, DP-FinDiff achieves 16-42% higher utility than DP baselines at comparable privacy levels, demonstrating its promise for safe and effective data sharing in sensitive domains.
[672] ML-Tool-Bench: Tool-Augmented Planning for ML Tasks
Yaswanth Chittepu, Raghavendra Addanki, Tung Mai, Anup Rao, Branislav Kveton
Main category: cs.LG
TL;DR: A benchmark for evaluating tool-augmented ML agents with 61 specialized tools and 15 Kaggle challenges, featuring in-memory object management and showing that standard approaches struggle with complex ML pipelines.
Details
Motivation: Existing tool-use benchmarks fail to evaluate the sophisticated planning capabilities required for ML agents that orchestrate end-to-end data science workflows, including data analysis, feature engineering, model selection, and hyperparameter optimization.Method: Introduced a comprehensive benchmark with 61 specialized tools and 15 Kaggle tabular ML challenges, featuring in-memory named object management. Proposed two approaches: 1) shaped deterministic rewards with structured textual feedback, and 2) problem decomposition into sub-tasks.
Result: Standard ReAct-style approaches struggle with generating valid tool sequences for complex ML pipelines, and tree search methods with LLM-based evaluation underperform due to inconsistent state scoring. The proposed approach improves over ReAct by 16.52 percentile positions using GPT-4o.
Conclusion: The work provides a foundation for developing more capable tool-augmented planning ML agents by addressing limitations in existing approaches and demonstrating improved performance through shaped rewards and problem decomposition.
Abstract: The development of autonomous machine learning (ML) agents capable of end-to-end data science workflows represents a significant frontier in artificial intelligence. These agents must orchestrate complex sequences of data analysis, feature engineering, model selection, and hyperparameter optimization, tasks that require sophisticated planning and iteration. While recent work on building ML agents has explored using large language models (LLMs) for direct code generation, tool-augmented approaches offer greater modularity and reliability. However, existing tool-use benchmarks focus primarily on task-specific tool selection or argument extraction for tool invocation, failing to evaluate the sophisticated planning capabilities required for ML Agents. In this work, we introduce a comprehensive benchmark for evaluating tool-augmented ML agents using a curated set of 61 specialized tools and 15 tabular ML challenges from Kaggle. Our benchmark goes beyond traditional tool-use evaluation by incorporating an in-memory named object management, allowing agents to flexibly name, save, and retrieve intermediate results throughout the workflows. We demonstrate that standard ReAct-style approaches struggle to generate valid tool sequences for complex ML pipelines, and that tree search methods with LLM-based evaluation underperform due to inconsistent state scoring. To address these limitations, we propose two simple approaches: 1) using shaped deterministic rewards with structured textual feedback, and 2) decomposing the original problem into a sequence of sub-tasks, which significantly improves trajectory validity and task performance. Using GPT-4o, our approach improves over ReAct by 16.52 percentile positions, taking the median across all Kaggle challenges. We believe our work provides a foundation for developing more capable tool-augmented planning ML agents.
[673] Using physics-inspired Singular Learning Theory to understand grokking & other phase transitions in modern neural networks
Anish Lakkapragada
Main category: cs.LG
TL;DR: Empirical study of Singular Learning Theory (SLT) in toy models, testing free energy scaling and local learning coefficients across various neural network families to understand phase transitions and theory-practice gaps.
Details
Motivation: Classical statistical learning theory fails to explain modern neural networks due to their non-identifiable (singular) nature. Singular Learning Theory (SLT), grounded in algebraic geometry, offers a promising framework to close this theory-practice gap, but needs empirical validation.Method: Two-pronged empirical approach: 1) Test Arrhenius-style rate hypothesis for SLT free energy using grokking modulo-arithmetic models and Anthropic’s Toy Models of Superposition; 2) Measure how local learning coefficient scales with problem difficulty across controlled network families including polynomial regressors, low-rank linear networks, and low-rank autoencoders.
Result: Experiments recover known scaling laws while also revealing meaningful deviations from theoretical expectations. The study demonstrates SLT’s utility for understanding neural network phase transitions and identifies open research questions.
Conclusion: The paper illustrates the merits of Singular Learning Theory for explaining neural network behavior, particularly phase transitions, and provides empirical evidence supporting its framework while highlighting areas needing further theoretical development.
Abstract: Classical statistical inference and learning theory often fail to explain the success of modern neural networks. A key reason is that these models are non-identifiable (singular), violating core assumptions behind PAC bounds and asymptotic normality. Singular learning theory (SLT), a physics-inspired framework grounded in algebraic geometry, has gained popularity for its ability to close this theory-practice gap. In this paper, we empirically study SLT in toy settings relevant to interpretability and phase transitions. First, we understand the SLT free energy $\mathcal{F}_n$ by testing an Arrhenius-style rate hypothesis using both a grokking modulo-arithmetic model and Anthropic’s Toy Models of Superposition. Second, we understand the local learning coefficient $λ_α$ by measuring how it scales with problem difficulty across several controlled network families (polynomial regressors, low-rank linear networks, and low-rank autoencoders). Our experiments recover known scaling laws while others yield meaningful deviations from theoretical expectations. Overall, our paper illustrates the many merits of SLT for understanding neural network phase transitions, and poses open research questions for the field.
[674] Flow Matching for Tabular Data Synthesis
Bahrul Ilmi Nasution, Floor Eijkelboom, Mark Elliot, Richard Allmendinger, Christian A. Naesseth
Main category: cs.LG
TL;DR: Flow matching methods, especially TabbyFlow, outperform diffusion models for tabular data synthesis with better utility, lower computational cost (≤100 steps), and better privacy-utility tradeoffs.
Details
Motivation: Need for effective privacy-preserving synthetic data generation methods, with flow matching offering a promising alternative to diffusion models for tabular data synthesis.Method: Comprehensive empirical study comparing flow matching (FM and variational FM) with state-of-the-art diffusion methods (TabDDPM and TabSyn) for tabular data synthesis, evaluating OT and VP probability paths, and comparing deterministic vs stochastic samplers.
Result: Flow matching, particularly TabbyFlow, outperforms diffusion baselines; achieves better performance with remarkably low function evaluations (≤100 steps); OT path shows superior performance while VP has potential for lower disclosure risk; stochastic flows preserve marginal distributions and can generate high utility data with reduced disclosure risk.
Conclusion: Flow matching is a superior approach for tabular data synthesis compared to diffusion models, offering better performance, computational efficiency, and flexible privacy-utility tradeoffs through different probability paths and stochastic sampling.
Abstract: Synthetic data generation is an important tool for privacy-preserving data sharing. While diffusion models have set recent benchmarks, flow matching (FM) offers a promising alternative. This paper presents different ways to implement flow matching for tabular data synthesis. We provide a comprehensive empirical study that compares flow matching (FM and variational FM) with a state-of-the-art diffusion method (TabDDPM and TabSyn) in tabular data synthesis. We evaluate both the standard Optimal Transport (OT) and the Variance Preserving (VP) probability paths, and also compare deterministic and stochastic samplers – something possible when learning to generate using \textit{variational} flow matching – characterising the empirical relationship between data utility and privacy risk. Our key findings reveal that flow matching, particularly TabbyFlow, outperforms diffusion baselines. Flow matching methods also achieves better performance with remarkably low function evaluations ($\leq$ 100 steps), offering a substantial computational advantage. The choice of probability path is also crucial, as using the OT path demonstrates superior performance, while VP has potential for producing synthetic data with lower disclosure risk. Lastly, our results show that making flows stochastic not only preserves marginal distributions but, in some instances, enables the generation of high utility synthetic data with reduced disclosure risk.
[675] Towards Precision Protein-Ligand Affinity Prediction Benchmark: A Complete and Modification-Aware DAVIS Dataset
Ming-Hsiu Wu, Ziqian Xie, Shuiwang Ji, Degui Zhi
Main category: cs.LG
TL;DR: Researchers created a more realistic protein-ligand binding dataset by adding 4,032 kinase-ligand pairs with natural protein modifications (substitutions, insertions, deletions, phosphorylation) to DAVIS, then proposed three benchmarks to test model generalization on modified proteins.
Details
Motivation: Current AI models for drug discovery overfit to oversimplified datasets that don't represent naturally occurring proteins with modifications, limiting their real-world applicability in precision medicine.Method: Curated a complete, modification-aware version of DAVIS dataset with 4,032 kinase-ligand pairs involving protein modifications. Proposed three benchmark settings: Augmented Dataset Prediction, Wild-Type to Modification Generalization, and Few-Shot Modification Generalization.
Result: Docking-based models generalize better in zero-shot settings, while docking-free models overfit to wild-type proteins but improve with fine-tuning on modified examples. The new dataset reveals model limitations in handling protein modifications.
Conclusion: The curated dataset and benchmarks provide a foundation for developing models that better generalize to protein modifications, advancing precision medicine in drug discovery by addressing real-world biological complexity.
Abstract: Advancements in AI for science unlocks capabilities for critical drug discovery tasks such as protein-ligand binding affinity prediction. However, current models overfit to existing oversimplified datasets that does not represent naturally occurring and biologically relevant proteins with modifications. In this work, we curate a complete and modification-aware version of the widely used DAVIS dataset by incorporating 4,032 kinase-ligand pairs involving substitutions, insertions, deletions, and phosphorylation events. This enriched dataset enables benchmarking of predictive models under biologically realistic conditions. Based on this new dataset, we propose three benchmark settings-Augmented Dataset Prediction, Wild-Type to Modification Generalization, and Few-Shot Modification Generalization-designed to assess model robustness in the presence of protein modifications. Through extensive evaluation of both docking-free and docking-based methods, we find that docking-based model generalize better in zero-shot settings. In contrast, docking-free models tend to overfit to wild-type proteins and struggle with unseen modifications but show notable improvement when fine-tuned on a small set of modified examples. We anticipate that the curated dataset and benchmarks offer a valuable foundation for developing models that better generalize to protein modifications, ultimately advancing precision medicine in drug discovery. The benchmark is available at: https://github.com/ZhiGroup/DAVIS-complete
[676] Exploiting Function-Family Structure in Analog Circuit Optimization
Zhuohua Liu, Kaiqi Huang, Qinxin Mei, Yuanqi Hu, Wei W. Xing
Main category: cs.LG
TL;DR: Circuit Prior Network (CPN) uses pre-trained tabular foundation models with physics-informed priors to optimize analog circuits, achieving much better performance than traditional Gaussian processes with far fewer evaluations.
Details
Motivation: Traditional Gaussian-process surrogates for analog circuit optimization impose globally smooth, stationary priors that are misaligned with the actual physics of circuits (exponential device laws, rational transfer functions, regime-dependent dynamics), leading to poor performance at realistic sample sizes (50-100 evaluations).Method: Circuit Prior Network (CPN) combines a tabular foundation model (TabPFN v2) with Direct Expected Improvement (DEI), computing expected improvement exactly under discrete posteriors rather than Gaussian approximations. It uses pre-trained models encoding circuit physics primitives instead of hand-crafted per-circuit models.
Result: Across 6 circuits and 25 baselines, CPN achieves R² ≈ 0.99 in small-sample regimes where GP-Matérn attains only R² = 0.16 on Bandgap circuits. It delivers 1.05-3.81× higher Figure of Merit with 3.34-11.89× fewer iterations.
Conclusion: Structure-matched priors enable reliable circuit optimization without per-circuit engineering, suggesting a shift from hand-crafting models as priors toward systematic physics-informed structure identification.
Abstract: Analog circuit optimization is typically framed as black-box search over arbitrary smooth functions, yet device physics constrains performance mappings to structured families: exponential device laws, rational transfer functions, and regime-dependent dynamics. Off-the-shelf Gaussian-process surrogates impose globally smooth, stationary priors that are misaligned with these regime-switching primitives and can severely misfit highly nonlinear circuits at realistic sample sizes (50–100 evaluations). We demonstrate that pre-trained tabular models encoding these primitives enable reliable optimization without per-circuit engineering. Circuit Prior Network (CPN) combines a tabular foundation model (TabPFN v2) with Direct Expected Improvement (DEI), computing expected improvement exactly under discrete posteriors rather than Gaussian approximations. Across 6 circuits and 25 baselines, structure-matched priors achieve $R^2 \approx 0.99$ in small-sample regimes where GP-Matérn attains only $R^2 = 0.16$ on Bandgap, deliver $1.05$–$3.81\times$ higher FoM with $3.34$–$11.89\times$ fewer iterations, and suggest a shift from hand-crafting models as priors toward systematic physics-informed structure identification. Our code will be made publicly available upon paper acceptance.
[677] Graph Data Augmentation with Contrastive Learning on Covariate Distribution Shift
Fanlong Zeng, Wensheng Gan
Main category: cs.LG
TL;DR: MPAIACL is a new method using contrastive learning for adversarial invariant augmentation to address covariate distribution shift in graph neural networks.
Details
Motivation: Existing GNNs fail to handle covariate distribution shift, and current methods don't fully utilize latent space information. The latent space contains rich information that could be better leveraged.Method: MPAIACL (More Powerful Adversarial Invariant Augmentation using Contrastive Learning) uses contrastive learning to unlock the full potential of vector representations by harnessing their intrinsic information from the latent space.
Result: MPAIACL demonstrates robust generalization and effectiveness, performing well compared with other baselines across various public OOD datasets.
Conclusion: The proposed MPAIACL method successfully addresses covariate distribution shift in graph data by leveraging contrastive learning to better utilize latent space information, showing strong performance on OOD datasets.
Abstract: Covariate distribution shift occurs when certain structural features present in the test set are absent from the training set. It is a common type of out-of-distribution (OOD) problem, frequently encountered in real-world graph data with complex structures. Existing research has revealed that most out-of-the-box graph neural networks (GNNs) fail to account for covariate shifts. Furthermore, we observe that existing methods aimed at addressing covariate shifts often fail to fully leverage the rich information contained within the latent space. Motivated by the potential of the latent space, we introduce a new method called MPAIACL for More Powerful Adversarial Invariant Augmentation using Contrastive Learning. MPAIACL leverages contrastive learning to unlock the full potential of vector representations by harnessing their intrinsic information. Through extensive experiments, MPAIACL demonstrates its robust generalization and effectiveness, as it performs well compared with other baselines across various public OOD datasets. The code is publicly available at https://github.com/flzeng1/MPAIACL.
[678] Upcycled and Merged MoE Reward Model for Mitigating Reward Hacking
Lingling Fu
Main category: cs.LG
TL;DR: Proposes upcycle-and-merge MoE reward modeling to mitigate reward hacking in RLHF while reducing computational overhead.
Details
Motivation: Conventional reward models in RLHF suffer from reward hacking/over-optimization where policies exploit shortcuts for high scores that don't reflect true human preferences. MoE-based models improve discrimination but introduce substantial computational overhead.Method: 1) Upcycle a dense reward model into MoE architecture with shared expert for general knowledge and normal experts for instruction-specific patterns. 2) Apply routing-weight normalization and merge experts back into dense model through learnable weight-averaging mechanism to preserve performance while reducing inference cost.
Result: Experimental results show the method effectively mitigates reward hacking across various model scales while significantly reducing inference costs compared to standard MoE approaches.
Conclusion: The upcycle-and-merge MoE structure improves both robustness (against reward hacking) and efficiency of RLHF reward models, highlighting its potential for practical deployment.
Abstract: Reward models play a critical role in Reinforcement Learning from Human Feedback (RLHF) by assessing the consistency between generated outputs and human preferences. However, conventional reward models are prone to reward hacking or over-optimization, where the policy exploits shortcut patterns to obtain high reward scores that do not reflect true human preference. Although Mixture-of-Experts (MoE)-based reward models can enhance discriminative capability, they typically introduce substantial computational overhead. To address these challenges, we propose an upcycle and merge MoE reward modeling approach. We first upcycle a dense reward model into a MoE architecture, where a shared expert captures general knowledge, while normal experts specialize in instruction-specific patterns. We then apply routing-weight normalization and merge experts back into a dense model through a learnable weight-averaging mechanism, preserving performance gains while significantly reducing inference cost. Experimental results demonstrate that our method effectively mitigates reward hacking across various model scales. Our work highlights the potential of upcycle and merge MoE structures for improving both robustness and efficiency of RLHF reward models.
[679] ESMC: MLLM-Based Embedding Selection for Explainable Multiple Clustering
Xinyue Wang, Yuheng Jia, Hui Liu, Junhui Hou
Main category: cs.LG
TL;DR: A method that uses MLLMs for user-driven clustering by leveraging hidden text token embeddings and pseudo-label learning, enabling customizable clustering based on user-defined criteria.
Details
Motivation: Traditional deep clustering methods only provide one fixed clustering result per dataset, which may not meet user needs. The paper aims to leverage MLLMs for user-driven clustering that adapts to user-specified semantic requirements.Method: Discovers that MLLMs’ hidden states of text tokens strongly relate to corresponding features, uses these embeddings for clustering based on user-defined criteria, and employs a lightweight clustering head with pseudo-label learning to enhance accuracy.
Result: Extensive experiments demonstrate competitive performance on diverse datasets and metrics, showing the method’s effectiveness for user-driven clustering.
Conclusion: The proposed approach successfully enables user-driven clustering using MLLMs, overcoming limitations of traditional deep clustering methods by providing customizable clustering based on user-defined semantic requirements.
Abstract: Typical deep clustering methods, while achieving notable progress, can only provide one clustering result per dataset. This limitation arises from their assumption of a fixed underlying data distribution, which may fail to meet user needs and provide unsatisfactory clustering outcomes. Our work investigates how multi-modal large language models (MLLMs) can be leveraged to achieve user-driven clustering, emphasizing their adaptability to user-specified semantic requirements. However, directly using MLLM output for clustering has risks for producing unstructured and generic image descriptions instead of feature-specific and concrete ones. To address these issues, our method first discovers that MLLMs’ hidden states of text tokens are strongly related to the corresponding features, and leverages these embeddings to perform clusterings from any user-defined criteria. We also employ a lightweight clustering head augmented with pseudo-label learning, significantly enhancing clustering accuracy. Extensive experiments demonstrate its competitive performance on diverse datasets and metrics.
[680] Deep Learning for Modeling and Dispatching Hybrid Wind Farm Power Generation
Zach Lawrence, Jessica Yao, Chris Qin
Main category: cs.LG
TL;DR: Two deep learning frameworks for hybrid wind farms: COVE-NN (LSTM-based dispatch strategy) reduces COVE by 32.3%, and a power generation modeling framework improves RMSE by 9.5% and power curve similarity by 18.9%.
Details
Motivation: Hybrid wind farms with integrated energy storage need data-driven dispatch strategies to maximize wind energy value using localized grid demand and market conditions. Synthetic power generation data from atmospheric conditions can improve robustness of these strategies.Method: Developed two deep learning frameworks: 1) COVE-NN - an LSTM-based dispatch strategy tailored to individual wind farms, and 2) a power generation modeling framework that creates synthetic power generation data from atmospheric conditions.
Result: COVE-NN reduced annual COVE by 32.3% over 43 years of simulated operations at Pyron site. Power generation modeling framework reduced RMSE by 9.5% and improved power curve similarity by 18.9% when validated on Palouse wind farm.
Conclusion: These models enable more robust, data-driven dispatch strategies for hybrid wind farms and can potentially be extended to other renewable energy systems.
Abstract: Wind farms with integrated energy storage, or hybrid wind farms, are able to store energy and dispatch it to the grid following an operational strategy. For individual wind farms with integrated energy storage capacity, data-driven dispatch strategies using localized grid demand and market conditions as input parameters stand to maximize wind energy value. Synthetic power generation data modeled on atmospheric conditions provide another avenue for improving the robustness of data-driven dispatch strategies. To these ends, the present work develops two deep learning frameworks: COVE-NN, an LSTM-based dispatch strategy tailored to individual wind farms, which reduced annual COVE by 32.3% over 43 years of simulated operations in a case study at the Pyron site; and a power generation modeling framework that reduced RMSE by 9.5% and improved power curve similarity by 18.9% when validated on the Palouse wind farm. Together, these models pave the way for more robust, data-driven dispatch strategies and potential extensions to other renewable energy systems.
[681] REM: Evaluating LLM Embodied Spatial Reasoning through Multi-Frame Trajectories
Jacob Thompson, Emiliano Garcia-Lopez, Yonatan Bisk
Main category: cs.LG
TL;DR: MLLMs lack human-like spatial reasoning despite video training; REM benchmark evaluates embodied spatial reasoning in 3D environments, revealing models struggle with complexity humans handle easily.
Details
Motivation: Humans develop viewpoint-independent cognitive maps through navigation, enabling intuitive spatial reasoning. Despite extensive video training, multimodal large language models (MLLMs) lack this fundamental spatial reasoning capability, which is critical for embodied AI applications.Method: Introduce REM (Reasoning over Embodied Multi-Frame Trajectories), a benchmark using controllable 3D environments for long-horizon embodied spatial reasoning. REM systematically evaluates key aspects like object permanence/distinction, spatial relationships, and numerical tracking across dynamic embodied viewpoints.
Result: Current best-performing models show promising overall performance but become increasingly unreliable at moderate complexity levels that humans handle easily. Models struggle with developing robust spatial representations from sequential visual input.
Conclusion: MLLMs face significant challenges in spatial reasoning from visual sequences. REM provides targeted metrics and diagnostics to foster improved spatial understanding in future models for embodied AI applications.
Abstract: Humans build viewpoint-independent cognitive maps through navigation, enabling intuitive reasoning about object permanence and spatial relations. We argue that multimodal large language models (MLLMs), despite extensive video training, lack this fundamental spatial reasoning capability, a critical limitation for embodied applications. To demonstrate these limitations and drive research, we introduce REM (Reasoning over Embodied Multi-Frame Trajectories), a benchmark using controllable 3D environments for long-horizon embodied spatial reasoning. REM systematically evaluates key aspects like object permanence/distinction, spatial relationships, and numerical tracking across dynamic embodied viewpoints. Our evaluation shows that the best-performing current models exhibit promising overall performance, but become increasingly unreliable at even moderate complexity levels easily handled by humans. These findings highlight challenges MLLMs face in developing robust spatial representations from sequential visual input. Consequently, REM provides targeted metrics and diagnostics to foster improved spatial understanding in future models.
[682] Preventing Model Collapse via Contraction-Conditioned Neural Filters
Zongjian Han, Yiran Liang, Ruiwen Wang, Yiwei Luo, Yilin Huang, Xiaotong Song, Dongqing Wei
Main category: cs.LG
TL;DR: Neural network filter method using contraction operators prevents model collapse in recursive generative model training without requiring increasing sample sizes.
Details
Motivation: Address model collapse in recursive training of generative models, eliminating the need for superlinear sample growth required by existing methods.Method: Design neural filter that learns to satisfy contraction conditions using specialized neural network architectures and loss functions within an unbiased estimation framework.
Result: Theoretical analysis shows estimation errors converge probabilistically with constant sample sizes when contraction conditions are satisfied; experiments demonstrate effective prevention of model collapse under fixed sample sizes.
Conclusion: Provides an end-to-end solution for practical applications by enabling stable recursive training without increasing sample requirements through learned contraction conditions.
Abstract: This paper presents a neural network filter method based on contraction operators to address model collapse in recursive training of generative models. Unlike \cite{xu2024probabilistic}, which requires superlinear sample growth ($O(t^{1+s})$), our approach completely eliminates the dependence on increasing sample sizes within an unbiased estimation framework by designing a neural filter that learns to satisfy contraction conditions. We develop specialized neural network architectures and loss functions that enable the filter to actively learn contraction conditions satisfying Assumption 2.3 in exponential family distributions, thereby ensuring practical application of our theoretical results. Theoretical analysis demonstrates that when the learned contraction conditions are satisfied, estimation errors converge probabilistically even with constant sample sizes, i.e., $\limsup_{t\to\infty}\mathbb{P}(|\mathbf{e}_t|>δ)=0$ for any $δ>0$. Experimental results show that our neural network filter effectively learns contraction conditions and prevents model collapse under fixed sample size settings, providing an end-to-end solution for practical applications.
[683] Forecasting India’s Demographic Transition Under Fertility Policy Scenarios Using hybrid LSTM-PINN Model
Subarna Khanra, Vijay Kumar Kukreja, Indu Bala
Main category: cs.LG
TL;DR: A hybrid LSTM-PINN model integrates policy-aware fertility functions for India’s demographic forecasting from 2024-2054, showing how different fertility policies affect age distribution, dependency ratios, and workforce size.
Details
Motivation: Demographic forecasting is crucial for policy planning in rapidly evolving nations like India, where fertility transitions, policy interventions, and age-structured dynamics interact complexly. Current methods may not adequately capture these interactions and physical constraints.Method: A hybrid modeling framework combining Physics-Informed Neural Networks (PINNs) with Long Short-Term Memory (LSTM) networks. PINNs embed the governing transport-reaction partial differential equation with India-specific demographic indicators, while LSTM layers capture temporal dependencies. The model incorporates policy-aware fertility functions and is applied to three fertility-policy scenarios.
Result: Fertility policies substantially shape future age distribution, dependency ratios, and workforce size. Stricter controls intensify ageing and reduce labor force participation, while relaxed policies support workforce growth but increase population pressure. The hybrid LSTM-PINN approach proves effective for demographic forecasting with both accuracy and interpretability.
Conclusion: The hybrid LSTM-PINN is an effective approach for demographic forecasting, offering actionable insights for India’s policy debates. The findings highlight the need for balanced fertility interventions to ensure sustainable socio-economic development, as extreme policies have significant trade-offs between workforce growth and population pressure.
Abstract: Demographic forecasting remains a fundamental challenge for policy planning in rapidly evolving nations such as India, where fertility transitions, policy interventions, and age structured dynamics interact in complex ways. In this study, we present a hybrid modelling framework that integrates policy-aware fertility functions into a Physics-Informed Neural Network (PINN) enhanced with Long Short-Term Memory (LSTM) networks to capture physical constraints and temporal dependencies in population dynamics. The model is applied to India’s age structured population from 2024 to 2054 under three fertility-policy scenarios: continuation of current fertility decline, stricter population control, and relaxed fertility promotion. The governing transport-reaction partial differential equation is formulated with India-specific demographic indicators, including age-specific fertility and mortality rates. PINNs embed the core population equation and policy-driven fertility changes, while LSTM layers improve long-term forecasting across decades. Results show that fertility policies substantially shape future age distribution, dependency ratios, and workforce size. Stricter controls intensify ageing and reduce labour force participation, whereas relaxed policies support workforce growth but increase population pressure. Our findings suggest that the hybrid LSTM-PINN is an effective approach for demographic forecasting, offering accuracy with interpretability. Beyond methodological novelty, this work provides actionable insights for India’s demographic policy debates, highlighting the need for balanced fertility interventions to ensure sustainable socio-economic development.
[684] Provable Benefit of Sign Descent: A Minimal Model Under Heavy-Tailed Class Imbalance
Robin Yadav, Shuo Xie, Tianhao Wang, Zhiyuan Li
Main category: cs.LG
TL;DR: The paper analyzes why adaptive optimization methods like Adam outperform Gradient Descent in LLM pretraining, focusing on how heavy-tailed class imbalance in language data enables faster convergence for coordinate-wise algorithms like Sign descent compared to normalized GD.
Details
Motivation: Recent studies propose smoothness assumptions to explain advantages of adaptive algorithms, but it's unclear how these assumptions manifest in language modeling tasks. The paper aims to analyze the benefit of ℓ∞-norm descent (sign descent) directly from data distribution properties, specifically heavy-tailed class imbalance in language data.Method: The authors propose a minimal yet representative setting of next-token prediction where they can provably show faster convergence of coordinate-wise algorithms like Sign descent (steepest descent w.r.t. ℓ∞ norm) over normalized GD (steepest descent w.r.t. ℓ₂ norm) in the presence of heavy-tailed class imbalance.
Result: The paper provides theoretical analysis showing that Sign descent converges faster than normalized GD when dealing with heavy-tailed class imbalance, which is characteristic of language modeling data distributions.
Conclusion: Heavy-tailed class imbalance in language data explains why coordinate-wise adaptive optimization methods like Sign descent outperform traditional gradient descent methods in LLM pretraining, providing a data-driven explanation rather than relying solely on smoothness assumptions.
Abstract: Adaptive optimization methods (such as Adam) play a major role in LLM pretraining, significantly outperforming Gradient Descent (GD). Recent studies have proposed new smoothness assumptions on the loss function to explain the advantages of adaptive algorithms with structured preconditioners, e.g., coordinate-wise or layer-wise, and steepest descent methods w.r.t. non-euclidean norms, e.g., $\ell_\infty$ norm or spectral norm, over GD. However, it remains unclear how these smoothness assumptions manifest in language modelling tasks. In this work, we aim to analyze the benefit of $\ell_\infty$-norm descent (a.k.a. sign descent) directly from properties of the data distribution, namely, heavy-tailed class imbalance. We propose a minimal yet representative setting of next-token prediction, where we can provably show faster convergence of coordinate-wise algorithms such as Sign descent (steepest descent w.r.t. $\ell_\infty$ norm) over normalized GD (steepest descent w.r.t. to $\ell_2$ norm) in the presence of heavy tail class imbalance.
[685] Text Mining Analysis of Symptom Patterns in Medical Chatbot Conversations
Hamed Razavi
Main category: cs.LG
TL;DR: Researchers use NLP methods to analyze symptom patterns in medical chatbot conversations, finding structured clinical topics and symptom relationships that can enhance diagnostic support.
Details
Motivation: With the rapid growth of digital health systems, there's a need to better understand how they interpret patient-reported symptoms. Chatbots in healthcare provide clinical support and generate text-based data that could reveal meaningful clinical patterns for improved symptom interpretation and decision support.Method: Multi-method NLP approach using the Medical Conversations to Disease Dataset (960 multi-turn dialogues across 24 clinical conditions). Methods include: Latent Dirichlet Allocation (LDA) for latent symptom themes, K-Means for symptom clustering, Transformer-based Named Entity Recognition (NER) for medical concept extraction, and Apriori algorithm for frequent symptom pair discovery.
Result: Analysis revealed coherent clinically relevant topic structures, moderate clustering cohesiveness, and high-confidence symptom relationships (e.g., fever-headache, rash-itchiness). Conversational medical data shows potential as valuable diagnostic signals for early symptom interpretation.
Conclusion: The research demonstrates that unstructured medical chatbot dialogues can be converted into actionable symptom knowledge, providing an extensible framework to enhance future chatbot performance, dependability, and clinical utility in tele-health systems.
Abstract: The fast growth of digital health systems has led to a need to better comprehend how they interpret and represent patient-reported symptoms. Chatbots have been used in healthcare to provide clinical support and enhance the user experience, making it possible to provide meaningful clinical patterns from text-based data through chatbots. The proposed research utilises several different natural language processing methods to study the occurrences of symptom descriptions in medicine as well as analyse the patterns that emerge through these conversations within medical bots. Through the use of the Medical Conversations to Disease Dataset which contains 960 multi-turn dialogues divided into 24 Clinical Conditions, a standardised representation of conversations between patient and bot is created for further analysis by computational means. The multi-method approach uses a variety of tools, including Latent Dirichlet Allocation (LDA) to identify latent symptom themes, K-Means to group symptom descriptions by similarity, Transformer-based Named Entity Recognition (NER) to extract medical concepts, and the Apriori algorithm to discover frequent symptom pairs. Findings from the analysis indicate a coherent structure of clinically relevant topics, moderate levels of clustering cohesiveness and several high confidence rates on the relationships between symptoms like fever headache and rash itchiness. The results support the notion that conversational medical data can be a valuable diagnostic signal for early symptom interpretation, assist in strengthening decision support and improve how users interact with tele-health technology. By demonstrating a method for converting unstructured free-flowing dialogue into actionable knowledge regarding symptoms this work provides an extensible framework to further enhance future performance, dependability and clinical utility of selecting medical chatbots.
[686] AI Agent for Source Finding by SoFiA-2 for SKA-SDC2
Xingchen Zhou, Nan Li, Peng Jia, Yingfeng Liu, Furen Deng, Shuanghao Shu, Ying Li, Liang Cao, Huanyuan Shan, Ayodeji Ibitoye
Main category: cs.LG
TL;DR: AI agent using reinforcement learning (Soft Actor-Critic) automatically optimizes parameters for radio source extraction programs like SoFiA, outperforming manual benchmarks on SKA Science Data Challenge dataset.
Details
Motivation: Source extraction in next-generation radio sky surveys (like SKA) requires complex parameter tuning in programs like SoFiA and Aegean. Manual parameter optimization is non-trivial and time-consuming, as results heavily depend on multiple parameters across different modules.Method: Proposed framework uses reinforcement learning agent (Soft Actor-Critic algorithm) to automatically optimize source extraction parameters. The agent interacts with environment by adjusting parameters and receives feedback via SDC2 score, progressively learning to select better parameter sets.
Result: After training, the AI agent automatically identifies optimal parameter configurations that outperform Team SoFiA’s benchmark within only 100 evaluation steps, with reduced time consumption compared to manual tuning.
Conclusion: The RL-based framework successfully addresses complex parameter tuning for radio source extraction and could be applied to similar problems beyond radio surveys. However, high-quality training sets with representative observations and ground truth catalogs are essential for success.
Abstract: Source extraction is crucial in analyzing data from next-generation, large-scale sky surveys in radio bands, such as the Square Kilometre Array (SKA). Several source extraction programs, including SoFiA and Aegean, have been developed to address this challenge. However, finding optimal parameter configurations when applying these programs to real observations is non-trivial. For example, the outcomes of SoFiA intensely depend on several key parameters across its preconditioning, source-finding, and reliability-filtering modules. To address this issue, we propose a framework to automatically optimize these parameters using an AI agent based on a state-of-the-art reinforcement learning (RL) algorithm, i.e., Soft Actor-Critic (SAC). The SKA Science Data Challenge 2 (SDC2) dataset is utilized to assess the feasibility and reliability of this framework. The AI agent interacts with the environment by adjusting parameters based on the feedback from the SDC2 score defined by the SDC2 Team, progressively learning to select parameter sets that yield improved performance. After sufficient training, the AI agent can automatically identify an optimal parameter configuration that outperform the benchmark set by Team SoFiA within only 100 evaluation steps and with reduced time consumption. Our approach could address similar problems requiring complex parameter tuning, beyond radio band surveys and source extraction. Yet, high-quality training sets containing representative observations and catalogs of ground truth are essential.
[687] What Is Preference Optimization Doing, How and Why?
Yue Wang, Qizhou Wang, Zizhuo Zhang, Ang Li, Gang Niu, Bo Han, Masashi Sugiyama
Main category: cs.LG
TL;DR: The paper analyzes optimization dynamics of preference optimization methods DPO and PPO, revealing DPO has stable targets while PPO has dynamic targets balancing exploration-exploitation, and examines different roles of positive/negative learning and loss reweighting components.
Details
Motivation: To fill the gap in understanding why DPO is considered supervised learning while PPO is reinforcement learning, by analyzing their optimization dynamics and underlying algorithmic behaviors.Method: Analyzes optimization dynamics by examining target directions of gradient updates and roles of three key components: positive learning, negative learning, and loss reweighting. Conducts ablation studies to examine how controlling these dynamics impacts performance.
Result: Reveals DPO follows stable targets while PPO follows dynamic targets balancing exploration-exploitation. Shows positive/negative learning jointly shape targets in DPO but offset each other, while in PPO negative learning supports exploration. Loss reweighting in DPO acts as regularizer, while in PPO indicates distinct roles of token groups.
Conclusion: The analysis deepens understanding of preference optimization methods and provides insights for developing more preference-aligned LLMs by revealing fundamental differences in optimization dynamics between DPO and PPO.
Abstract: Preference optimization (PO) is indispensable for large language models (LLMs), with methods such as direct preference optimization (DPO) and proximal policy optimization (PPO) achieving great success. A common belief is that DPO is supervised learning while PPO is reinforcement learning, yet deeper analyses for the reasons underlying these differences remain lacking. To fill this gap, we analyze their optimization dynamics, revealing distinct algorithmic behaviors and comprehending their underlying causes. First, we examine the target directions of gradient-based updates and find that DPO follows stable targets, whereas PPO follows dynamic targets that balance exploration and exploitation, thus validating the common belief from a new perspective. Second, we examine the roles of positive learning, negative learning, and loss reweighting, which are three key components in PO methods. Our analyses reveal that these components play fairly different roles. In DPO, positive and negative learning jointly shape the learning targets meanwhile mutually offset each other. However, loss reweighting in DPO acts less as a reward signal but more as a regularizer to mitigate overfitting. In PPO, negative learning primarily supports exploration rather than determining the targets. Meanwhile, loss reweighting, related to absolute values of token-level advantages, indicates the distinct roles of token groups in updating targets. Given these findings, we conduct carefully designed ablation studies to further examine how controlling these dynamics impacts optimization efficiency and practical performance. The insights gained from our analyses not only deepen the understanding of PO methods but also inspire the development of more preference-aligned LLMs.
[688] Sigma: The Key for Vision-Language-Action Models toward Telepathic Alignment
Libo Wang
Main category: cs.LG
TL;DR: Researchers developed “Sigma” - a VLA model for humanoid robots that creates a time-updatable thought space between semantics and control, achieving telepathic communication without retraining the base model.
Details
Motivation: To address the lack of a time-updatable mediating thought space between semantics and continuous control in humanoid robot cognitive systems, enabling better semantic alignment and intention-driven behavior.Method: Built Sigma VLA model on RTX 4090 using pi05_base foundation, preprocessed svla_so101_pickplace dataset, designed custom architecture combining deep semantic understanding and association, optimized data preprocessing, LoRA fine-tuning, and inference-stage adapter.
Result: Sigma showed stable decrease in control MSE across vector, fragment, and trajectory timescales while maintaining telepathy norm and semantic-text alignment quality unchanged, demonstrating mind-responsive alignment control.
Conclusion: The architecture successfully quantifies mind-responsive alignment control through deep semantic understanding and association without retraining the base model, providing reproducible experience for semantic alignment in humanoid robots.
Abstract: To address the gap in humanoid robot cognitive systems regarding the lack of a time-updable mediating thought space between semantics and continuous control, this study constructs and trains a VLA model named “Sigma” that runs on a single RTX 4090. It uses the open-source pi05_base model as a foundation and preprocesses svla_so101_pickplace into a training dataset. The researcher independently designed an architecture for a vision-language-action model that combines deep semantic understanding and association to achieve telepathic communication. The training process involved repeated optimizations of data preprocessing, LoRA fine-tuning, and the inference-stage adapter. The experiment employed offline closed-loop replay, comparing Sigma with the untuned pure pi05_base_base model under data conditions. Results showed that Sigma exhibited a stable decrease in control MSE across vector, fragment, and entire trajectory timescales, while maintaining the telepathy norm and semantic-text alignment quality unchanged. It demonstrates that mind-responsive alignment control is quantified through an architecture that combines deep understanding of semantics and association without retraining the base model, which provides reproducible experience for semantic alignment and intention-driven behavior in humanoid robots.
[689] Limitations of Using Identical Distributions for Training and Testing When Learning Boolean Functions
Jordi Pérez-Guijarro
Main category: cs.LG
TL;DR: Matching training and test distributions is not always optimal for generalization, assuming one-way functions exist, but standard conclusions hold for uniform distributions with regular target functions.
Details
Motivation: To investigate whether identical training and test distributions are always optimal for generalization, especially when distributions differ, challenging the common assumption in machine learning.Method: Theoretical analysis assuming existence of one-way functions, examining generalization behavior under distribution mismatch conditions, with special consideration of uniform distributions and regular target functions.
Result: Surprisingly, matching distributions is not always optimal for generalization when one-way functions exist, contradicting standard learning methods. However, with regular target functions and uniform distribution, the standard conclusion (matching is optimal) holds.
Conclusion: Distribution matching is not universally optimal for generalization, revealing a fundamental limitation in standard learning assumptions, though regularity conditions can restore conventional wisdom in specific cases like uniform distributions.
Abstract: When the distributions of the training and test data do not coincide, the problem of understanding generalization becomes considerably more complex, prompting a variety of questions. In this work, we focus on a fundamental one: Is it always optimal for the training distribution to be identical to the test distribution? Surprisingly, assuming the existence of one-way functions, we find that the answer is no. That is, matching distributions is not always the best scenario, which contrasts with the behavior of most learning methods. Nonetheless, we also show that when certain regularities are imposed on the target functions, the standard conclusion is recovered in the case of the uniform distribution.
[690] Estimating the Effective Rank of Vision Transformers via Low-Rank Factorization
Liyu Zerihun
Main category: cs.LG
TL;DR: The paper introduces a framework to estimate a model’s intrinsic dimensionality by treating learned representations as projections onto low-rank subspaces, defining effective rank as a region where factorized students achieve 85-95% of teacher accuracy.
Details
Motivation: Deep networks are heavily over-parameterized but their learned representations often exhibit low-rank structure. There's a need to characterize the intrinsic dimensionality of these models to understand their true capacity and enable parameter compression.Method: Train a full-rank teacher model, factorize its weights at multiple ranks, then train each factorized student via distillation to measure performance as a function of rank. Define effective rank as the smallest contiguous set of ranks where students reach 85-95% of teacher accuracy, using monotone PCHIP interpolation for stabilization. Also define effective knee as the rank maximizing perpendicular distance between smoothed accuracy curve and its endpoint secant.
Result: On ViT-B/32 fine-tuned on CIFAR-100, factorizing linear blocks with distillation yields effective-rank region of [16, 34] and effective knee at r* ~ 31. At rank 32, student achieves 69.46% accuracy vs. 73.35% for teacher (~94.7% of baseline) with substantial parameter compression.
Conclusion: The framework provides a practical tool for estimating effective-rank regions and knees across architectures and datasets, offering a systematic way to characterize the intrinsic dimensionality of deep models and enabling parameter-efficient representations.
Abstract: Deep networks are heavily over-parameterized, yet their learned representations often admit low-rank structure. We introduce a framework for estimating a model’s intrinsic dimensionality by treating learned representations as projections onto a low-rank subspace of the model’s full capacity. Our approach: train a full-rank teacher, factorize its weights at multiple ranks, and train each factorized student via distillation to measure performance as a function of rank. We define effective rank as a region, not a point: the smallest contiguous set of ranks for which the student reaches 85-95% of teacher accuracy. To stabilize estimates, we fit accuracy vs. rank with a monotone PCHIP interpolant and identify crossings of the normalized curve. We also define the effective knee as the rank maximizing perpendicular distance between the smoothed accuracy curve and its endpoint secant; an intrinsic indicator of where marginal gains concentrate. On ViT-B/32 fine-tuned on CIFAR-100 (one seed, due to compute constraints), factorizing linear blocks and training with distillation yields an effective-rank region of approximately [16, 34] and an effective knee at r* ~ 31. At rank 32, the student attains 69.46% top-1 accuracy vs. 73.35% for the teacher (~94.7% of baseline) while achieving substantial parameter compression. We provide a framework to estimate effective-rank regions and knees across architectures and datasets, offering a practical tool for characterizing the intrinsic dimensionality of deep models.
[691] Soft Quality-Diversity Optimization
Saeed Hedayatian, Stefanos Nikolaidis
Main category: cs.LG
TL;DR: Soft QD reformulates Quality-Diversity optimization without discretization, enabling better scalability to high-dimensional problems through a differentiable algorithm called SQUAD.
Details
Motivation: Current QD methods face challenges in large solution spaces (storage impractical) and high-dimensional behavior spaces (discretization ineffective due to curse of dimensionality). The paper aims to overcome these limitations by developing a QD approach that doesn't require discretization.Method: Proposes Soft QD, an alternative framing of QD problems that eliminates the need for discretization. Derives a novel differentiable QD algorithm called SQUAD (Soft QD Using Approximated Diversity) based on this formulation.
Result: SQUAD demonstrates competitive performance with state-of-the-art methods on standard benchmarks while offering better scalability to higher dimensional problems. The Soft QD formulation shows desirable properties like monotonicity and relates to the widely used QD Score metric.
Conclusion: Soft QD provides a promising alternative to traditional discretization-based QD methods, particularly for high-dimensional problems where current approaches struggle due to the curse of dimensionality.
Abstract: Quality-Diversity (QD) algorithms constitute a branch of optimization that is concerned with discovering a diverse and high-quality set of solutions to an optimization problem. Current QD methods commonly maintain diversity by dividing the behavior space into discrete regions, ensuring that solutions are distributed across different parts of the space. The QD problem is then solved by searching for the best solution in each region. This approach to QD optimization poses challenges in large solution spaces, where storing many solutions is impractical, and in high-dimensional behavior spaces, where discretization becomes ineffective due to the curse of dimensionality. We present an alternative framing of the QD problem, called \emph{Soft QD}, that sidesteps the need for discretizations. We validate this formulation by demonstrating its desirable properties, such as monotonicity, and by relating its limiting behavior to the widely used QD Score metric. Furthermore, we leverage it to derive a novel differentiable QD algorithm, \emph{Soft QD Using Approximated Diversity (SQUAD)}, and demonstrate empirically that it is competitive with current state of the art methods on standard benchmarks while offering better scalability to higher dimensional problems.
[692] Causal Invariance and Counterfactual Learning Driven Cooperative Game for Multi-Label Classification
Yijia Fan, Jusheng Zhang, Kaitong Cai, Jing Yang, Keze Wang
Main category: cs.LG
TL;DR: CCG framework combines causal inference with cooperative game theory to address label imbalance, spurious correlations, and distribution shifts in multi-label classification, particularly improving rare label prediction.
Details
Motivation: Multi-label classification suffers from label imbalance, spurious correlations, and distribution shifts that particularly harm rare label prediction performance. Current approaches lack robustness to these challenges.Method: Causal Cooperative Game (CCG) framework treats MLC as cooperative multi-player interaction, combining Neural Structural Equation Models for causal discovery, counterfactual curiosity reward for feature learning, causal invariance loss for generalization, and specialized rare label enhancement.
Result: CCG substantially outperforms strong baselines in both rare label prediction and overall robustness, with ablation studies and qualitative analysis validating component efficacy and interpretability.
Conclusion: Synergizing causal inference with cooperative game theory shows strong potential for advancing multi-label learning by addressing fundamental challenges of label imbalance and distribution shifts.
Abstract: Multi-label classification (MLC) remains vulnerable to label imbalance, spurious correlations, and distribution shifts, challenges that are particularly detrimental to rare label prediction. To address these limitations, we introduce the Causal Cooperative Game (CCG) framework, which conceptualizes MLC as a cooperative multi-player interaction. CCG unifies explicit causal discovery via Neural Structural Equation Models with a counterfactual curiosity reward to drive robust feature learning. Furthermore, it incorporates a causal invariance loss to ensure generalization across diverse environments, complemented by a specialized enhancement strategy for rare labels. Extensive benchmarking demonstrates that CCG substantially outperforms strong baselines in both rare label prediction and overall robustness. Through rigorous ablation studies and qualitative analysis, we validate the efficacy and interpretability of our components, underscoring the potential of synergizing causal inference with cooperative game theory for advancing multi-label learning.
[693] ReJump: A Tree-Jump Representation for Analyzing and Improving LLM Reasoning
Yuchen Zeng, Shuibai Zhang, Wonjun Kang, Shutong Wu, Lynnix Zou, Ying Fan, Heeju Kim, Ziqian Lin, Jungtaek Kim, Hyung Il Koo, Dimitris Papailiopoulos, Kangwook Lee
Main category: cs.LG
TL;DR: ReJump is a framework for analyzing Large Reasoning Models by representing reasoning traces as tree node visitation orders, enabling quantification of various reasoning behaviors like exploration, exploitation, and verification.
Details
Motivation: Large Reasoning Models achieve impressive performance on complex tasks but their underlying reasoning "algorithms" remain poorly understood, creating a need for systematic analysis tools.Method: Proposes ReJump framework that represents reasoning traces as visitation orders over tree nodes, with transitions (jumps) capturing behaviors like calculation, backtracking, and verification. Uses an LLM agent to extract reasoning traces into ReJump format for analysis.
Result: Models with similar accuracy exhibit distinct reasoning behaviors; different tasks favor different reasoning styles; distilled LRMs differ from teachers; CoT-prompted LLMs differ from LRMs; ReJump can improve reasoning quality through Best-of-N selection and prompt selection.
Conclusion: ReJump provides a systematic framework for analyzing and understanding reasoning behaviors in LRMs, revealing diverse reasoning strategies and enabling improvements in reasoning quality through guided selection methods.
Abstract: Large Reasoning Models (LRMs) are Large Language Models (LLMs) explicitly trained to generate long-form Chain-of-Thoughts (CoTs), achieving impressive success on challenging tasks like math and programming. However, their underlying reasoning “algorithms” remain poorly understood. To investigate this, we propose ReJump, which represents a reasoning trace as a visitation order over nodes in a tree of intermediate problem-solving steps. Transitions between nodes, which we term jumps, include adjacent moves that capture behaviors such as calculation, and non-adjacent moves that capture behaviors such as backtracking and verification. ReJump enables analyzing LLM reasoning with diverse metrics that quantify exploration, exploitation, overthinking, forgetting, and verification. Using our proposed LLM agent to extract reasoning traces into ReJump format, we evaluate state-of-the-art LRMs on two tasks and find that models with similar accuracy can exhibit distinct reasoning behaviors, while different tasks favor different reasoning styles (e.g., varying balance between exploration and exploitation). To further understand how learning strategies shape reasoning, we use ReJump to compare distilled LRMs with their teachers, CoT-prompted LLMs with LRMs, and to examine how the number of reasoning examples and reinforcement learning affect reasoning behavior. Finally, we show that ReJump can improve reasoning quality at test time through strategies such as ReJump-guided Best-of-N selection and prompt selection. Our code is publicly available at https://github.com/UW-Madison-Lee-Lab/ReJump.
[694] Uncertainty Quantification for Deep Regression using Contextualised Normalizing Flows
Adriel Sosa Marco, John Daniel Kirwan, Alexia Toumpa, Simos Gerasimou
Main category: cs.LG
TL;DR: MCNF is a post hoc uncertainty quantification method that provides both prediction intervals and full predictive distributions without requiring model retraining.
Details
Motivation: Existing uncertainty quantification methods either overlook distributional information (ignoring multimodal/asymmetric distributions) or require major architectural changes and retraining of models.Method: MCNF operates as a post hoc method on top of trained predictive models, requiring no retraining of the underlying model while producing both prediction intervals and full conditioned predictive distributions.
Result: MCNF provides well-calibrated uncertainty estimates, competitive with state-of-the-art methods, and offers richer distributional information for downstream decision-making tasks.
Conclusion: MCNF is a practical uncertainty quantification approach that addresses limitations of existing methods by providing comprehensive distributional information without model modifications.
Abstract: Quantifying uncertainty in deep regression models is important both for understanding the confidence of the model and for safe decision-making in high-risk domains. Existing approaches that yield prediction intervals overlook distributional information, neglecting the effect of multimodal or asymmetric distributions on decision-making. Similarly, full or approximated Bayesian methods, while yielding the predictive posterior density, demand major modifications to the model architecture and retraining. We introduce MCNF, a novel post hoc uncertainty quantification method that produces both prediction intervals and the full conditioned predictive distribution. MCNF operates on top of the underlying trained predictive model; thus, no predictive model retraining is needed. We provide experimental evidence that the MCNF-based uncertainty estimate is well calibrated, is competitive with state-of-the-art uncertainty quantification methods, and provides richer information for downstream decision-making tasks.
[695] Prediction-space knowledge markets for communication-efficient federated learning on multimedia tasks
Wenzhang Du
Main category: cs.LG
TL;DR: KTA v2 is a federated learning method that uses prediction-space knowledge trading with logit sharing and personalized teacher ensembles to achieve high accuracy with dramatically reduced communication costs compared to traditional parameter-based FL methods.
Details
Motivation: Federated learning suffers from statistical heterogeneity and communication constraints, especially with large models. Classic parameter-averaging methods like FedAvg transmit full model weights and can diverge under non-IID data, creating a need for more efficient and robust FL approaches.Method: KTA v2 uses a two-stage knowledge trading market: 1) Clients locally train on private data and share only logits on a small public reference set, 2) Server constructs client-client similarity graphs in prediction space, combines with reference-set accuracy to form personalized teacher ensembles, and sends back soft targets for a second-stage distillation update. This can be interpreted as approximate block-coordinate descent with prediction-space regularization.
Result: KTA v2 consistently outperforms local-only baselines and strong parameter-based methods (FedAvg, FedProx), and substantially improves over FedMD-style global teacher. On CIFAR-10 with ResNet-18, achieves 57.7% test accuracy using ~1/1100 of FedAvg’s communication. On AG News, attains 89.3% accuracy with ~1/300 of FedAvg’s traffic.
Conclusion: KTA v2 demonstrates that prediction-space knowledge trading with personalized teacher ensembles is an effective approach for federated learning, achieving strong performance with dramatically reduced communication overhead while handling statistical heterogeneity better than parameter-based methods.
Abstract: Federated learning (FL) enables collaborative training over distributed multimedia data but suffers acutely from statistical heterogeneity and communication constraints, especially when clients deploy large models. Classic parameter-averaging methods such as FedAvg transmit full model weights and can diverge under nonindependent and identically distributed (non-IID) data. We propose KTA v2, a prediction-space knowledge trading market for FL. Each round, clients locally train on their private data, then share only logits on a small public reference set. The server constructs a client-client similarity graph in prediction space, combines it with reference-set accuracy to form per-client teacher ensembles, and sends back personalized soft targets for a second-stage distillation update. This two-stage procedure can be interpreted as approximate block-coordinate descent on a unified objective with prediction-space regularization. Experiments on FEMNIST, CIFAR-10 and AG News show that, under comparable or much lower communication budgets, KTA v2 consistently outperforms a local-only baseline and strong parameter-based methods (FedAvg, FedProx), and substantially improves over a FedMD-style global teacher. On CIFAR-10 with ResNet-18, KTA v2 reaches 57.7% test accuracy using approximately 1/1100 of FedAvg’s communication, while on AG News it attains 89.3% accuracy with approximately 1/300 of FedAvg’s traffic.
[696] Topological Federated Clustering via Gravitational Potential Fields under Local Differential Privacy
Yunbo Long, Jiaquan Zhang, Xi Chen, Alexandra Brintrup
Main category: cs.LG
TL;DR: GFC is a novel one-shot federated clustering method using gravitational potential fields and topological persistence to handle non-IID data under local differential privacy, achieving better privacy-accuracy trade-offs than distance-based approaches.
Details
Motivation: Existing one-shot federated clustering methods under LDP suffer from instability with pairwise centroid distances or neighborhood rankings, degrading severely under strong LDP noise and data heterogeneity. There's a need for methods that preserve privacy while maintaining accuracy without iterative communication.Method: GFC transforms privatized client centroids into a global gravitational potential field where true cluster centers emerge as topologically persistent singularities. Key innovations: (1) client-side compactness-aware perturbation encoding local cluster geometry as “mass” values, and (2) server-side topological aggregation extracting stable centroids through persistent homology analysis of the potential field’s superlevel sets.
Result: Theoretically establishes closed-form bound between privacy budget ε and centroid estimation error, proving potential field’s Lipschitz smoothing properties exponentially suppress noise in high-density regions. Empirically outperforms state-of-the-art methods on ten benchmarks, especially under strong LDP constraints (ε < 1), while maintaining comparable performance at lower privacy budgets.
Conclusion: By reformulating federated clustering as a topological persistence problem in a physics-inspired space, GFC achieves unprecedented privacy-accuracy trade-offs without iterative communication, providing a new perspective for privacy-preserving distributed learning.
Abstract: Clustering non-independent and identically distributed (non-IID) data under local differential privacy (LDP) in federated settings presents a critical challenge: preserving privacy while maintaining accuracy without iterative communication. Existing one-shot methods rely on unstable pairwise centroid distances or neighborhood rankings, degrading severely under strong LDP noise and data heterogeneity. We present Gravitational Federated Clustering (GFC), a novel approach to privacy-preserving federated clustering that overcomes the limitations of distance-based methods under varying LDP. Addressing the critical challenge of clustering non-IID data with diverse privacy guarantees, GFC transforms privatized client centroids into a global gravitational potential field where true cluster centers emerge as topologically persistent singularities. Our framework introduces two key innovations: (1) a client-side compactness-aware perturbation mechanism that encodes local cluster geometry as “mass” values, and (2) a server-side topological aggregation phase that extracts stable centroids through persistent homology analysis of the potential field’s superlevel sets. Theoretically, we establish a closed-form bound between the privacy budget $ε$ and centroid estimation error, proving the potential field’s Lipschitz smoothing properties exponentially suppress noise in high-density regions. Empirically, GFC outperforms state-of-the-art methods on ten benchmarks, especially under strong LDP constraints ($ε< 1$), while maintaining comparable performance at lower privacy budgets. By reformulating federated clustering as a topological persistence problem in a synthetic physics-inspired space, GFC achieves unprecedented privacy-accuracy trade-offs without iterative communication, providing a new perspective for privacy-preserving distributed learning.
[697] City-Conditioned Memory for Multi-City Traffic and Mobility Forecasting
Wenzhang Du
Main category: cs.LG
TL;DR: CityCond is a lightweight city-conditioned memory layer that improves spatio-temporal forecasting across cities with varying data availability, working with multiple backbones and showing strongest gains for high-capacity models.
Details
Motivation: Deploying traffic forecasting models across multiple cities is challenging due to differences in network size/topology, varying data availability, and limited historical data for new cities. Existing per-city training approaches have high maintenance costs and poor transfer to data-scarce cities.Method: CityCond combines a city-ID encoder with an optional shared memory bank (CityMem). It takes a city index and backbone hidden states to produce city-conditioned features fused through gated residual connections. The layer can be attached to various backbones (GRU, TCN, Transformer, GNN, STGCN).
Result: CityCond yields consistent improvements across 14+ model variants and 3 random seeds, with largest gains for high-capacity backbones like Transformers and STGCNs. CityMem reduces Transformer error by ~1/3 in full-data settings and brings substantial gains in low-data and cross-city transfer.
Conclusion: CityCond serves as a reusable design pattern for scalable, multi-city forecasting under realistic data constraints, enabling better performance across full-data, low-data, and cross-city few-shot transfer scenarios.
Abstract: Deploying spatio-temporal forecasting models across many cities is difficult: traffic networks differ in size and topology, data availability can vary by orders of magnitude, and new cities may provide only a short history of logs. Existing deep traffic models are typically trained per city and backbone, creating high maintenance cost and poor transfer to data-scarce cities. We ask whether a single, backbone-agnostic layer can condition on “which city this sequence comes from”, improve accuracy in full- and low-data regimes, and support better cross-city adaptation with minimal code changes. We propose CityCond, a light-weight city-conditioned memory layer that augments existing spatio-temporal backbones. CityCond combines a city-ID encoder with an optional shared memory bank (CityMem). Given a city index and backbone hidden states, it produces city-conditioned features fused through gated residual connections. We attach CityCond to five representative backbones (GRU, TCN, Transformer, GNN, STGCN) and evaluate three regimes: full-data, low-data, and cross-city few-shot transfer on METR-LA and PEMS-BAY. We also run auxiliary experiments on SIND, a drone-based multi-agent trajectory dataset from a signalized intersection in Tianjin (we focus on pedestrian tracks). Across more than fourteen model variants and three random seeds, CityCond yields consistent improvements, with the largest gains for high-capacity backbones such as Transformers and STGCNs. CityMem reduces Transformer error by roughly one third in full-data settings and brings substantial gains in low-data and cross-city transfer. On SIND, simple city-ID conditioning modestly improves low-data LSTM performance. CityCond can therefore serve as a reusable design pattern for scalable, multi-city forecasting under realistic data constraints.
[698] Robust Probabilistic Load Forecasting for a Single Household: A Comparative Study from SARIMA to Transformers on the REFIT Dataset
Midhun Manoj
Main category: cs.LG
TL;DR: The paper addresses probabilistic forecasting for risk management using volatile household data with structural gaps, finding that advanced deep learning models (TFT and LSTM) outperform classical methods in handling non-linear, regime-switching behavior.
Details
Motivation: Probabilistic forecasting is crucial for modern risk management to quantify uncertainty in critical systems, but existing methods struggle with volatile datasets containing structural data gaps like the REFIT household dataset.Method: First conducted comparative experiments to select Seasonal Imputation method over linear interpolation to handle data gaps, then systematically evaluated a hierarchy of models from classical baselines (SARIMA, Prophet) to machine learning (XGBoost) and advanced deep learning (LSTM, Temporal Fusion Transformer).
Result: Classical models failed to capture non-linear, regime-switching behavior. LSTM provided the most well-calibrated probabilistic forecast, while TFT emerged as the superior all-round model with best point forecast accuracy (RMSE 481.94) and produced safer, more cautious prediction intervals that effectively captured extreme volatility.
Conclusion: Advanced deep learning architectures, particularly Temporal Fusion Transformer, are superior for probabilistic forecasting on volatile datasets with structural gaps, offering both accurate point forecasts and reliable uncertainty quantification for risk management applications.
Abstract: Probabilistic forecasting is essential for modern risk management, allowing decision-makers to quantify uncertainty in critical systems. This paper tackles this challenge using the volatile REFIT household dataset, which is complicated by a large structural data gap. We first address this by conducting a rigorous comparative experiment to select a Seasonal Imputation method, demonstrating its superiority over linear interpolation in preserving the data’s underlying distribution. We then systematically evaluate a hierarchy of models, progressing from classical baselines (SARIMA, Prophet) to machine learning (XGBoost) and advanced deep learning architectures (LSTM). Our findings reveal that classical models fail to capture the data’s non-linear, regime-switching behavior. While the LSTM provided the most well-calibrated probabilistic forecast, the Temporal Fusion Transformer (TFT) emerged as the superior all-round model, achieving the best point forecast accuracy (RMSE 481.94) and producing safer, more cautious prediction intervals that effectively capture extreme volatility.
[699] The Spectral Dimension of NTKs is Constant: A Theory of Implicit Regularization, Finite-Width Stability, and Scalable Estimation
Praveen Anilkumar Shukla
Main category: cs.LG
TL;DR: The paper studies the intrinsic complexity of overparameterized deep networks through the effective rank of the Neural Tangent Kernel Gram matrix, showing it converges to a constant limit for infinite-width networks and remains stable for finite-width networks.
Details
Motivation: Modern deep networks are heavily overparameterized but generalize well, suggesting low intrinsic complexity not captured by parameter counts. The authors aim to quantify this complexity at initialization through the effective rank of the NTK Gram matrix.Method: 1) Theoretical analysis of effective rank for infinite-width NTK with i.i.d. data, proving constant-limit law. 2) Finite-width stability analysis showing effective rank changes by O_p(m^{-1/2}) with width m. 3) Design of scalable estimator using random output probes and CountSketch of parameter Jacobians with proven unbiasedness and consistency.
Result: 1) Proved constant-limit law for effective rank with sub-Gaussian concentration. 2) Established finite-width stability bounds. 3) Experimental results on CIFAR-10 with ResNet-20/56 show effective rank ≈1.0-1.3 with near-zero slopes in n, consistent with theory, and kernel-moment predictions match fitted constants.
Conclusion: The effective rank of NTK Gram matrices provides a meaningful measure of network complexity that remains low and stable, explaining why overparameterized networks can generalize well despite having many parameters.
Abstract: Modern deep networks are heavily overparameterized yet often generalize well, suggesting a form of low intrinsic complexity not reflected by parameter counts. We study this complexity at initialization through the effective rank of the Neural Tangent Kernel (NTK) Gram matrix, $r_{\text{eff}}(K) = (\text{tr}(K))^2/|K|F^2$. For i.i.d. data and the infinite-width NTK $k$, we prove a constant-limit law $\lim{n\to\infty} \mathbb{E}[r_{\text{eff}}(K_n)] = \mathbb{E}[k(x, x)]^2 / \mathbb{E}[k(x, x’)^2] =: r_\infty$, with sub-Gaussian concentration. We further establish finite-width stability: if the finite-width NTK deviates in operator norm by $O_p(m^{-1/2})$ (width $m$), then $r_{\text{eff}}$ changes by $O_p(m^{-1/2})$. We design a scalable estimator using random output probes and a CountSketch of parameter Jacobians and prove conditional unbiasedness and consistency with explicit variance bounds. On CIFAR-10 with ResNet-20/56 (widths 16/32) across $n \in {10^3, 5\times10^3, 10^4, 2.5\times10^4, 5\times10^4}$, we observe $r_{\text{eff}} \approx 1.0\text{–}1.3$ and slopes $\approx 0$ in $n$, consistent with the theory, and the kernel-moment prediction closely matches fitted constants.
[700] Towards Active Synthetic Data Generation for Finetuning Language Models
Samuel Kessler, Menglin Xia, Daniel Madrigal Diaz, Dongge Han, Helia Heshemi, Saravan Rajmohan, Victor Ruehle, Jordan T. Ash
Main category: cs.LG
TL;DR: Iterative synthetic data generation guided by student model state outperforms static generation for language model finetuning, with simple active learning selection criteria being most effective.
Details
Motivation: Current synthetic data generation for language model finetuning typically happens before training (static), but iterative generation guided by the student's current state could be more effective for improving model capabilities.Method: Closed-loop iterative synthetic data generation where teacher model generates samples guided by current student state, using simple active learning selection criteria to curate finetuning data.
Result: Iterative generation with active learning selection outperforms static generation for fixed compute budgets, validated across four mathematical/logical reasoning datasets with four different small language models.
Conclusion: Iterative synthetic data generation guided by student state is more effective than static generation, and simple active learning methods outperform more complex LLM-specific approaches for data selection.
Abstract: A common and effective means for improving language model capabilities involves finetuning a student'' language model's parameters on generations from a more proficient teacher’’ model. Termed ``synthetic data’’, these generations are often produced before any student finetuning, but some work has considered generating new synthetic samples as training progresses. This paper studies and advocates for the latter case, where data are generated in an iterative, closed-loop fashion that is guided by the current state of the student model. For a fixed budget of generated samples, or a budget in terms of compute spent querying a teacher, we show that this curation of finetuning data affords improved student performance over static generation. Further, while there have been several LLM-specific methods proposed that operate in this regime, we find that simple, inexpensive selection criteria from the active learning literature tend to be most performant. We validate these claims across four mathematical and logical reasoning datasets using four different small language models.
[701] HBLLM: Wavelet-Enhanced High-Fidelity 1-Bit Quantization for LLMs
Ningning Chen, Weicai Ye, Ying Jiang
Main category: cs.LG
TL;DR: HBLLM is a wavelet-enhanced 1-bit post-training quantization method for LLMs that uses Haar wavelet transforms and structure-aware grouping to achieve high-fidelity quantization with minimal storage overhead.
Details
Motivation: To develop an efficient 1-bit quantization method for LLMs that maintains high model fidelity while significantly reducing storage requirements, addressing the challenge of deploying large models in resource-constrained environments.Method: Uses Haar wavelet transforms for frequency decomposition, with two structure-aware grouping strategies: (1) frequency-aware multi-parameter intra-row grouping, and (2) ℓ₂-norm-based saliency-driven column selection. For non-salient weights, employs shared mean quantization across groups within each frequency band.
Result: Achieves state-of-the-art 1-bit quantization performance: perplexity of 6.71 on LLaMA2-13B with average weight storage of only 1.08 bits. Demonstrated effectiveness on OPT and LLaMA models.
Conclusion: HBLLM provides an effective wavelet-enhanced quantization approach that significantly improves fidelity in 1-bit LLM quantization while maintaining minimal storage overhead, enabling more efficient deployment of large language models.
Abstract: We introduce HBLLM, a wavelet-enhanced high-fidelity $1$-bit post-training quantization method for Large Language Models (LLMs). By leveraging Haar wavelet transforms to enhance expressive capacity through frequency decomposition, HBLLM significantly improves quantization fidelity while maintaining minimal overhead. This approach features two innovative structure-aware grouping strategies: (1) frequency-aware multi-parameter intra-row grouping and (2) $\ell_2$-norm-based saliency-driven column selection. For non-salient weights, a shared mean is employed across quantization groups within each frequency band to optimize storage efficiency. Experiments conducted on the OPT and LLaMA models demonstrate that HBLLM achieves state-of-the-art performance in $1$-bit quantization, attaining a perplexity of $6.71$ on LLaMA$2$-$13$B with an average weight storage of only $1.08$ bits. Code available at: https://github.com/Yeyke/HBLLM.
[702] Associative Syntax and Maximal Repetitions reveal context-dependent complexity in fruit bat communication
Luigi Assom
Main category: cs.LG
TL;DR: Unsupervised method analyzes fruit-bat vocalizations, finding associative syntax, context-dependent syllable use, and heavy-tail distributions indicating combinatorial complexity, with higher complexity in conflict contexts.
Details
Motivation: To develop unsupervised methods for analyzing graded vocal systems (using fruit-bats as case study) and evaluate communication complexity in relation to behavioral context.Method: 1) Improved unsupervised vocal unit labeling via manifold learning on mel-spectrograms; 2) Encoded vocalizations as syllabic sequences to analyze syntax; 3) Extracted Maximal Repetitions (MRs) to evaluate syntactical structures; 4) Compared unsupervised labels based on acoustic similarity.
Result: Found: i) Associative syntax (F1 > 0.9); ii) Context-dependent syllable use (p < 0.05); iii) Heavy-tail MR distribution (truncated power-law, α < 2); iv) Mother-pup interactions characterized by repetitions; v) Conflict contexts show higher complexity (longer MRs, more interconnected sequences) than non-agonistic contexts.
Conclusion: Communicative complexity is higher in disagreement scenarios, reflecting lower compressibility of information, with unsupervised methods revealing structured patterns in graded vocal systems.
Abstract: This study presents an unsupervised method to infer discreteness, syntax and temporal structures of fruit-bats vocalizations, as a case study of graded vocal systems, and evaluates the complexity of communication patterns in relation with behavioral context. The method improved the baseline for unsupervised labeling of vocal units (i.e. syllables) through manifold learning, by investigating how dimen- sionality reduction on mel-spectrograms affects labeling, and comparing it with unsupervised labels based on acoustic similarity. We then encoded vocalizations as syllabic sequences to analyze the type of syntax, and extracted the Maximal Repetitions (MRs) to evaluate syntactical structures. We found evidence for: i) associative syntax, rather than combinatorial (context classification is unaffected by permutation of sequences, F 1 > 0.9); ii) context-dependent use of syllables (Wilcoxon rank-sum tests, p-value < 0.05); iii) heavy-tail distribution of MRs (truncated power-law, exponent α < 2), indicative of mechanism encoding com- binatorial complexity. Analysis of MRs and syllabic transition networks revealed that mother-pupil interactions were characterized by repetitions, while commu- nication in conflict-contexts exhibited higher complexity (longer MRs and more interconnected vocal sequences) than non-agonistic contexts. We propose that communicative complexity is higher in scenarios of disagreement, reflecting lower compressibility of information.
[703] Light-Weight Benchmarks Reveal the Hidden Hardware Cost of Zero-Shot Tabular Foundation Models
Aayam Bansal, Ishaan Gangwani
Main category: cs.LG
TL;DR: Zero-shot tabular foundation models (FMs) show substantial hardware-accuracy trade-offs compared to traditional tree ensembles, requiring 40,000x more latency and 9GB VRAM for marginal accuracy gains.
Details
Motivation: To characterize the hardware footprint of zero-shot foundation models for tabular data prediction, which remains poorly understood despite their promise of training-free operation.Method: A fully reproducible benchmark comparing two open FMs (TabPFN-1.0 and TabICL-base) against tuned XGBoost, LightGBM, and Random Forest baselines on four public datasets (Adult-Income, Higgs-100k, Wine-Quality, California-Housing), measuring test accuracy, wall-clock latency, peak CPU RAM, and peak GPU VRAM on a single NVIDIA T4 GPU.
Result: Tree ensembles equal or surpass FM accuracy on three datasets while being dramatically more efficient (≤0.40s latency, ≤150MB RAM, zero VRAM). TabICL achieved only 0.8 percentage-point gain on Higgs but required ~40,000x more latency (960s) and 9GB VRAM. TabPFN matched tree-model accuracy on Wine and Housing but peaked at 4GB VRAM and couldn’t process the full 100k-row Higgs table.
Conclusion: Current tabular FMs impose substantial hardware-versus-accuracy trade-offs compared to traditional tree ensembles, highlighting the need for efficiency-oriented research in foundation models for tabular data.
Abstract: Zero-shot foundation models (FMs) promise training-free prediction on tabular data, yet their hardware footprint remains poorly characterized. We present a fully reproducible benchmark that reports test accuracy together with wall-clock latency, peak CPU RAM, and peak GPU VRAM on four public datasets: Adult-Income, Higgs-100k, Wine-Quality, and California-Housing. Two open FMs (TabPFN-1.0 and TabICL-base) are compared against tuned XGBoost, LightGBM, and Random Forest baselines on a single NVIDIA T4 GPU. The tree ensembles equal or surpass FM accuracy on three datasets while completing full-test batches in <= 0.40 s and <= 150 MB RAM, using zero VRAM. TabICL achieves a 0.8 percentage-point gain on Higgs but requires roughly 40,000 times more latency (960 s) and 9 GB VRAM. TabPFN matches tree-model accuracy on Wine and Housing but peaks at 4 GB VRAM and cannot process the full 100k-row Higgs table. These results quantify the substantial hardware-versus-accuracy trade-offs in current tabular FMs and provide an open baseline for future efficiency-oriented research.
[704] Beyond High-Entropy Exploration: Correctness-Aware Low-Entropy Segment-Based Advantage Shaping for Reasoning LLMs
Xinzhu Chen, Xuesheng Li, Zhongxiang Sun, Weijie Yu
Main category: cs.LG
TL;DR: LESS is a reinforcement learning framework that modulates advantages for low-entropy segments in reasoning trajectories, improving accuracy and robustness in math reasoning tasks.
Details
Motivation: Current RLVR approaches focus on high-entropy tokens for exploration but overlook that most reasoning trajectories consist of low-entropy segments encoding stable structural patterns. The overlap of low-entropy segments across correct responses correlates with model accuracy, while incorrect responses exhibit stable but unproductive patterns.Method: Proposed LESS (Low-Entropy Segment Suppression) framework that performs fine-grained advantage modulation: amplifies segments unique to correct responses, suppresses those unique to incorrect ones, neutralizes segments shared by both, while preserving high-entropy exploration in the underlying RL algorithm.
Result: LESS consistently improves accuracy over strong RL baselines across three backbones and six math benchmarks, and achieves stronger robustness of the performance floor when instantiated on top of GRPO.
Conclusion: Low-entropy segments in reasoning trajectories contain valuable structural patterns that should be modulated based on correctness, leading to improved performance in reinforcement learning for reasoning tasks.
Abstract: Reinforcement Learning with Verifiable Rewards (RLVR) has become a central approach for improving the reasoning ability of large language models. Recent work studies RLVR through token entropy, arguing that high-entropy tokens drive exploration and should receive stronger updates. However, they overlook the fact that most of a reasoning trajectory consists of low-entropy segments that encode stable and reusable structural patterns. Through qualitative and quantitative analyses, we find that the overlap of low-entropy segments across correct responses strongly correlates with model accuracy, while overlaps involving incorrect responses exhibit stable but unproductive patterns. Motivated by these findings, we propose LESS, a correctness-aware reinforcement framework that performs fine-grained advantage modulation over low-entropy segments. LESS amplifies segments unique to correct responses, suppresses those unique to incorrect ones, and neutralizes segments shared by both, while preserving high-entropy exploration in the underlying RL algorithm. Instantiated on top of the popular GRPO, LESS consistently improves accuracy over strong RL baselines across three backbones and six math benchmarks, achieves stronger robustness of the performance floor.
[705] Mode-Conditioning Unlocks Superior Test-Time Scaling
Chen Henry Wu, Sachin Goyal, Aditi Raghunathan
Main category: cs.LG
TL;DR: ModC framework addresses diversity collapse in parallel sampling by explicitly allocating compute across reasoning modes, achieving significant efficiency gains and improved scaling across various reasoning benchmarks.
Details
Motivation: Parallel sampling promises substantial test-time scaling gains, but its effectiveness is limited by diversity collapse where models concentrate on few modes and repeat the same mistakes, underutilizing data diversity.Method: Propose mode-conditioning (ModC) framework that explicitly allocates test-time compute across reasoning modes using either specialist models or mode-specific prefixes. Also show gradient clustering enables ModC without explicit mode labels.
Result: ModC consistently improves scaling across controlled graph-search tasks and large-scale reasoning benchmarks (0.5B to 7B models). On OpenThoughts, fine-tuning Qwen2.5-7B with ModC achieves 4x efficiency gain over standard training while improving maximum Pass@k. Gradient clustering yields up to 10% gains on NuminaMath. ModC also improves reinforcement learning and boosts diversity-inducing RL methods.
Conclusion: Standard training underutilizes data diversity, and ModC provides a simple, effective remedy for unlocking the full benefits of diversity in test-time scaling, addressing the fundamental limitation of diversity collapse in parallel sampling.
Abstract: Parallel sampling promises substantial gains in test-time scaling, but its effectiveness is sharply limited by diversity collapse, where models concentrate on a few modes and repeated samples produce the same mistakes. We propose the mode-conditioning (ModC) framework, which explicitly allocates test-time compute across reasoning modes using either specialist models or mode-specific prefixes. ModC consistently improves scaling across controlled graph-search tasks and large-scale reasoning benchmarks, spanning model families and sizes from 0.5B to 7B. On OpenThoughts, fine-tuning Qwen2.5-7B with ModC achieves a 4x efficiency gain over standard training while also improving the maximum attainable Pass@k. We further show that gradient clustering enables ModC without explicit mode labels, yielding up to 10% gains on datasets such as NuminaMath. Finally, we show that ModC improves reinforcement learning (RL) and can further boost diversity-inducing RL methods. These results demonstrate that standard training underutilizes the diversity in data, and that ModC provides a simple, effective remedy for unlocking the full benefits of diversity in test-time scaling.
[706] Partially Equivariant Reinforcement Learning in Symmetry-Breaking Environments
Junwoo Chang, Minwoo Park, Joohwan Seo, Roberto Horowitz, Jongmin Lee, Jongeun Choi
Main category: cs.LG
TL;DR: The paper introduces Partially group-Invariant MDP (PI-MDP) framework and practical RL algorithms (PE-DQN and PE-SAC) that selectively apply group-invariant or standard Bellman backups to handle local symmetry-breaking in real-world environments while maintaining equivariance benefits.
Details
Motivation: Real-world environments rarely have perfect group symmetries due to dynamics, actuation limits, and reward design that break symmetries locally. Standard group-invariant approaches propagate errors from local symmetry-breaking across the entire state-action space, causing global value estimation errors.Method: Introduces Partially group-Invariant MDP (PI-MDP) framework that selectively applies group-invariant Bellman backups where symmetry holds and standard Bellman backups where symmetry is broken. Develops practical RL algorithms: PE-DQN for discrete control and PE-SAC for continuous control.
Result: Experiments across Grid-World, locomotion, and manipulation benchmarks show that PE-DQN and PE-SAC significantly outperform baselines, demonstrating improved sample efficiency and generalizability.
Conclusion: Selective symmetry exploitation through the PI-MDP framework enables robust and sample-efficient RL by mitigating error propagation from local symmetry-breaking while maintaining equivariance benefits where appropriate.
Abstract: Group symmetries provide a powerful inductive bias for reinforcement learning (RL), enabling efficient generalization across symmetric states and actions via group-invariant Markov Decision Processes (MDPs). However, real-world environments almost never realize fully group-invariant MDPs; dynamics, actuation limits, and reward design usually break symmetries, often only locally. Under group-invariant Bellman backups for such cases, local symmetry-breaking introduces errors that propagate across the entire state-action space, resulting in global value estimation errors. To address this, we introduce Partially group-Invariant MDP (PI-MDP), which selectively applies group-invariant or standard Bellman backups depending on where symmetry holds. This framework mitigates error propagation from locally broken symmetries while maintaining the benefits of equivariance, thereby enhancing sample efficiency and generalizability. Building on this framework, we present practical RL algorithms – Partially Equivariant (PE)-DQN for discrete control and PE-SAC for continuous control – that combine the benefits of equivariance with robustness to symmetry-breaking. Experiments across Grid-World, locomotion, and manipulation benchmarks demonstrate that PE-DQN and PE-SAC significantly outperform baselines, highlighting the importance of selective symmetry exploitation for robust and sample-efficient RL.
[707] Pay Attention Later: From Vector Space Diffusion to Linearithmic Spectral Phase-Locking
Alper Yıldırım, İbrahim Yücedağ
Main category: cs.LG
TL;DR: PRISM replaces standard Transformers with harmonic representations to solve catastrophic forgetting, achieving lossless plasticity for novel concepts while maintaining reasoning capabilities.
Details
Motivation: Standard Transformers suffer from "Semantic Alignment Tax" - high optimization costs to organize chaotic initialization, and "Catastrophic Rigidity" - inability to adapt to novel concepts without destroying pre-trained capabilities. This creates a plasticity-stability dilemma.Method: Introduces Phase-Resonant Intelligent Spectral Model (PRISM) that encodes semantic identity as resonant frequencies in complex domain (C^d) and replaces quadratic self-attention with O(N log N) Gated Harmonic Convolutions. Also introduces Iterative Semantic Map Refinement (ISMR) diagnostic protocol.
Result: On WMT14 translation: Standard Transformer slightly better on static benchmarks (23.88 vs 21.40 BLEU) but fails plasticity-stability test (-10.55 BLEU degradation, 60% acquisition). PRISM achieves 96% 5-shot acquisition with minimal degradation (-0.84 BLEU), demonstrating Lossless Plasticity.
Conclusion: Harmonic representations effectively decouple memory from reasoning, offering structural solution to plasticity-stability dilemma for real-time knowledge adaptation. Scaling cannot solve alignment barrier (20-layer no faster than 1-layer).
Abstract: Standard Transformers suffer from a “Semantic Alignment Tax”, a prohibitive optimization cost required to organize a chaotic initialization into a coherent geometric map via local gradient diffusion. We hypothesize that this reliance on diffusive learning creates “Catastrophic Rigidity”, rendering models unable to adapt to novel concepts without destroying their pre-trained reasoning capabilities. To isolate this phenomenon, we introduce Iterative Semantic Map Refinement (ISMR), a diagnostic protocol revealing that alignment is a fixed geometric barrier that scaling cannot solve; a 20-layer model overcomes this barrier no faster than a 1-layer model. We introduce the Phase-Resonant Intelligent Spectral Model (PRISM). PRISM encodes semantic identity as resonant frequencies in the complex domain (C^d) and replaces quadratic self-attention with linearithmic O(N log N) Gated Harmonic Convolutions. We validate PRISM on the WMT14 translation task. While the Standard Transformer maintains a slight edge in general competence on static benchmarks (23.88 vs 21.40 BLEU), it fails the “Plasticity-Stability” stress test completely. When injected with novel concepts, the Transformer suffers Catastrophic Forgetting, degrading by -10.55 BLEU points while achieving only 60% acquisition. In contrast, PRISM demonstrates Lossless Plasticity, achieving 96% 5-shot acquisition with negligible degradation (-0.84 BLEU). These results suggest that harmonic representations effectively decouple memory from reasoning, offering a structural solution to the plasticity-stability dilemma in real-time knowledge adaptation.
[708] D-CTNet: A Dual-Branch Channel-Temporal Forecasting Network with Frequency-Domain Correction
Shaoxun Wang, Xingjun Zhang, Kun Xia, Qianyang Li, Jiawei Cao, Zhendong Tan
Main category: cs.LG
TL;DR: D-CTNet: A patch-based dual-branch network for multivariate time series forecasting that decouples temporal and channel dependencies while addressing non-stationary distribution shifts through frequency-domain correction.
Details
Motivation: Multivariate time series forecasting is crucial for collaborative industrial systems (Digital Twin, maintenance), but faces challenges: need to decouple complex inter-variable dependencies while handling non-stationary distribution shifts caused by environmental changes in collaborative environments.Method: Proposes Patch-Based Dual-Branch Channel-Temporal Forecasting Network (D-CTNet) with: 1) parallel dual-branch design (linear temporal modeling + channel attention) to decouple and jointly learn intra-channel temporal patterns and multivariate correlations, 2) global patch attention fusion for long-range dependencies, 3) Frequency-Domain Stationarity Correction mechanism for adaptive suppression of distribution shifts via spectrum alignment.
Result: Evaluations on seven benchmark datasets show better forecasting accuracy and robustness compared with state-of-the-art methods.
Conclusion: D-CTNet shows promise as a new forecasting engine for industrial collaborative systems by effectively addressing both dependency decoupling and non-stationary distribution shift challenges.
Abstract: Accurate Multivariate Time Series (MTS) forecasting is crucial for collaborative design of complex systems, Digital Twin building, and maintenance ahead of time. However, the collaborative industrial environment presents new challenges for MTS forecasting models: models should decouple complex inter-variable dependencies while addressing non-stationary distribution shift brought by environmental changes. To address these challenges and improve collaborative sensing reliability, we propose a Patch-Based Dual-Branch Channel-Temporal Forecasting Network (D-CTNet). Particularly, with a parallel dual-branch design incorporating linear temporal modeling layer and channel attention mechanism, our method explicitly decouples and jointly learns intra-channel temporal evolution patterns and dynamic multivariate correlations. Furthermore, a global patch attention fusion module goes beyond the local window scope to model long range dependencies. Most importantly, aiming at non-stationarity, a Frequency-Domain Stationarity Correction mechanism adaptively suppresses distribution shift impacts from environment change by spectrum alignment. Evaluations on seven benchmark datasets show that our model achieves better forecasting accuracy and robustness compared with state-of-the-art methods. Our work shows great promise as a new forecasting engine for industrial collaborative systems.
[709] Memory-Integrated Reconfigurable Adapters: A Unified Framework for Settings with Multiple Tasks
Susmit Agrawal, Krishn Vishwas Kher, Saksham Mittal, Swarnim Maheshwari, Vineeth N. Balasubramanian
Main category: cs.LG
TL;DR: MIRA is a unified framework combining Hopfield-style associative memory with adapter modules to handle both domain generalization and continual learning, achieving state-of-the-art performance by enabling rapid task switching and preventing catastrophic forgetting.
Details
Motivation: The brain seamlessly handles domain shifts and sequential tasks without catastrophic forgetting, using neuromodulatory systems like dopamine and acetylcholine. Current deep learning approaches treat domain generalization and continual learning as separate problems, despite their biological connection. No prior work has explored associative memories (integral to biological systems) to jointly address these challenges.Method: MIRA integrates Hopfield-style associative memory modules atop a shared backbone. It learns associative memory keys post-hoc to index and retrieve affine combinations of stored adapter updates for any given task or domain on a per-sample basis. The framework uses adapter-based modulation with associative memory to enable rapid task switching.
Result: MIRA achieves state-of-the-art out-of-distribution accuracy in domain generalization benchmarks. In incremental learning settings, it outperforms architectures explicitly designed to handle catastrophic forgetting using generic continual learning algorithms. The framework demonstrates enhanced adaptability and retention.
Conclusion: By unifying adapter-based modulation with biologically inspired associative memory, MIRA delivers rapid task switching and enduring knowledge retention in a single extensible architecture. This charts a path toward more versatile and memory-augmented AI systems that better emulate biological learning capabilities.
Abstract: Organisms constantly pivot between tasks such as evading predators, foraging, traversing rugged terrain, and socializing, often within milliseconds. Remarkably, they preserve knowledge of once-learned environments sans catastrophic forgetting, a phenomenon neuroscientists hypothesize, is due to a singular neural circuitry dynamically overlayed by neuromodulatory agents such as dopamine and acetylcholine. In parallel, deep learning research addresses analogous challenges via domain generalization (DG) and continual learning (CL), yet these methods remain siloed, despite the brains ability to perform them seamlessly. In particular, prior work has not explored architectures involving associative memories (AMs), which are an integral part of biological systems, to jointly address these tasks. We propose Memory-Integrated Reconfigurable Adapters (MIRA), a unified framework that integrates Hopfield-style associative memory modules atop a shared backbone. Associative memory keys are learned post-hoc to index and retrieve an affine combination of stored adapter updates for any given task or domain on a per-sample basis. By varying only the task-specific objectives, we demonstrate that MIRA seamlessly accommodates domain shifts and sequential task exposures under one roof. Empirical evaluations on standard benchmarks confirm that our AM-augmented architecture significantly enhances adaptability and retention: in DG, MIRA achieves SoTA out-of-distribution accuracy, and in incremental learning settings, it outperforms architectures explicitly designed to handle catastrophic forgetting using generic CL algorithms. By unifying adapter-based modulation with biologically inspired associative memory, MIRA delivers rapid task switching and enduring knowledge retention in a single extensible architecture, charting a path toward more versatile and memory-augmented AI systems.
[710] Multi-Modal AI for Remote Patient Monitoring in Cancer Care
Yansong Liu, Ronnie Stafford, Pramit Khetrapal, Huriye Kocadag, Graça Carvalho, Patricia de Winter, Maryam Imran, Amelia Snook, Adamos Hadjivasiliou, D. Vijay Anand, Weining Lin, John Kelly, Yukun Zhou, Ivana Drobnjak
Main category: cs.LG
TL;DR: Multi-modal AI framework for remote cancer patient monitoring using wearable sensors, surveys, and clinical data to predict adverse events with 83.9% accuracy.
Details
Motivation: Cancer patients face uncertainties and unmonitored side effects between clinic visits, creating a critical care gap that needs bridging through continuous remote monitoring.Method: Developed multi-modal AI framework integrating HALO-X platform data (demographics, wearable sensors, daily surveys, clinical events) with models adapted for asynchronous, incomplete real-world RPM data to forecast continuous risk of adverse events.
Result: Model achieved 83.9% accuracy (AUROC=0.70) on 2.1M data points from 84 patients over 6,080 patient-days; identified previous treatments, wellness check-ins, and daily max heart rate as key predictive features; case study showed early warning capability.
Conclusion: Establishes feasibility of multi-modal AI RPM for cancer care and offers path toward proactive patient support, demonstrating potential to bridge care gaps between clinical visits.
Abstract: For patients undergoing systemic cancer therapy, the time between clinic visits is full of uncertainties and risks of unmonitored side effects. To bridge this gap in care, we developed and prospectively trialed a multi-modal AI framework for remote patient monitoring (RPM). This system integrates multi-modal data from the HALO-X platform, such as demographics, wearable sensors, daily surveys, and clinical events. Our observational trial is one of the largest of its kind and has collected over 2.1 million data points (6,080 patient-days) of monitoring from 84 patients. We developed and adapted a multi-modal AI model to handle the asynchronous and incomplete nature of real-world RPM data, forecasting a continuous risk of future adverse events. The model achieved an accuracy of 83.9% (AUROC=0.70). Notably, the model identified previous treatments, wellness check-ins, and daily maximum heart rate as key predictive features. A case study demonstrated the model’s ability to provide early warnings by outputting escalating risk profiles prior to the event. This work establishes the feasibility of multi-modal AI RPM for cancer care and offers a path toward more proactive patient support.(Accepted at Europe NeurIPS 2025 Multimodal Representation Learning for Healthcare Workshop)
[711] WUSH: Near-Optimal Adaptive Transforms for LLM Quantization
Jiale Chen, Vage Egiazarian, Torsten Hoefler, Dan Alistarh
Main category: cs.LG
TL;DR: WUSH introduces optimal data-aware linear transforms for joint weight-activation quantization that combine Hadamard backbone with data-dependent components, outperforming fixed orthogonal transforms.
Details
Motivation: Fixed orthogonal transforms (like Hadamard) ignore data statistics and their optimality is not understood, while extreme weights/activations in LLMs reduce quantization resolution by stretching dynamic range.Method: Derives closed-form optimal linear blockwise transforms for joint weight-activation quantization using data-free quantizers. Combines Hadamard backbone with data-dependent components based on second-order moments, creating non-orthogonal but structured transforms.
Result: WUSH transform is provably optimal under mild assumptions, remains structured for efficient implementation, and consistently improves upon Hadamard transform for common numerical formats.
Conclusion: Data-aware transforms outperform fixed orthogonal transforms for quantization, with WUSH providing optimal structured solution for improving quantization resolution in LLMs.
Abstract: Quantization to low bitwidth is a standard approach for deploying large language models, however, a few extreme weights and activations stretch the dynamic range and reduce the effective resolution of the quantizer. A common mitigation approach is to apply some fixed orthogonal transforms, such as Hadamard matrices, before quantization, which typically reduces the dynamic range. Yet, these transforms ignore the statistics of the data, and their optimality is currently not understood. In this work, we derive, for the first time, closed-form optimal linear blockwise transforms for joint weight-activation quantization using standard data-free quantizers for common numerical formats. Specifically, we provide derivations of the optimal adaptive (data-aware) transforms for round-to-nearest (RTN), AbsMax-scaled block quantizers for both integer and floating-point formats. The resulting construction, which we call WUSH, combines a Hadamard backbone with a data-dependent component based on second-order moments, yielding a non-orthogonal transform that is provably optimal under mild assumptions and remains structured for efficient implementation. Preliminary experimental results show that our approach consistently improves upon the Hadamard transform for common formats.
[712] Goal-Driven Reward by Video Diffusion Models for Reinforcement Learning
Qi Wang, Mian Wu, Yuyang Zhang, Mingqi Yuan, Wenyao Zhang, Haoxiang You, Yunbo Wang, Xin Jin, Xiaokang Yang, Wenjun Zeng
Main category: cs.LG
TL;DR: Using pretrained video diffusion models as reward functions for RL agents, eliminating need for hand-designed rewards by leveraging video-level alignment and frame-level goal states.
Details
Motivation: RL often requires carefully designed reward functions that are challenging to create and don't generalize well across tasks. The paper aims to leverage the rich world knowledge in pretrained video diffusion models to provide goal-driven reward signals without ad-hoc reward design.Method: 1) Use off-the-shelf video diffusion models pretrained on large-scale video datasets as reward functions. 2) For video-level rewards: finetune pretrained video diffusion model on domain-specific datasets, then use its video encoder to evaluate alignment between agent trajectories and generated goal videos. 3) For frame-level goals: identify most relevant frame from generated video using CLIP as goal state, then use learned forward-backward representation to compute probability of visiting goal state from given state-action pair as frame-level reward.
Result: Experiments on various Meta-World tasks demonstrate the effectiveness of the approach.
Conclusion: The method successfully leverages pretrained video diffusion models to provide informative reward signals for RL agents without requiring hand-designed reward functions, enabling more coherent and goal-driven trajectories.
Abstract: Reinforcement Learning (RL) has achieved remarkable success in various domains, yet it often relies on carefully designed programmatic reward functions to guide agent behavior. Designing such reward functions can be challenging and may not generalize well across different tasks. To address this limitation, we leverage the rich world knowledge contained in pretrained video diffusion models to provide goal-driven reward signals for RL agents without ad-hoc design of reward. Our key idea is to exploit off-the-shelf video diffusion models pretrained on large-scale video datasets as informative reward functions in terms of video-level and frame-level goals. For video-level rewards, we first finetune a pretrained video diffusion model on domain-specific datasets and then employ its video encoder to evaluate the alignment between the latent representations of agent’s trajectories and the generated goal videos. To enable more fine-grained goal-achievement, we derive a frame-level goal by identifying the most relevant frame from the generated video using CLIP, which serves as the goal state. We then employ a learned forward-backward representation that represents the probability of visiting the goal state from a given state-action pair as frame-level reward, promoting more coherent and goal-driven trajectories. Experiments on various Meta-World tasks demonstrate the effectiveness of our approach.
[713] Stabilizing Reinforcement Learning with LLMs: Formulation and Practices
Chujie Zheng, Kai Dang, Bowen Yu, Mingze Li, Huiqiang Jiang, Junrong Lin, Yuqiong Liu, An Yang, Jingren Zhou, Junyang Lin
Main category: cs.LG
TL;DR: This paper explains why token-level surrogate objectives can optimize sequence-level rewards in RL with LLMs, identifies conditions for validity (minimized training-inference discrepancy and policy staleness), and provides stabilization techniques including importance sampling, clipping, and Routing Replay for MoE models.
Details
Motivation: The motivation is to provide a principled understanding of why surrogate token-level objectives work for sequence-level RL optimization with LLMs, and to develop stable training recipes for large-scale RL with MoE models, addressing training instability issues that arise in practice.Method: The method uses a first-order approximation to show when token-level surrogate objectives become valid for sequence-level reward optimization. It identifies two key conditions: minimized training-inference discrepancy and policy staleness. The paper then validates this through extensive experiments with a 30B MoE model, testing various stabilization techniques including importance sampling correction, clipping, and Routing Replay for MoE models.
Result: For on-policy training, basic policy gradient with importance sampling correction achieves highest stability. For off-policy training (to accelerate convergence), combining clipping and Routing Replay is essential to mitigate instability from policy staleness. Once stabilized, prolonged optimization yields comparable final performance regardless of cold-start initialization.
Conclusion: The paper provides principled explanations for RL stabilization techniques and practical recipes for stable RL training with large LLMs, particularly MoE models. The insights should facilitate future research in RL with LLMs by addressing training instability challenges.
Abstract: This paper proposes a novel formulation for reinforcement learning (RL) with large language models, explaining why and under what conditions the true sequence-level reward can be optimized via a surrogate token-level objective in policy gradient methods such as REINFORCE. Specifically, through a first-order approximation, we show that this surrogate becomes increasingly valid only when both the training-inference discrepancy and policy staleness are minimized. This insight provides a principled explanation for the crucial role of several widely adopted techniques in stabilizing RL training, including importance sampling correction, clipping, and particularly Routing Replay for Mixture-of-Experts (MoE) models. Through extensive experiments with a 30B MoE model totaling hundreds of thousands of GPU hours, we show that for on-policy training, the basic policy gradient algorithm with importance sampling correction achieves the highest training stability. When off-policy updates are introduced to accelerate convergence, combining clipping and Routing Replay becomes essential to mitigate the instability caused by policy staleness. Notably, once training is stabilized, prolonged optimization consistently yields comparable final performance regardless of cold-start initialization. We hope that the shared insights and the developed recipes for stable RL training will facilitate future research.
[714] Subgroup Validity in Machine Learning for Echocardiogram Data
Cynthia Feeney, Shane Williams, Benjamin S. Wessler, Michael C. Hughes
Main category: cs.LG
TL;DR: Current open echocardiogram datasets lack adequate sociodemographic reporting and subgroup analysis, raising concerns about subgroup validity for AI models in cardiac ultrasound interpretation.
Details
Motivation: Echocardiogram datasets enable training deep learning models for automated cardiac ultrasound interpretation, but underreported patient demographics and unevaluated subgroup performance raise validity concerns that must be addressed before clinical deployment.Method: The authors improved sociodemographic reporting for TMED-2 and MIMIC-IV-ECHO datasets, analyzed six open echocardiogram datasets for demographic completeness, and performed exploratory subgroup analysis of two published aortic stenosis detection models on TMED-2.
Result: Analysis revealed no consideration of gender-diverse patients, insufficient patient counts for many racial/ethnic groups, and insufficient evidence for subgroup validity across sex, racial, and ethnic subgroups in aortic stenosis detection models.
Conclusion: More data for underrepresented subgroups, improved demographic reporting, and subgroup-focused analyses are needed to prove subgroup validity in echocardiogram AI models before clinical deployment.
Abstract: Echocardiogram datasets enable training deep learning models to automate interpretation of cardiac ultrasound, thereby expanding access to accurate readings of diagnostically-useful images. However, the gender, sex, race, and ethnicity of the patients in these datasets are underreported and subgroup-specific predictive performance is unevaluated. These reporting deficiencies raise concerns about subgroup validity that must be studied and addressed before model deployment. In this paper, we show that current open echocardiogram datasets are unable to assuage subgroup validity concerns. We improve sociodemographic reporting for two datasets: TMED-2 and MIMIC-IV-ECHO. Analysis of six open datasets reveals no consideration of gender-diverse patients and insufficient patient counts for many racial and ethnic groups. We further perform an exploratory subgroup analysis of two published aortic stenosis detection models on TMED-2. We find insufficient evidence for subgroup validity for sex, racial, and ethnic subgroups. Our findings highlight that more data for underrepresented subgroups, improved demographic reporting, and subgroup-focused analyses are needed to prove subgroup validity in future work.
[715] Upper Approximation Bounds for Neural Oscillators
Zifeng Huang, Konstantin M. Zuev, Yong Xia, Michael Beer
Main category: cs.LG
TL;DR: Neural oscillators (2nd-order ODEs + MLP) have theoretical approximation bounds derived for causal operators and stable dynamical systems, showing polynomial error scaling that avoids the curse of dimensionality.
Details
Motivation: Neural oscillators show competitive performance for learning causal mappings in temporal sequences, but lack theoretical quantification of their architectural capacities, which is needed for robust application in science and engineering.Method: Analyze neural oscillator consisting of second-order ODE followed by multilayer perceptron (MLP). Derive upper approximation bounds for: 1) causal/uniformly continuous operators between continuous temporal function spaces, and 2) uniformly asymptotically incrementally stable second-order dynamical systems. Proof method also applies to state-space models with linear time-continuous complex RNN + MLP.
Result: Approximation error for second-order dynamical systems scales polynomially with reciprocals of MLP widths, mitigating curse of parametric complexity. Two numerical cases validate decay rates of established error bounds.
Conclusion: Theoretical foundation established for neural oscillators, showing they can effectively approximate causal operators and stable dynamical systems with polynomial error scaling, supporting their practical application in science and engineering.
Abstract: Neural oscillators, originating from the second-order ordinary differential equations (ODEs), have demonstrated competitive performance in stably learning causal mappings between long-term sequences or continuous temporal functions. However, theoretically quantifying the capacities of their neural network architectures remains a significant challenge. In this study, the neural oscillator consisting of a second-order ODE followed by a multilayer perceptron (MLP) is considered. Its upper approximation bound for approximating causal and uniformly continuous operators between continuous temporal function spaces and that for approximating uniformly asymptotically incrementally stable second-order dynamical systems are derived. The established proof method of the approximation bound for approximating the causal continuous operators can also be directly applied to state-space models consisting of a linear time-continuous complex recurrent neural network followed by an MLP. Theoretical results reveal that the approximation error of the neural oscillator for approximating the second-order dynamical systems scales polynomially with the reciprocals of the widths of two utilized MLPs, thus mitigating the curse of parametric complexity. The decay rates of two established approximation error bounds are validated through two numerical cases. These results provide a robust theoretical foundation for the effective application of the neural oscillator in science and engineering.
[716] ZIP-RC: Zero-overhead Inference-time Prediction of Reward and Cost for Adaptive and Interpretable Generation
Rohin Manvi, Joey Hong, Tim Seyde, Maxime Labonne, Mathias Lechner, Sergey Levine
Main category: cs.LG
TL;DR: ZIP-RC enables LLMs to predict reward and remaining length during inference with zero overhead, allowing adaptive sampling decisions based on expected utility of continuing vs. stopping.
Details
Motivation: LLMs lack introspection for anticipating success/computation needs, leading to inefficient fixed-budget sampling (like Best-of-N) and no confidence signals. This causes high costs, latency, and undermines trust.Method: ZIP-RC reuses unused logits in forward passes to output joint distribution over final reward and remaining length. Computes sampling utility (expected max reward, compute, latency) and uses meta-actions to decide whether to continue/stop sampling.
Result: On math benchmarks, improves accuracy by up to 12% over majority voting at equal/lower average cost, creates smooth Pareto frontiers between quality, compute, and latency.
Conclusion: ZIP-RC provides real-time reward-cost introspection enabling adaptive, efficient reasoning without extra models or inference overhead.
Abstract: Large language models excel at reasoning but lack key aspects of introspection, including anticipating their own success and the computation required to achieve it. Humans use real-time introspection to decide how much effort to invest, when to make multiple attempts, when to stop, and when to signal success or failure. Without this, LLMs struggle to make intelligent meta-cognition decisions. Test-time scaling methods like Best-of-N drive up cost and latency by using a fixed budget of samples regardless of the marginal benefit of each one at any point in generation, and the absence of confidence signals can mislead people, prevent appropriate escalation to better tools, and undermine trustworthiness. Learned verifiers or reward models can provide confidence estimates, but do not enable adaptive inference and add substantial cost by requiring extra models or forward passes. We present ZIP-RC, an adaptive inference method that equips models with zero-overhead inference-time predictions of reward and cost. At every token, ZIP-RC reuses reserved or unused logits in the same forward pass as next-token prediction to output a joint distribution over final reward and remaining length – no extra models, architecture change, or inference overhead. This full joint distribution is used to compute a sampling utility which is the linear combination of the expected maximum reward, total compute, and latency of set of samples if generated to completion. During inference, we maximize this utility with meta-actions that determine which prefix of tokens to continue or initiate sampling from. On mixed-difficulty mathematical benchmarks, ZIP-RC improves accuracy by up to 12% over majority voting at equal or lower average cost, and traces smooth Pareto frontiers between quality, compute, and latency. By providing real-time reward-cost introspection, ZIP-RC enables adaptive, efficient reasoning.
[717] Operator-Theoretic Framework for Gradient-Free Federated Learning
Mohit Kumar, Mathias Brucker, Alexander Valentinitsch, Adnan Husakovic, Ali Abbas, Manuela Geiß, Bernhard A. Moser
Main category: cs.LG
TL;DR: Operator-theoretic federated learning framework using RKHS mapping for gradient-free optimization with privacy, communication efficiency, and FHE compatibility.
Details
Motivation: Address federated learning challenges: heterogeneity, strict communication/computation limits, privacy requirements, while ensuring performance comparable to gradient-based methods.Method: Map L²-optimal solution to RKHS via forward operator, approximate with data, map back with inverse operator. Use Kernel Affine Hull Machines with space folding property, clients transfer scalar space folding measures. Private protocol uses noise-perturbed data matrices in one step.
Result: Gradient-free FL matches/exceeds gradient-based fine-tuning (gains up to 23.7 points). Differential privacy with kernel smoothing mitigates accuracy loss. FHE-compatible global rule requires only integer min/equality operations per test point with practical latencies.
Conclusion: Framework provides provable guarantees with low communication, supports private knowledge transfer via scalar summaries, yields FHE-compatible prediction rule as mathematically grounded alternative to gradient-based FL under heterogeneity.
Abstract: Federated learning must address heterogeneity, strict communication and computation limits, and privacy while ensuring performance. We propose an operator-theoretic framework that maps the $L^2$-optimal solution into a reproducing kernel Hilbert space (RKHS) via a forward operator, approximates it using available data, and maps back with the inverse operator, yielding a gradient-free scheme. Finite-sample bounds are derived using concentration inequalities over operator norms, and the framework identifies a data-dependent hypothesis space with guarantees on risk, error, robustness, and approximation. Within this space we design efficient kernel machines leveraging the space folding property of Kernel Affine Hull Machines. Clients transfer knowledge via a scalar space folding measure, reducing communication and enabling a simple differentially private protocol: summaries are computed from noise-perturbed data matrices in one step, avoiding per-round clipping and privacy accounting. The induced global rule requires only integer minimum and equality-comparison operations per test point, making it compatible with fully homomorphic encryption (FHE). Across four benchmarks, the gradient-free FL method with fixed encoder embeddings matches or outperforms strong gradient-based fine-tuning, with gains up to 23.7 points. In differentially private experiments, kernel smoothing mitigates accuracy loss in high-privacy regimes. The global rule admits an FHE realization using $Q \times C$ encrypted minimum and $C$ equality-comparison operations per test point, with operation-level benchmarks showing practical latencies. Overall, the framework provides provable guarantees with low communication, supports private knowledge transfer via scalar summaries, and yields an FHE-compatible prediction rule offering a mathematically grounded alternative to gradient-based federated learning under heterogeneity.
[718] AltNet: Addressing the Plasticity-Stability Dilemma in Reinforcement Learning
Mansi Maheshwari, John C. Raisbeck, Bruno Castro da Silva
Main category: cs.LG
TL;DR: AltNet uses twin networks that alternate roles to restore plasticity without performance drops in reinforcement learning.
Details
Motivation: Neural networks in RL suffer from plasticity loss over time, and existing reset methods cause dangerous performance drops in real-world settings.Method: Twin networks alternate roles: one learns actively in environment, the other learns off-policy from interactions and replay buffer. Active network periodically resets while passive network becomes new active network.
Result: AltNet restores plasticity, improves sample efficiency, achieves higher performance, and avoids performance drops in DeepMind Control Suite tasks.
Conclusion: AltNet outperforms baseline methods and state-of-the-art reset techniques while maintaining safety by eliminating performance degradation during resets.
Abstract: Neural networks have shown remarkable success in supervised learning when trained on a single task using a fixed dataset. However, when neural networks are trained on a reinforcement learning task, their ability to continue learning from new experiences declines over time. This decline in learning ability is known as plasticity loss. To restore plasticity, prior work has explored periodically resetting the parameters of the learning network, a strategy that often improves overall performance. However, such resets come at the cost of a temporary drop in performance, which can be dangerous in real-world settings. To overcome this instability, we introduce AltNet, a reset-based approach that restores plasticity without performance degradation by leveraging twin networks. The use of twin networks anchors performance during resets through a mechanism that allows networks to periodically alternate roles: one network learns as it acts in the environment, while the other learns off-policy from the active network’s interactions and a replay buffer. At fixed intervals, the active network is reset and the passive network, having learned from prior experiences, becomes the new active network. AltNet restores plasticity, improving sample efficiency and achieving higher performance, while avoiding performance drops that pose risks in safety-critical settings. We demonstrate these advantages in several high-dimensional control tasks from the DeepMind Control Suite, where AltNet outperforms various relevant baseline methods, as well as state-of-the-art reset-based techniques.
[719] HalluGraph: Auditable Hallucination Detection for Legal RAG Systems via Knowledge Graph Alignment
Valentin Noël, Elimane Yassine Seidou, Charly Ken Capo-Chichi, Ghanem Amari
Main category: cs.LG
TL;DR: HalluGraph is a graph-theoretic framework for detecting hallucinations in legal AI systems by structurally aligning knowledge graphs from context, query, and response to provide verifiable guarantees about source document fidelity.
Details
Motivation: Legal AI systems need accountability when citing legal documents, as existing hallucination detectors using semantic similarity can tolerate dangerous entity substitutions (parties, dates, legal provisions) with material consequences.Method: Graph-theoretic framework that quantifies hallucinations through structural alignment between knowledge graphs extracted from context, query, and response. Uses two metrics: Entity Grounding (EG) measures if response entities appear in source documents, and Relation Preservation (RP) verifies if asserted relationships are supported by context.
Result: On structured control documents (>400 words, >20 entities), HalluGraph achieves near-perfect discrimination with AUC = 0.979. Maintains robust performance (AUC ≈ 0.89) on challenging generative legal tasks, consistently outperforming semantic similarity baselines.
Conclusion: HalluGraph provides the transparency and traceability required for high-stakes legal applications, enabling full audit trails from generated assertions back to source passages, addressing critical accountability challenges in legal AI systems.
Abstract: Legal AI systems powered by retrieval-augmented generation (RAG) face a critical accountability challenge: when an AI assistant cites case law, statutes, or contractual clauses, practitioners need verifiable guarantees that generated text faithfully represents source documents. Existing hallucination detectors rely on semantic similarity metrics that tolerate entity substitutions, a dangerous failure mode when confusing parties, dates, or legal provisions can have material consequences. We introduce HalluGraph, a graph-theoretic framework that quantifies hallucinations through structural alignment between knowledge graphs extracted from context, query, and response. Our approach produces bounded, interpretable metrics decomposed into \textit{Entity Grounding} (EG), measuring whether entities in the response appear in source documents, and \textit{Relation Preservation} (RP), verifying that asserted relationships are supported by context. On structured control documents, HalluGraph achieves near-perfect discrimination ($>$400 words, $>$20 entities), HalluGraph achieves $AUC = 0.979$, while maintaining robust performance ($AUC \approx 0.89$) on challenging generative legal task, consistently outperforming semantic similarity baselines. The framework provides the transparency and traceability required for high-stakes legal applications, enabling full audit trails from generated assertions back to source passages.
[720] FMTK: A Modular Toolkit for Composable Time Series Foundation Model Pipelines
Hetvi Shastri, Pragya Sharma, Walid A. Hanafy, Mani Srivastava, Prashant Shenoy
Main category: cs.LG
TL;DR: FMTK is an open-source toolkit for building and fine-tuning time-series foundation model pipelines with standardized abstractions, requiring only ~7 lines of code.
Details
Motivation: Current time-series foundation model pipelines require ad hoc, model-specific implementations that lack modularity and reproducibility, creating barriers to flexible composition across models and tasks.Method: FMTK introduces standardized backbone and component abstractions that enable flexible composition of TSFM pipelines. It provides a lightweight, extensible toolkit for constructing and fine-tuning these pipelines.
Result: FMTK achieves correctness and performance with an average of just seven lines of code, enabling modular and reproducible TSFM pipeline construction.
Conclusion: FMTK addresses the modularity and reproducibility challenges in time-series foundation model pipelines through standardized abstractions, making TSFM application development more accessible and efficient.
Abstract: Foundation models (FMs) have opened new avenues for machine learning applications due to their ability to adapt to new and unseen tasks with minimal or no further training. Time-series foundation models (TSFMs) – FMs trained on time-series data – have shown strong performance on classification, regression, and imputation tasks. Recent pipelines combine TSFMs with task-specific encoders, decoders, and adapters to improve performance; however, assembling such pipelines typically requires ad hoc, model-specific implementations that hinder modularity and reproducibility. We introduce FMTK, an open-source, lightweight and extensible toolkit for constructing and fine-tuning TSFM pipelines via standardized backbone and component abstractions. FMTK enables flexible composition across models and tasks, achieving correctness and performance with an average of seven lines of code. https://github.com/umassos/FMTK
[721] Adaptive-lambda Subtracted Importance Sampled Scores in Machine Unlearning for DDPMs and VAEs
MohammadParsa Dini, Human Jafari, Sajjad Amini, MohammadMahdi Mojahedian
Main category: cs.LG
TL;DR: Adaptive-lambda SISS improves machine unlearning for diffusion models by dynamically inferring optimal mixing weights instead of using fixed ones, achieving better trade-offs between forgetting and retention.
Details
Motivation: Existing machine unlearning methods for generative models use fixed mixing weights (lambda) which are suboptimal because unlearning strength needs to vary across samples and training stages. This limits effectiveness in complying with right-to-be-forgotten requirements while preserving model performance.Method: Proposes Adaptive-lambda SISS where lambda becomes a latent variable dynamically inferred at each training step using a lightweight inference network. The network parameterizes an adaptive posterior over lambda conditioned on contextual features from SISS loss terms. Also extends to score-based unlearning, introduces hybrid objectives combining Score Forgetting Distillation with SISS, and proposes RL formulation treating unlearning as sequential decision process.
Result: Experiments on augmented MNIST benchmark show Adaptive-lambda SISS substantially outperforms original static-lambda SISS, achieving stronger removal of forgotten classes while better preserving generation quality on the retain set.
Conclusion: Adaptive-lambda approach provides a principled extension to machine unlearning that dynamically adjusts unlearning strength, enabling better trade-offs between forgetting undesired content and preserving model performance on retained data.
Abstract: Machine Unlearning is essential for large generative models (VAEs, DDPMs) to comply with the right to be forgotten and prevent undesired content generation without costly retraining. Existing approaches, such as Static-lambda SISS for diffusion models, rely on a fixed mixing weight lambda, which is suboptimal because the required unlearning strength varies across samples and training stages. We propose Adaptive-lambda SISS, a principled extension that turns lambda into a latent variable dynamically inferred at each training step. A lightweight inference network parameterizes an adaptive posterior over lambda, conditioned on contextual features derived from the instantaneous SISS loss terms (retain/forget losses and their gradients). This enables joint optimization of the diffusion model and the lambda-inference mechanism via a variational objective, yielding significantly better trade-offs. We further extend the adaptive-lambda principle to score-based unlearning and introduce a multi-class variant of Score Forgetting Distillation. In addition, we present two new directions: (i) a hybrid objective combining the data-free efficiency of Score Forgetting Distillation with the direct gradient control of SISS, and (ii) a Reinforcement Learning formulation that treats unlearning as a sequential decision process, learning an optimal policy over a state space defined by the model’s current memory of the forget set. Experiments on an augmented MNIST benchmark show that Adaptive-lambda SISS substantially outperforms the original static-lambda SISS, achieving stronger removal of forgotten classes while better preserving generation quality on the retain set.
[722] Agentic Policy Optimization via Instruction-Policy Co-Evolution
Han Zhou, Xingchen Wan, Ivan Vulić, Anna Korhonen
Main category: cs.LG
TL;DR: INSPO introduces an instruction-policy co-evolution framework that dynamically optimizes LLM agent instructions during reinforcement learning, outperforming static instruction approaches.
Details
Motivation: Current RLVR methods use static, manually designed instructions that may be suboptimal for the base model and don't adapt as the agent's policy improves through environment interaction.Method: INSPO maintains a dynamic population of instruction candidates, samples them with questions, attributes RL reward signals to instructions, prunes low performers, and generates new instructions through on-policy reflection using LLM-based optimization of past experience.
Result: INSPO substantially outperforms strong baselines with static instructions on multi-turn retrieval and reasoning tasks, discovering innovative instructions that guide agents toward more strategic reasoning paths with minimal computational overhead.
Conclusion: Dynamic instruction optimization integrated into the RL loop enables more effective agent reasoning, demonstrating that instruction-policy co-evolution is a promising approach for advancing LLM agent capabilities.
Abstract: Reinforcement Learning with Verifiable Rewards (RLVR) has advanced the reasoning capability of large language models (LLMs), enabling autonomous agents that can conduct effective multi-turn and tool-integrated reasoning. While instructions serve as the primary protocol for defining agents, RLVR typically relies on static and manually designed instructions. However, those instructions may be suboptimal for the base model, and the optimal instruction may change as the agent’s policy improves and explores the interaction with the environment. To bridge the gap, we introduce INSPO, a novel Instruction-Policy co-evolution framework that integrates instruction optimization as a dynamic component of the reinforcement learning (RL) loop. INSPO maintains a dynamic population of instruction candidates that are sampled with questions, where reward signals in RL loops are automatically attributed to each instruction, and low performers are periodically pruned. New instructions are generated and verified through an on-policy reflection mechanism, where an LLM-based optimizer analyzes past experience from a replay buffer and evolves more effective strategies given the current policy. We conduct extensive experiments on multi-turn retrieval and reasoning tasks, demonstrating that INSPO substantially outperforms strong baselines relying on static instructions. INSPO discovers innovative instructions that guide the agent toward more strategic reasoning paths, achieving substantial performance gains with only a marginal increase in computational overhead.
[723] PIANO: Physics-informed Dual Neural Operator for Precipitation Nowcasting
Seokhyun Chin, Junghwan Park, Woojin Cho
Main category: cs.LG
TL;DR: Physics-informed neural operator (PIANO) combines satellite imagery with physics constraints for improved precipitation nowcasting, showing better accuracy for moderate and heavy precipitation events with low seasonal variability.
Details
Motivation: Current precipitation nowcasting methods are computationally expensive and restrictive, limiting access for many countries. There's a need for more accessible and accurate forecasting methods for disaster early warning.Method: Proposes PIANO (physics-informed dual neural operator) that enforces advection-diffusion equation during training using PINN loss to predict satellite imagery, then uses a generative model to convert satellite images to radar images for precipitation nowcasting.
Result: Notable improvement in moderate (4mm/h) and heavy (8mm/h) precipitation event prediction compared to baselines. Demonstrates low seasonal variability in predictions, indicating robustness for generalization.
Conclusion: PIANO shows potential as a physics-informed approach for precipitation nowcasting and serves as a good baseline for future work in this direction.
Abstract: Precipitation nowcasting, key for early warning of disasters, currently relies on computationally expensive and restrictive methods that limit access to many countries. To overcome this challenge, we propose precipitation nowcasting using satellite imagery with physics constraints for improved accuracy and physical consistency. We use a novel physics-informed dual neural operator (PIANO) structure to enforce the fundamental equation of advection-diffusion during training to predict satellite imagery using a PINN loss. Then, we use a generative model to convert satellite images to radar images, which are used for precipitation nowcasting. Compared to baseline models, our proposed model shows a notable improvement in moderate (4mm/h) precipitation event prediction alongside short-term heavy (8mm/h) precipitation event prediction. It also demonstrates low seasonal variability in predictions, indicating robustness for generalization. This study suggests the potential of the PIANO and serves as a good baseline for physics-informed precipitation nowcasting.
[724] AlignSAE: Concept-Aligned Sparse Autoencoders
Minglai Yang, Xinyu Guo, Mihai Surdeanu, Liangming Pan
Main category: cs.LG
TL;DR: AlignSAE improves interpretability of LLMs by aligning sparse autoencoder features with human-defined concepts through a two-stage curriculum, enabling precise concept control.
Details
Motivation: LLMs encode knowledge in opaque parametric spaces that are hard to inspect or control. Current sparse autoencoders produce entangled features that don't reliably align with human concepts, limiting interpretability and controllability.Method: AlignSAE uses a “pre-train, then post-train” curriculum: 1) initial unsupervised training of sparse autoencoder, 2) supervised post-training to bind specific concepts to dedicated latent slots while preserving general reconstruction capacity.
Result: AlignSAE creates an interpretable interface where specific relations can be inspected and controlled without interference. It enables precise causal interventions like reliable “concept swaps” by targeting single, semantically aligned slots.
Conclusion: The method successfully aligns SAE features with human-defined ontologies, providing better interpretability and control over LLM representations while maintaining reconstruction quality.
Abstract: Large Language Models (LLMs) encode factual knowledge within hidden parametric spaces that are difficult to inspect or control. While Sparse Autoencoders (SAEs) can decompose hidden activations into more fine-grained, interpretable features, they often struggle to reliably align these features with human-defined concepts, resulting in entangled and distributed feature representations. To address this, we introduce AlignSAE, a method that aligns SAE features with a defined ontology through a “pre-train, then post-train” curriculum. After an initial unsupervised training phase, we apply supervised post-training to bind specific concepts to dedicated latent slots while preserving the remaining capacity for general reconstruction. This separation creates an interpretable interface where specific relations can be inspected and controlled without interference from unrelated features. Empirical results demonstrate that AlignSAE enables precise causal interventions, such as reliable “concept swaps”, by targeting single, semantically aligned slots.
[725] Bayesian dynamic scheduling of multipurpose batch processes under incomplete look-ahead information
Taicheng Zheng, Dan Li, Jie Li
Main category: cs.LG
TL;DR: Bayesian dynamic scheduling method for multipurpose batch processes that uses Bayesian Networks to quantify disturbance impacts and guide rescheduling, outperforming traditional periodic rescheduling in cost and system nervousness.
Details
Motivation: Multipurpose batch processes operate in dynamic environments with disturbances (processing delays, demand changes), but existing scheduling methods assume complete look-ahead information which is unrealistic. This leads to suboptimal long-term costs and high system nervousness when only incomplete information is available.Method: Proposes a Bayesian dynamic scheduling method that learns a Bayesian Network from the probability distribution of disturbances. The network represents how likely each operation will be impacted by disturbances. During online execution, when new disturbances are observed, the method updates the posterior distribution to guide rescheduling strategies.
Result: The method was compared with existing periodic rescheduling strategy (which generates new schedules from scratch at fixed intervals) on four benchmark problems. Computational results show statistically better long-term costs and system nervousness. Theoretically proved that if disturbances are mutually independent, the impact-quantifying variables satisfy Bayesian Network independence assumptions.
Conclusion: The Bayesian dynamic scheduling method effectively addresses disturbances in multipurpose batch processes with incomplete information. The approach can be extended to other scheduling problems (job shop scheduling, continuous processes) by defining problem-specific dependencies between operations.
Abstract: Multipurpose batch processes become increasingly popular in manufacturing industries since they adapt to low-volume, high-value products and shifting demands. These processes often operate in a dynamic environment, which faces disturbances such as processing delays and demand changes. To minimise long-term cost and system nervousness (i.e., disruptive changes to schedules), schedulers must design rescheduling strategies to address such disturbances effectively. Existing methods often assume complete look-ahead information over the scheduling horizon. This assumption contrasts with realistic situations where schedulers can only access incomplete look-ahead information. Sticking with existing methods may lead to suboptimal long-term costs and high-level system nervousness. In this work we propose a Bayesian dynamic scheduling method. Our method relies on learning a Bayesian Network from the probability distribution of disturbances. Specifically, the Bayesian Network represents how likely each operation will be impacted by disturbances. During the online execution, when new disturbances become observed, this method updates the posterior distribution and therefore guides the rescheduling strategy. We compare our method with the existing periodic rescheduling strategy (which generates new schedules from scratch at fixed intervals) on four benchmark problems. Computational results show that our method achieves statistically better long-term costs and system nervousness. In the theoretical aspect, we prove that if disturbances are mutually independent, the impact-quantifying variables inherently satisfy the independence assumptions required by Bayesian Networks. As an implication, practitioners can extend the method to other scheduling problems (such as job shop scheduling and continuous processes), given that they define the problem-specific dependencies between operations.
[726] Efficiently Learning Branching Networks for Multitask Algorithmic Reasoning
Dongyue Li, Zhenshuo Zhang, Minxuan Duan, Edgar Dobriban, Hongyang R. Zhang
Main category: cs.LG
TL;DR: AutoBRANE: Efficient algorithm for multitask algorithmic reasoning that learns optimal branching neural network structures to reduce negative interference between tasks with different execution steps.
Details
Motivation: Multitask algorithmic reasoning is challenging due to negative interference when training different algorithms together, as their execution steps differ. Current approaches struggle to design a single model that performs well on multiple algorithmic tasks simultaneously.Method: Propose branching neural networks with AutoBRANE algorithm that efficiently searches for optimal k-ary tree structures over n tasks. Uses convex relaxation at each layer to approximate optimal task partition, clustering tasks using gradient-based affinity scores. Works with any base model (GNNs, LLMs).
Result: Gradient features estimate task performance within 5% error across GNNs and LLMs. On CLRS benchmark: outperforms strongest multitask GNN by 3.7%, reduces runtime 48%, memory 26%. On text-based graph reasoning: improves 3.2% over best baseline. On large graph dataset (21M edges, 500 tasks): 28% accuracy gain, 4.5× runtime reduction.
Conclusion: AutoBRANE enables efficient multitask algorithmic reasoning by learning hierarchical branching structures that cluster related algorithms, reducing negative interference while improving performance and efficiency across diverse reasoning tasks.
Abstract: Algorithmic reasoning – the ability to perform step-by-step logical inference – has become a core benchmark for evaluating reasoning in graph neural networks (GNNs) and large language models (LLMs). Ideally, one would like to design a single model capable of performing well on multiple algorithmic reasoning tasks simultaneously. However, this is challenging when the execution steps of algorithms differ from one another, causing negative interference when they are trained together. We propose branching neural networks, a principled architecture for multitask algorithmic reasoning. Searching for the optimal $k$-ary tree with $L$ layers over $n$ algorithmic tasks is combinatorial, requiring exploration of up to $k^{nL}$ possible structures. We develop AutoBRANE, an efficient algorithm that reduces this search to $O(nL)$ time by solving a convex relaxation at each layer to approximate an optimal task partition. The method clusters tasks using gradient-based affinity scores and can be used on top of any base model, including GNNs and LLMs. We validate AutoBRANE on a broad suite of graph-algorithmic and text-based reasoning benchmarks. We show that gradient features estimate true task performance within 5% error across four GNNs and four LLMs (up to 34B parameters). On the CLRS benchmark, it outperforms the strongest single multitask GNN by 3.7% and the best baseline by 1.2%, while reducing runtime by 48% and memory usage by 26%. The learned branching structures reveal an intuitively reasonable hierarchical clustering of related algorithms. On three text-based graph reasoning benchmarks, AutoBRANE improves over the best non-branching multitask baseline by 3.2%. Finally, on a large graph dataset with 21M edges and 500 tasks, AutoBRANE achieves a 28% accuracy gain over existing multitask and branching architectures, along with a 4.5$\times$ reduction in runtime.
[727] World Model Robustness via Surprise Recognition
Geigh Zollicoffer, Tanush Chopra, Mingkuan Yan, Xiaoxu Ma, Kenneth Eaton, Mark Riedl
Main category: cs.LG
TL;DR: World model-based RL agents use surprise-based rejection sampling to maintain performance under sensor noise and OOD distractions.
Details
Motivation: Real-world AI systems face unpredictable distractions and OOD noise that can destabilize policies and cause unsafe behavior, while robust training alone cannot anticipate all possible OOD conditions.Method: Developed an algorithm leveraging world model’s surprise measure to reduce noise impact. Introduced multi-representation and single-representation rejection sampling for robustness to multiple or single faulty sensors.
Result: Methods preserved performance relative to baselines under varying noise types/levels across CARLA and Safety Gymnasium environments. Enhanced stability of two SOTA world models (Cosmos and DreamerV3) with different architectures.
Conclusion: Surprise-based rejection sampling provides robust approach across world modeling domains for handling sensor noise and OOD distractions in real-world deployment.
Abstract: AI systems deployed in the real world must contend with distractions and out-of-distribution (OOD) noise that can destabilize their policies and lead to unsafe behavior. While robust training can reduce sensitivity to some forms of noise, it is infeasible to anticipate all possible OOD conditions. To mitigate this issue, we develop an algorithm that leverages a world model’s inherent measure of surprise to reduce the impact of noise in world model–based reinforcement learning agents. We introduce both multi-representation and single-representation rejection sampling, enabling robustness to settings with multiple faulty sensors or a single faulty sensor. While the introduction of noise typically degrades agent performance, we show that our techniques preserve performance relative to baselines under varying types and levels of noise across multiple environments within self-driving simulation domains (CARLA and Safety Gymnasium). Furthermore, we demonstrate that our methods enhance the stability of two state-of-the-art world models with markedly different underlying architectures: Cosmos and DreamerV3. Together, these results highlight the robustness of our approach across world modeling domains. We release our code at https://github.com/Bluefin-Tuna/WISER .
[728] Projection-Free CNN Pruning via Frank-Wolfe with Momentum: Sparser Models with Less Pretraining
Hamza ElMokhtar Shili, Natasha Patnaik, Isabelle Ruble, Kathryn Jarjoura, Daniel Suarez Aguirre
Main category: cs.LG
TL;DR: FW with momentum pruning yields sparser, more accurate CNNs than magnitude pruning, achieving good results with minimal dense pre-training.
Details
Motivation: Motivated by the Lottery Ticket Hypothesis, which suggests smaller sub-networks within pre-trained networks can perform well, the paper investigates pruning methods specifically for CNNs in image classification tasks.Method: Compares three pruning approaches: simple magnitude-based pruning, Frank-Wolfe style pruning, and FW with momentum on a CNN trained on MNIST. Experiments track test accuracy, loss, sparsity, and inference time while varying dense pre-training from 1 to 10 epochs.
Result: FW with momentum produces pruned networks that are both sparser and more accurate than the original dense model and simple pruning baselines, with minimal inference-time overhead. It achieves good accuracy with only a few epochs of pre-training, eliminating the need for full dense model training.
Conclusion: Frank-Wolfe with momentum is an effective pruning method for CNNs that can create sparser, more accurate networks with reduced pre-training requirements, supporting the Lottery Ticket Hypothesis in convolutional architectures.
Abstract: We investigate algorithmic variants of the Frank-Wolfe (FW) optimization method for pruning convolutional neural networks. This is motivated by the “Lottery Ticket Hypothesis”, which suggests the existence of smaller sub-networks within larger pre-trained networks that perform comparatively well (if not better). Whilst most literature in this area focuses on Deep Neural Networks more generally, we specifically consider Convolutional Neural Networks for image classification tasks. Building on the hypothesis, we compare simple magnitude-based pruning, a Frank-Wolfe style pruning scheme, and an FW method with momentum on a CNN trained on MNIST. Our experiments track test accuracy, loss, sparsity, and inference time as we vary the dense pre-training budget from 1 to 10 epochs. We find that FW with momentum yields pruned networks that are both sparser and more accurate than the original dense model and the simple pruning baselines, while incurring minimal inference-time overhead in our implementation. Moreover, FW with momentum reaches these accuracies after only a few epochs of pre-training, indicating that full pre-training of the dense model is not required in this setting.
[729] Dynamic Algorithm for Explainable k-medians Clustering under lp Norm
Konstantin Makarychev, Ilias Papanikolaou, Liren Shan
Main category: cs.LG
TL;DR: First algorithm for explainable k-medians under all lp norms (p≥1) with O(p(log k)^{1+1/p-1/p²}) approximation, plus dynamic implementation for streaming data.
Details
Motivation: Need interpretable clustering where users can trace and understand cluster assignments via threshold decision trees, especially for evolving datasets requiring dynamic updates.Method: Construct threshold decision trees that partition data into k clusters while minimizing k-medians objective under lp norms; also develop dynamic algorithm with amortized O(d log³ k) update time and O(log k) recourse.
Result: Achieves O(p(log k)^{1+1/p-1/p²}) approximation for any p≥1; improves existing O(log^{3/2}k) bound for p=2; matches tight log k+O(1) bound for p=1 up to O(log log k) factor.
Conclusion: First comprehensive solution for explainable k-medians across all lp norms with strong approximation guarantees and practical dynamic implementation for real-world applications.
Abstract: We study the problem of explainable k-medians clustering introduced by Dasgupta, Frost, Moshkovitz, and Rashtchian (2020). In this problem, the goal is to construct a threshold decision tree that partitions data into k clusters while minimizing the k-medians objective. These trees are interpretable because each internal node makes a simple decision by thresholding a single feature, allowing users to trace and understand how each point is assigned to a cluster. We present the first algorithm for explainable k-medians under lp norm for every finite p >= 1. Our algorithm achieves an O(p(log k)^{1 + 1/p - 1/p^2}) approximation to the optimal k-medians cost for any p >= 1. Previously, algorithms were known only for p = 1 and p = 2. For p = 2, our algorithm improves upon the existing bound of O(log^{3/2}k), and for p = 1, it matches the tight bound of log k + O(1) up to a multiplicative O(log log k) factor. We show how to implement our algorithm in a dynamic setting. The dynamic algorithm maintains an explainable clustering under a sequence of insertions and deletions, with amortized update time O(d log^3 k) and O(log k) recourse, making it suitable for large-scale and evolving datasets.
[730] Fiber Bundle Networks: A Geometric Machine Learning Paradigm
Dong Liu
Main category: cs.LG
TL;DR: FiberNet integrates differential geometry with ML, reformulating classification as geometric optimization on fiber bundles for interpretability and efficiency.
Details
Motivation: Traditional deep neural networks lack interpretability as black-box function approximators. The authors aim to create a framework that provides both interpretability and efficiency by integrating geometric principles from differential geometry into machine learning.Method: FiberNet reformulates classification as geometric optimization on fiber bundles: categories form the base space, wavelet-transformed features lie in fibers. Two key innovations: (1) learnable Riemannian metrics to identify important frequency components, (2) variational prototype optimization via energy minimization. Classification uses Voronoi tessellation under learned metrics.
Result: The framework provides clear geometric interpretability where each prototype defines a decision region and test samples are assigned to nearest prototypes. This enables simultaneous achievement of interpretability and efficiency, which are typically difficult to obtain together in conventional deep learning.
Conclusion: Integrating fiber bundle theory with machine learning offers a promising approach to achieve both interpretability and efficiency in classification tasks, addressing a key limitation of traditional deep learning methods.
Abstract: We propose Fiber Bundle Networks (FiberNet), a novel machine learning framework integrating differential geometry with machine learning. Unlike traditional deep neural networks relying on black-box function fitting, we reformulate classification as interpretable geometric optimization on fiber bundles, where categories form the base space and wavelet-transformed features lie in the fibers above each category. We introduce two innovations: (1) learnable Riemannian metrics identifying important frequency feature components, (2) variational prototype optimization through energy function minimization. Classification is performed via Voronoi tessellation under the learned Riemannian metric, where each prototype defines a decision region and test samples are assigned to the nearest prototype, providing clear geometric interpretability. This work demonstrates that the integration of fiber bundle with machine learning provides interpretability and efficiency, which are difficult to obtain simultaneously in conventional deep learning.
[731] Open-Set Domain Adaptation Under Background Distribution Shift: Challenges and A Provably Efficient Solution
Shravan Chaudhari, Yoav Wald, Suchi Saria
Main category: cs.LG
TL;DR: A method called OURS is developed to solve open-set recognition even when the background distribution shifts, with theoretical guarantees and empirical validation showing it outperforms existing methods.
Details
Motivation: Real-world ML systems face data shifts where new classes emerge (open-set recognition) and known class distributions change. Existing open-set recognition guarantees assume fixed background distributions, but this paper addresses the challenging case where background distributions shift.Method: Developed OURS method with theoretical guarantees under assumptions that novel classes are separable from non-novel classes. Created techniques to make the method scalable and robust, with comprehensive empirical evaluations on image and text data.
Result: OURS significantly outperforms existing open-set recognition methods under background shift. Provides new insights into how factors like novel class size influence performance, an aspect not extensively explored in prior work.
Conclusion: The paper presents a theoretically-grounded and empirically-validated method for open-set recognition that works even when background distributions shift, advancing the field beyond fixed-distribution assumptions.
Abstract: As we deploy machine learning systems in the real world, a core challenge is to maintain a model that is performant even as the data shifts. Such shifts can take many forms: new classes may emerge that were absent during training, a problem known as open-set recognition, and the distribution of known categories may change. Guarantees on open-set recognition are mostly derived under the assumption that the distribution of known classes, which we call \emph{the background distribution}, is fixed. In this paper we develop \ours{}, a method that is guaranteed to solve open-set recognition even in the challenging case where the background distribution shifts. We prove that the method works under benign assumptions that the novel class is separable from the non-novel classes, and provide theoretical guarantees that it outperforms a representative baseline in a simplified overparameterized setting. We develop techniques to make \ours{} scalable and robust, and perform comprehensive empirical evaluations on image and text data. The results show that \ours{} significantly outperforms existing open-set recognition methods under background shift. Moreover, we provide new insights into how factors such as the size of the novel class influences performance, an aspect that has not been extensively explored in prior work.
[732] From Regression to Classification: Exploring the Benefits of Categorical Representations of Energy in MLIPs
Ahmad Ali
Main category: cs.LG
TL;DR: MLIPs using classification instead of regression: predict energy/force distributions via histograms with cross-entropy loss, achieving similar accuracy while enabling uncertainty quantification.
Details
Motivation: Current MLIPs use scalar regression to predict single energy/force values, which lacks uncertainty quantification. The authors propose a classification approach to provide richer supervision and principled uncertainty estimation.Method: Convert scalar energy/force targets into histograms, predict categorical distributions over energy/force values using cross-entropy loss, and quantify epistemic uncertainty through distribution entropy.
Result: The categorical formulation achieves absolute error performance comparable to regression baselines while enabling uncertainty quantification through predicted distribution entropy.
Conclusion: Classification-based MLIPs offer a viable alternative to regression approaches, providing similar accuracy while adding principled uncertainty quantification capabilities absent in traditional scalar regression methods.
Abstract: Density Functional Theory (DFT) is a widely used computational method for estimating the energy and behavior of molecules. Machine Learning Interatomic Potentials (MLIPs) are models trained to approximate DFT-level energies and forces at dramatically lower computational cost. Many modern MLIPs rely on a scalar regression formulation; given information about a molecule, they predict a single energy value and corresponding forces while minimizing absolute error with DFT’s calculations. In this work, we explore a multi-class classification formulation that predicts a categorical distribution over energy/force values, providing richer supervision through multiple targets. Most importantly, this approach offers a principled way to quantify model uncertainty. In particular, our method predicts a histogram of the energy/force distribution, converts scalar targets into histograms, and trains the model using cross-entropy loss. Our results demonstrate that this categorical formulation can achieve absolute error performance comparable to regression baselines. Furthermore, this representation enables the quantification of epistemic uncertainty through the entropy of the predicted distribution, offering a measure of model confidence absent in scalar regression approaches.
[733] 2D-ThermAl: Physics-Informed Framework for Thermal Analysis of Circuits using Generative AI
Soumyadeep Chandra, Sayeed Shafayet Chowdhury, Kaushik Roy
Main category: cs.LG
TL;DR: ThermAl is a physics-informed generative AI framework that predicts full-chip thermal distributions from activity profiles, achieving 0.71°C RMSE and 200× speedup over traditional FEM methods for early-stage thermal analysis.
Details
Motivation: Traditional FEM-based thermal simulations are computationally expensive for early-stage IC design, requiring multiple redesign cycles to address late-stage thermal failures. There's a need for fast, accurate thermal analysis during early design phases.Method: ThermAl uses a physics-informed generative AI framework with a hybrid U-Net architecture enhanced with positional encoding and Boltzmann regularizer to maintain physical fidelity. Trained on extensive COMSOL-generated heat dissipation maps from simple logic gates to complex designs.
Result: Achieves 0.71°C RMSE for temperature mapping, runs ~200× faster than conventional FEM tools. Maintains high accuracy (<2.2% full-scale RMSE) across extended temperature range 25-95°C through cross-validation, demonstrating generalization beyond nominal operating conditions.
Conclusion: ThermAl provides fast, accurate early-stage thermal analysis for IC design, enabling hotspot detection and thermal pattern learning. The framework shows strong generalization across temperature ranges and has potential for integration into large-scale EDA workflows.
Abstract: Thermal analysis is increasingly critical in modern integrated circuits, where non-uniform power dissipation and high transistor densities can cause rapid temperature spikes and reliability concerns. Traditional methods, such as FEM-based simulations offer high accuracy but computationally prohibitive for early-stage design, often requiring multiple iterative redesign cycles to resolve late-stage thermal failures. To address these challenges, we propose ‘ThermAl’, a physics-informed generative AI framework which effectively identifies heat sources and estimates full-chip transient and steady-state thermal distributions directly from input activity profiles. ThermAl employs a hybrid U-Net architecture enhanced with positional encoding and a Boltzmann regularizer to maintain physical fidelity. Our model is trained on an extensive dataset of heat dissipation maps, ranging from simple logic gates (e.g., inverters, NAND, XOR) to complex designs, generated via COMSOL. Experimental results demonstrate that ThermAl delivers precise temperature mappings for large circuits, with a root mean squared error (RMSE) of only 0.71°C, and outperforms conventional FEM tools by running up to ~200 times faster. We analyze performance across diverse layouts and workloads, and discuss its applicability to large-scale EDA workflows. While thermal reliability assessments often extend beyond 85°C for post-layout signoff, our focus here is on early-stage hotspot detection and thermal pattern learning. To ensure generalization beyond the nominal operating range 25-55°C, we additionally performed cross-validation on an extended dataset spanning 25-95°C maintaining a high accuracy (<2.2% full-scale RMSE) even under elevated temperature conditions representative of peak power and stress scenarios.
[734] A TinyML Reinforcement Learning Approach for Energy-Efficient Light Control in Low-Cost Greenhouse Systems
Mohamed Abdallah Salem, Manuel Cuevas Perez, Ahmed Harb Rabia
Main category: cs.LG
TL;DR: RL-based Q-learning implemented on low-power microcontroller for adaptive LED lighting control using LDR sensor feedback, achieving stabilization at 13 distinct light intensity levels with minimal overshooting.
Details
Motivation: To develop a lightweight, on-device reinforcement learning solution for energy-efficient lighting control in resource-constrained environments like agricultural systems, enabling adaptive environmental regulation without heavy computational requirements.Method: Implemented model-free Q-learning algorithm on low-power microcontroller to dynamically adjust LED brightness based on real-time LDR sensor feedback. System trained to stabilize at 13 distinct light intensity levels (L1-L13) within a 64-state space derived from LDR readings, with 130 total trials (10 episodes per target level).
Result: Experimental validation showed the agent could effectively learn to stabilize at varying light levels with minimal overshooting and smooth convergence, even with environmental perturbations. Performance evaluated through convergence speed, steps taken, and time to reach targets, with box plots and histograms analyzing training time and learning efficiency distribution.
Conclusion: Demonstrates feasibility of lightweight, on-device RL for energy-efficient lighting control and establishes groundwork for multi-modal environmental control applications in resource-constrained agricultural systems.
Abstract: This study presents a reinforcement learning (RL)-based control strategy for adaptive lighting regulation in controlled environments using a low-power microcontroller. A model-free Q-learning algorithm was implemented to dynamically adjust the brightness of a Light-Emitting Diode (LED) based on real-time feedback from a light-dependent resistor (LDR) sensor. The system was trained to stabilize at 13 distinct light intensity levels (L1 to L13), with each target corresponding to a specific range within the 64-state space derived from LDR readings. A total of 130 trials were conducted, covering all target levels with 10 episodes each. Performance was evaluated in terms of convergence speed, steps taken, and time required to reach target states. Box plots and histograms were generated to analyze the distribution of training time and learning efficiency across targets. Experimental validation demonstrated that the agent could effectively learn to stabilize at varying light levels with minimal overshooting and smooth convergence, even in the presence of environmental perturbations. This work highlights the feasibility of lightweight, on-device RL for energy-efficient lighting control and sets the groundwork for multi-modal environmental control applications in resource-constrained agricultural systems.
[735] Data assimilation and discrepancy modeling with shallow recurrent decoders
Yuxuan Bao, J. Nathan Kutz
Main category: cs.LG
TL;DR: DA-SHRED is a machine learning framework that bridges simulation-to-real gaps by combining shallow recurrent decoders with data assimilation to reconstruct full system states from sparse sensor measurements and recover missing dynamics.
Details
Motivation: Modern sensing faces challenges with data efficiency, real-time processing, and limited coverage. Simulation models often neglect small-scale processes, are sensitive to perturbations, or oversimplify parameter correlations, leading to reconstructions that diverge from reality. There's a critical need to integrate observational data with predictive simulations for accurate state estimation.Method: DA-SHRED uses a shallow recurrent decoder architecture that: 1) Learns latent space from reduced simulation models via SHRED, 2) Updates latent variables using real sensor data to reconstruct full system states, 3) Incorporates sparse identification of nonlinear dynamics (SINDy) regression in latent space to identify missing dynamics functionals.
Result: The framework successfully closes the SIM2REAL gap and recovers missing dynamics in highly complex systems, demonstrating that efficient temporal encoding combined with physics-informed correction enables robust data assimilation.
Conclusion: DA-SHRED provides an effective machine learning approach for data assimilation that bridges computational modeling with experimental sensor data, enabling accurate reconstruction of high-dimensional spatiotemporal fields from sparse measurements while identifying and correcting for missing dynamics in simulation models.
Abstract: The requirements of modern sensing are rapidly evolving, driven by increasing demands for data efficiency, real-time processing, and deployment under limited sensing coverage. Complex physical systems are often characterized through the integration of a limited number of point sensors in combination with scientific computations which approximate the dominant, full-state dynamics. Simulation models, however, inevitably neglect small-scale or hidden processes, are sensitive to perturbations, or oversimplify parameter correlations, leading to reconstructions that often diverge from the reality measured by sensors. This creates a critical need for data assimilation, the process of integrating observational data with predictive simulation models to produce coherent and accurate estimates of the full state of complex physical systems. We propose a machine learning framework for Data Assimilation with a SHallow REcurrent Decoder (DA-SHRED) which bridges the simulation-to-real (SIM2REAL) gap between computational modeling and experimental sensor data. For real-world physics systems modeling high-dimensional spatiotemporal fields, where the full state cannot be directly observed and must be inferred from sparse sensor measurements, we leverage the latent space learned from a reduced simulation model via SHRED, and update these latent variables using real sensor data to accurately reconstruct the full system state. Furthermore, our algorithm incorporates a sparse identification of nonlinear dynamics based regression model in the latent space to identify functionals corresponding to missing dynamics in the simulation model. We demonstrate that DA-SHRED successfully closes the SIM2REAL gap and additionally recovers missing dynamics in highly complex systems, demonstrating that the combination of efficient temporal encoding and physics-informed correction enables robust data assimilation.
[736] First On-Orbit Demonstration of a Geospatial Foundation Model
Andrew Du, Roberto Del Prete, Alejandro Mousist, Nick Manser, Fabrice Marre, Andrew Barton, Carl Seubert, Gabriele Meoni, Tat-Jun Chin
Main category: cs.LG
TL;DR: Compact Vision Transformer-based geospatial foundation models enable onboard AI execution on resource-constrained space hardware while maintaining performance.
Details
Motivation: Large geospatial foundation models (GeoFMs) have strong generalization for Earth observation tasks but are too big for deployment on resource-constrained space hardware, creating a barrier to onboard AI execution.Method: Develop compact variants of Vision Transformer-based GeoFMs using model compression and domain adaptation techniques to reduce size and resource demands while preserving downstream task performance.
Result: Evaluation across five downstream tasks and validation in two flight environments shows successful size reduction with maintained high performance. Demonstrated reliable on-orbit inference with IMAGIN-e payload aboard the International Space Station.
Conclusion: Model compression and domain adaptation enable transition from large GeoFMs to flight-ready, resource-efficient deployments, expanding feasibility of onboard AI for Earth observation missions.
Abstract: Geospatial foundation models (GeoFMs) promise broad generalisation capacity for Earth observation (EO) tasks, particularly under data-limited conditions. However, their large size poses a barrier to deployment on resource-constrained space hardware. To address this, we present compact variants of a Vision Transformer (ViT)-based GeoFM that preserve downstream task performance while enabling onboard execution. Evaluation across five downstream tasks and validation in two representative flight environments show that model compression and domain adaptation are critical to reducing size and resource demands while maintaining high performance under operational conditions. We further demonstrate reliable on-orbit inference with the IMAGIN-e payload aboard the International Space Station. These results establish a pathway from large GeoFMs to flight-ready, resource-efficient deployments, expanding the feasibility of onboard AI for EO missions.
[737] Teaching by Failure: Counter-Example-Driven Curricula for Transformer Self-Improvement
Harshil Vejendla
Main category: cs.LG
TL;DR: CEDC improves Transformer robustness by iteratively fine-tuning on automatically generated counter-examples from model failures, achieving up to 30x better length extrapolation without manual difficulty heuristics.
Details
Motivation: Transformers often fail to generalize to inputs longer or more complex than training data, showing brittle extrapolation. Current methods require manual difficulty heuristics or are computationally inefficient.Method: Counter-Example-Driven Curricula (CEDC): 1) Generate diverse candidate problems using current model, 2) Use fast executable verifier to identify incorrect predictions (counter-examples), 3) Fine-tune model on dataset enriched with discovered failures, iteratively.
Result: CEDC achieves up to 30x greater length extrapolation than baselines, is 3.75x more computationally efficient than uniform data augmentation, and requires no manual difficulty heuristics. Works on integer addition, sorting, Dyck-2 recognition, and text classification.
Conclusion: Verifier-guided, failure-driven learning is a simple, powerful, and efficient paradigm for enhancing Transformer generalization capabilities, with curriculum naturally adapting to target progressively more complex error modes.
Abstract: Transformer models often exhibit brittle extrapolation, failing on inputs that are longer or structurally more complex than those seen during training. We introduce Counter-Example-Driven Curricula (CEDC), an automated framework that improves model robustness by iteratively focusing on its own failures. At each step, CEDC uses the current model to generate a diverse set of candidate problems, employs a fast, executable verifier to identify incorrect predictions (counter-examples), and then fine-tunes the model on a dataset enriched with these discovered failures. We evaluate CEDC on a suite of algorithmic and natural language tasks, including integer addition, sorting, Dyck-2 language recognition, and three text classification benchmarks. Compared to static training and standard curriculum learning baselines, CEDC achieves up to 30x greater length extrapolation, is 3.75x more computationally efficient than uniform data augmentation, and requires no manual difficulty heuristics. We provide a detailed analysis of the counter-examples, showing how the curriculum naturally adapts to target progressively more complex error modes. Our findings establish verifier-guided, failure-driven learning as a simple, powerful, and efficient paradigm for enhancing the generalization capabilities of Transformer models.
[738] LGDC: Latent Graph Diffusion via Spectrum-Preserving Coarsening
Nagham Osman, Keyue Jiang, Davide Buffelli, Xiaowen Dong, Laura Toni
Main category: cs.LG
TL;DR: LGDC is a hybrid graph generation framework that combines autoregressive and diffusion models to capture both local and global graph structures efficiently.
Details
Motivation: Existing graph generation methods have a trade-off: autoregressive models capture local structures well but diffusion models excel at global patterns. The paper aims to combine both strengths.Method: LGDC uses spectrum-preserving coarsening-decoarsening to map graphs to a latent space, where diffusion generates latent graphs, then expansion restores detail - combining autoregressive and diffusion approaches.
Result: LGDC matches autoregressive models on locally structured datasets (Tree) and diffusion models on globally structured ones (Planar, Community-20), validating hybrid benefits.
Conclusion: The hybrid LGDC framework successfully combines strengths of both autoregressive and diffusion models for graph generation, capturing both local and global properties efficiently.
Abstract: Graph generation is a critical task across scientific domains. Existing methods fall broadly into two categories: autoregressive models, which iteratively expand graphs, and one-shot models, such as diffusion, which generate the full graph at once. In this work, we provide an analysis of these two paradigms and reveal a key trade-off: autoregressive models stand out in capturing fine-grained local structures, such as degree and clustering properties, whereas one-shot models excel at modeling global patterns, such as spectral distributions. Building on this, we propose LGDC (latent graph diffusion via spectrum-preserving coarsening), a hybrid framework that combines strengths of both approaches. LGDC employs a spectrum-preserving coarsening-decoarsening to bidirectionally map between graphs and a latent space, where diffusion efficiently generates latent graphs before expansion restores detail. This design captures both local and global properties with improved efficiency. Empirically, LGDC matches autoregressive models on locally structured datasets (Tree) and diffusion models on globally structured ones (Planar, Community-20), validating the benefits of hybrid generation.
[739] Learning to Reconstruct Temperature Field from Sparse Observations with Implicit Physics Priors
Shihang Li, Zhiqiang Gong, Weien Zhou, Yue Gao, Wen Yao
Main category: cs.LG
TL;DR: IPTR is an implicit physics-guided framework for temperature field reconstruction that leverages reference simulation data as priors to improve generalization across varying conditions.
Details
Motivation: Accurate temperature field reconstruction is crucial for thermal monitoring in engineering applications, but existing methods struggle with high measurement costs and distributional shifts across conditions. Current DNN-based approaches don't effectively use reference simulation data that encodes valuable thermal knowledge.Method: IPTR introduces a dual physics embedding module with two branches: (1) implicit physics-guided branch using cross-attention to distill latent physics from reference simulation data, and (2) auxiliary encoding branch with Fourier layers to capture spatial characteristics of target observations. The fused representation is decoded to reconstruct the full temperature field.
Result: Extensive experiments under single-condition, multi-condition, and few-shot settings show IPTR consistently outperforms existing methods, achieving state-of-the-art reconstruction accuracy and strong generalization capability.
Conclusion: IPTR effectively addresses the limitations of existing methods by leveraging reference simulation data as implicit physics priors, enabling robust temperature field reconstruction with improved generalization across varying operational conditions.
Abstract: Accurate reconstruction of temperature field of heat-source systems (TFR-HSS) is crucial for thermal monitoring and reliability assessment in engineering applications such as electronic devices and aerospace structures. However, the high cost of measurement acquisition and the substantial distributional shifts in temperature field across varying conditions present significant challenges for developing reconstruction models with robust generalization capabilities. Existing DNNs-based methods typically formulate TFR-HSS as a one-to-one regression problem based solely on target sparse measurements, without effectively leveraging reference simulation data that implicitly encode thermal knowledge. To address this limitation, we propose IPTR, an implicit physics-guided temperature field reconstruction framework that introduces sparse monitoring-temperature field pair from reference simulations as priors to enrich physical understanding. To integrate both reference and target information, we design a dual physics embedding module consisting of two complementary branches: an implicit physics-guided branch employing cross-attention to distill latent physics from the reference data, and an auxiliary encoding branch based on Fourier layers to capture the spatial characteristics of the target observation. The fused representation is then decoded to reconstruct the full temperature field. Extensive experiments under single-condition, multi-condition, and few-shot settings demonstrate that IPTR consistently outperforms existing methods, achieving state-of-the-art reconstruction accuracy and strong generalization capability.
[740] Know Thyself by Knowing Others: Learning Neuron Identity from Population Context
Vinam Arora, Divyansha Lachi, Ian J. Knight, Mehdi Azabou, Blake Richards, Cole L. Hurwitz, Josh Siegle, Eva L. Dyer
Main category: cs.LG
TL;DR: NuCLR is a self-supervised framework that learns neuron identity representations from neural activity using contrastive learning and spatiotemporal transformers, achieving SOTA performance on cell type and brain region decoding with strong cross-animal generalization.
Details
Motivation: Inferring neuron identity (cell type, connectivity, brain region) from neural activity remains challenging. Current methods lack general-purpose representations that can differentiate individual neurons across different conditions and animals.Method: NuCLR uses self-supervised contrastive learning to bring together views of the same neuron observed at different times and stimuli. It employs a spatiotemporal transformer that integrates neural activity in a permutation-equivariant manner to capture population context without fixed neuron ordering.
Result: Achieves state-of-the-art performance on cell type and brain region decoding across multiple electrophysiology and calcium imaging datasets. Shows strong zero-shot generalization to unseen animals and label efficiency (requires only small fraction of labeled samples). First systematic scaling analysis shows performance improves with more animals during pretraining.
Conclusion: Large, diverse neural datasets enable models to learn neuron identity representations that generalize across animals. NuCLR demonstrates the potential of self-supervised learning for extracting meaningful biological information from neural activity data.
Abstract: Neurons process information in ways that depend on their cell type, connectivity, and the brain region in which they are embedded. However, inferring these factors from neural activity remains a significant challenge. To build general-purpose representations that allow for resolving information about a neuron’s identity, we introduce NuCLR, a self-supervised framework that aims to learn representations of neural activity that allow for differentiating one neuron from the rest. NuCLR brings together views of the same neuron observed at different times and across different stimuli and uses a contrastive objective to pull these representations together. To capture population context without assuming any fixed neuron ordering, we build a spatiotemporal transformer that integrates activity in a permutation-equivariant manner. Across multiple electrophysiology and calcium imaging datasets, a linear decoding evaluation on top of NuCLR representations achieves a new state-of-the-art for both cell type and brain region decoding tasks, and demonstrates strong zero-shot generalization to unseen animals. We present the first systematic scaling analysis for neuron-level representation learning, showing that increasing the number of animals used during pretraining consistently improves downstream performance. The learned representations are also label-efficient, requiring only a small fraction of labeled samples to achieve competitive performance. These results highlight how large, diverse neural datasets enable models to recover information about neuron identity that generalize across animals. Code is available at https://github.com/nerdslab/nuclr.
[741] Sum Rate Maximization in STAR-RIS-UAV-Assisted Networks: A CA-DDPG Approach for Joint Optimization
Yujie Huang, Haibin Wan, Xiangcheng Li, Tuanfa Qin, Yun Li, Jun Li, Wen Chen
Main category: cs.LG
TL;DR: This paper proposes a CA-DDPG algorithm to optimize beamforming, phase shifts, and UAV positioning in STAR-RIS-UAV wireless systems to maximize sum rate.
Details
Motivation: STAR-RIS technology enables comprehensive signal control but needs enhanced flexibility; UAV integration can further improve system adaptability and performance in wireless communications.Method: Proposes a convolution-augmented deep deterministic policy gradient (CA-DDPG) algorithm with stochastic perturbation for exploration, optimizing beamforming, phase shifts, and UAV positioning through DRL.
Result: CA-DDPG effectively optimizes system parameters, improves system capacity, and outperforms other algorithms in simulation results.
Conclusion: The proposed CA-DDPG algorithm successfully enhances STAR-RIS-UAV system performance by balancing exploration and evaluation, achieving superior sum rate optimization.
Abstract: With the rapid advances in programmable materials, reconfigurable intelligent surfaces (RIS) have become a pivotal technology for future wireless communications. The simultaneous transmitting and reflecting reconfigurable intelligent surfaces (STAR-RIS) can both transmit and reflect signals, enabling comprehensive signal control and expanding application scenarios. This paper introduces an unmanned aerial vehicle (UAV) to further enhance system flexibility and proposes an optimization design for the spectrum efficiency of the STAR-RIS-UAV-assisted wireless communication system. We present a deep reinforcement learning (DRL) algorithm capable of iteratively optimizing beamforming, phase shifts, and UAV positioning to maximize the system’s sum rate through continuous interactions with the environment. To improve exploration in deterministic policies, we introduce a stochastic perturbation factor, which enhances exploration capabilities. As exploration is strengthened, the algorithm’s ability to accurately evaluate the state-action value function becomes critical. Thus, based on the deep deterministic policy gradient (DDPG) algorithm, we propose a convolution-augmented deep deterministic policy gradient (CA-DDPG) algorithm that balances exploration and evaluation to improve the system’s sum rate. The simulation results demonstrate that the CA-DDPG algorithm effectively interacts with the environment, optimizing the beamforming matrix, phase shift matrix, and UAV location, thereby improving system capacity and achieving better performance than other algorithms.
[742] Research on Milling Machine Predictive Maintenance Based on Machine Learning and SHAP Analysis in Intelligent Manufacturing Environment
Wen Zhao, Jiawen Ding, Xueting Huang, Yibo Zhang
Main category: cs.LG
TL;DR: Experimental study on predictive maintenance for industrial milling machines using AI4I 2020 dataset, comparing 8 ML models with XGBoost and random forest performing best, using SHAP analysis to identify key failure factors.
Details
Motivation: To address predictive maintenance needs in intelligent manufacturing by combining AI technology with manufacturing processes to improve production efficiency, reduce maintenance costs, and support digital transformation of the manufacturing industry.Method: Proposes a complete predictive maintenance experimental process with six main links: data preprocessing, model training, model evaluation, model selection, SHAP analysis, and result visualization. Compares eight machine learning models on the AI4I 2020 dataset.
Result: Integrated learning methods (XGBoost and random forest) performed best in milling machine fault prediction. SHAP analysis revealed processing temperature, torque, and speed as key factors influencing equipment failure.
Conclusion: The study provides methodological reference for predictive maintenance in intelligent manufacturing, combining AI and manufacturing technology to support digital transformation, improve production efficiency, and reduce maintenance costs.
Abstract: In the context of intelligent manufacturing, this paper conducts a series of experimental studies on the predictive maintenance of industrial milling machine equipment based on the AI4I 2020 dataset. This paper proposes a complete predictive maintenance experimental process combining artificial intelligence technology, including six main links: data preprocessing, model training, model evaluation, model selection, SHAP analysis, and result visualization. By comparing and analyzing the performance of eight machine learning models, it is found that integrated learning methods such as XGBoost and random forest perform well in milling machine fault prediction tasks. In addition, with the help of SHAP analysis technology, the influence mechanism of different features on equipment failure is deeply revealed, among which processing temperature, torque and speed are the key factors affecting failure. This study combines artificial intelligence and manufacturing technology, provides a methodological reference for predictive maintenance practice in an intelligent manufacturing environment, and has practical significance for promoting the digital transformation of the manufacturing industry, improving production efficiency and reducing maintenance costs.
[743] A Comparative Study of Machine Learning Algorithms for Electricity Price Forecasting with LIME-Based Interpretability
Xuanyi Zhao, Jiawen Ding, Xueting Huang, Yibo Zhang
Main category: cs.LG
TL;DR: This study compares eight machine learning models for electricity price forecasting using Spanish market data, finding KNN performs best with R²=0.865, and uses LIME for interpretability to reveal meteorological and supply-demand factors as key nonlinear influencers.
Details
Motivation: With rapid electricity market development and increased price volatility, accurate forecasting is crucial for power system operations and market decisions. Traditional linear models fail to capture complex nonlinear pricing characteristics, necessitating advanced machine learning approaches.Method: The study compares eight machine learning models (linear regression, ridge regression, decision tree, KNN, random forest, gradient boosting, SVR, and XGBoost) using Spanish electricity market data integrated with consumption, generation, and meteorological variables. LIME analysis is used for model interpretability.
Result: KNN achieves the best performance with R² of 0.865, MAE of 3.556, and RMSE of 5.240. LIME analysis reveals that meteorological factors and supply-demand indicators significantly influence price fluctuations through nonlinear relationships.
Conclusion: This work demonstrates the effectiveness of machine learning models in electricity price forecasting while improving decision transparency through interpretability analysis, showing that KNN performs best among the tested models and that interpretability tools like LIME can reveal important nonlinear relationships in pricing factors.
Abstract: With the rapid development of electricity markets, price volatility has significantly increased, making accurate forecasting crucial for power system operations and market decisions. Traditional linear models cannot capture the complex nonlinear characteristics of electricity pricing, necessitating advanced machine learning approaches. This study compares eight machine learning models using Spanish electricity market data, integrating consumption, generation, and meteorological variables. The models evaluated include linear regression, ridge regression, decision tree, KNN, random forest, gradient boosting, SVR, and XGBoost. Results show that KNN achieves the best performance with R^2 of 0.865, MAE of 3.556, and RMSE of 5.240. To enhance interpretability, LIME analysis reveals that meteorological factors and supply-demand indicators significantly influence price fluctuations through nonlinear relationships. This work demonstrates the effectiveness of machine learning models in electricity price forecasting while improving decision transparency through interpretability analysis.
[744] Neural Network Optimal Power Flow via Energy Gradient Flow and Unified Dynamics
Xuezhi Liu
Main category: cs.LG
TL;DR: Proposes an unsupervised neural network method for Optimal Power Flow using energy gradient flow to transform OPF into energy minimization, eliminating need for labeled data while ensuring physical consistency.
Details
Motivation: Traditional OPF methods have computational inefficiency, sensitivity to initial values, and poor batch computation. Supervised learning approaches require large pre-solved datasets and struggle with physical consistency guarantees.Method: Transforms OPF into energy minimization problem by constructing energy function measuring deviation from constraint manifold. Uses neural network dynamics and gradient flow to guide networks to learn optimal solutions satisfying both power flow constraints and cost minimization, trained unsupervised by minimizing physical residuals.
Result: Achieves true “end-to-end” physics-constrained learning without requiring labeled data, enabling networks to directly learn optimal solutions that satisfy physical constraints while minimizing costs.
Conclusion: The proposed unsupervised neural network approach with energy gradient flow provides an efficient, physically-consistent solution to OPF problems without the data requirements of supervised methods, addressing key limitations of traditional optimization algorithms.
Abstract: Optimal Power Flow (OPF) is a core optimization problem in power system operation and planning, aiming to minimize generation costs while satisfying physical constraints such as power flow equations, generator limits, and voltage limits. Traditional OPF solving methods typically employ iterative optimization algorithms (such as interior point methods, sequential quadratic programming, etc.), with limitations including low computational efficiency, initial value sensitivity, and low batch computation efficiency. Most existing deep learning-based OPF methods rely on supervised learning, requiring pre-solving large numbers of cases, and have difficulty guaranteeing physical consistency. This paper proposes an Optimal Power Flow solving method based on neural network dynamics and energy gradient flow, transforming OPF problems into energy minimization problems. By constructing an energy function to measure the degree of deviation from the constraint manifold, and guiding networks to learn optimal solutions that simultaneously satisfy power flow constraints and minimize costs through gradient flow. Neural networks are trained unsupervised by directly minimizing physical residuals, requiring no labeled data, achieving true “end-to-end” physics-constrained learning.
[745] CoSineVerifier: Tool-Augmented Answer Verification for Computation-Oriented Scientific Questions
Ruixiang Feng, Zhenwei An, Yuntao Wen, Ran Le, Yiming Jia, Chen Yang, Zongchao Chen, Lisi Chen, Shen Gao, Shuo Shang, Yang Song, Tao Zhang
Main category: cs.LG
TL;DR: CoSineVerifier is a tool-augmented verifier that uses external executors for precise computations and symbolic simplifications, achieving SOTA on STEM verification benchmarks and improving RLVR performance.
Details
Motivation: Existing verifiers struggle with computation-oriented scientific domains like algebraic equivalence checking and physical constant substitution, requiring more robust verification beyond simple semantic matching.Method: Two-stage pipeline: cold-start fine-tuning followed by multi-turn reinforcement learning with tool integration, leveraging external executors for precise computations and symbolic simplifications.
Result: Achieves state-of-the-art performance on VerifyBench-Hard and SCI-Bench; when used as reward model in RLVR, consistently outperforms both rubric-based and model-based verifiers on AIME'24 and AIME'25.
Conclusion: CoSineVerifier demonstrates strong generalization across STEM, general QA, and long-form reasoning tasks, showing potential to enhance LLM reasoning capabilities through tool-augmented verification.
Abstract: Answer verification methods are widely employed in language model training pipelines spanning data curation, evaluation, and reinforcement learning with verifiable rewards (RLVR). While prior work focus on developing unified verifiers applicable across multiple reasoning scenarios, significant challenges remain in computation-oriented scientific domains, such as algebraic equivalence checking and physical constant substitution. In this paper, we introduce \model, a tool-augmented verifier that leverages external executors to perform precise computations and symbolic simplifications. \model enables robust verification that goes beyond simple semantic matching. We propose a novel two-stage pipeline, which begin with cold-start fine-tuning and followed by multi-turn reinforcement learning with tool integration. Extensive experiments conducted on STEM subjects, general QA, and long-form reasoning tasks demonstrates strong generalization of \model. The results shows that the \model achieves state-of-the-art performance on VerifyBench-Hard and SCI-Bench. And we also employ our \model in RLVR as a reward model, the results show that it consistently outperforms both rubric-based and model-based verifiers on AIME'24 and AIME'25, demonstrating strong potential to enhance reasoning capabilities of LLM. Our model is released at \hyperlink{https://huggingface.co/Nanbeige/CoSineVerifier-Tool-4B}{https://huggingface.co/Nanbeige/CoSineVerifier-Tool-4B}.
[746] On the Tension Between Optimality and Adversarial Robustness in Policy Optimization
Haoran Li, Jiayu Lv, Congying Han, Zicheng Zhang, Anqi Li, Yan Liu, Tiande Guo, Nan Jiang
Main category: cs.LG
TL;DR: BARPO: A bilevel framework that modulates adversary strength to reconcile optimality and adversarial robustness in deep reinforcement learning, outperforming vanilla adversarially robust policy optimization.
Details
Motivation: Recent theoretical insights suggest potential alignment between optimality and adversarial robustness in deep RL, but there's a key gap between theory and practice. Standard policy optimization (SPO) converges to vulnerable policies with strong performance, while adversarially robust policy optimization (ARPO) favors robustness at the expense of returns.Method: Developed BARPO, a bilevel framework that unifies SPO and ARPO by modulating adversary strength. This approach facilitates navigability while preserving global optima by addressing the reshaping effect of the strongest adversary in ARPO that induces deceptive sticky FOSPs.
Result: BARPO consistently outperforms vanilla ARPO in extensive empirical evaluations, providing a practical approach to reconcile theoretical and empirical performance by balancing robustness and optimality.
Conclusion: BARPO offers a practical solution to bridge the theory-practice gap in adversarial robustness for deep RL, demonstrating that modulating adversary strength can help achieve both optimal performance and robustness without the tradeoffs observed in existing methods.
Abstract: Achieving optimality and adversarial robustness in deep reinforcement learning has long been regarded as conflicting goals. Nonetheless, recent theoretical insights presented in CAR suggest a potential alignment, raising the important question of how to realize this in practice. This paper first identifies a key gap between theory and practice by comparing standard policy optimization (SPO) and adversarially robust policy optimization (ARPO). Although they share theoretical consistency, a fundamental tension between robustness and optimality arises in practical policy gradient methods. SPO tends toward convergence to vulnerable first-order stationary policies (FOSPs) with strong natural performance, whereas ARPO typically favors more robust FOSPs at the expense of reduced returns. Furthermore, we attribute this tradeoff to the reshaping effect of the strongest adversary in ARPO, which significantly complicates the global landscape by inducing deceptive sticky FOSPs. This improves robustness but makes navigation more challenging. To alleviate this, we develop the BARPO, a bilevel framework unifying SPO and ARPO by modulating adversary strength, thereby facilitating navigability while preserving global optima. Extensive empirical results demonstrate that BARPO consistently outperforms vanilla ARPO, providing a practical approach to reconcile theoretical and empirical performance.
[747] Efficient Training of Diffusion Mixture-of-Experts Models: A Practical Recipe
Yahui Liu, Yang Yue, Jingyuan Zhang, Chenxi Sun, Yang Zhou, Wencong Zeng, Ruiming Tang, Guorui Zhou
Main category: cs.LG
TL;DR: Systematic exploration of architectural configurations for Diffusion MoE models reveals that careful tuning of expert modules, widths, expert counts, and attention encodings yields greater gains than routing innovations alone.
Details
Motivation: Current Diffusion MoE research focuses heavily on routing mechanisms while neglecting architectural configuration space, which remains under-explored despite its potential for significant performance improvements.Method: Systematically study architectural factors inspired by LLM MoE designs: DeepSeek-style expert modules, alternative intermediate widths, varying expert counts, and enhanced attention positional encodings. Apply these to both latent and pixel-space diffusion frameworks.
Result: Novel architectures that efficiently surpass strong baselines using equal or fewer activated parameters, with gains exceeding those from routing innovations alone. Practical training recipe provided.
Conclusion: Architectural configuration tuning is essential for unlocking Diffusion MoE potential, offering a practical and efficient approach that outperforms routing-focused methods.
Abstract: Recent efforts on Diffusion Mixture-of-Experts (MoE) models have primarily focused on developing more sophisticated routing mechanisms. However, we observe that the underlying architectural configuration space remains markedly under-explored. Inspired by the MoE design paradigms established in large language models (LLMs), we identify a set of crucial architectural factors for building effective Diffusion MoE models–including DeepSeek-style expert modules, alternative intermediate widths, varying expert counts, and enhanced attention positional encodings. Our systematic study reveals that carefully tuning these configurations is essential for unlocking the full potential of Diffusion MoE models, often yielding gains that exceed those achieved by routing innovations alone. Through extensive experiments, we present novel architectures that can be efficiently applied to both latent and pixel-space diffusion frameworks, which provide a practical and efficient training recipe that enables Diffusion MoE models to surpass strong baselines while using equal or fewer activated parameters. All code and models are publicly available at: https://github.com/yhlleo/EfficientMoE.
[748] Efficient Hyperparameter Search for Non-Stationary Model Training
Berivan Isik, Matthew Fahrbach, Dima Kuzmin, Nicolas Mayoraz, Emil Praun, Steffen Rendle, Raghavendra Vasudeva
Main category: cs.LG
TL;DR: A two-stage hyperparameter optimization framework reduces search costs by 10× for online learning systems by first efficiently identifying promising configurations, then fully training only those candidates.
Details
Motivation: Online learning systems (recommendation/advertising) require expensive model training that multiplies during hyperparameter search, creating significant computational costs for systems with shifting data distributions.Method: Two-stage paradigm: (1) efficiently identify most promising configurations using novel data reduction and prediction strategies designed for sequential, non-stationary data, (2) train only selected candidates to full potential. Focuses on accurate identification rather than peak performance in first stage.
Result: Reduces total hyperparameter search cost by up to 10× on Criteo 1TB dataset (largest suitable public benchmark) and delivers significant efficiency gains on industrial advertising system at two orders of magnitude larger scale.
Conclusion: The framework effectively addresses the unique challenges of hyperparameter optimization for online learning systems with non-stationary data, achieving substantial cost reductions while maintaining model quality.
Abstract: Online learning is the cornerstone of applications like recommendation and advertising systems, where models continuously adapt to shifting data distributions. Model training for such systems is remarkably expensive, a cost that multiplies during hyperparameter search. We introduce a two-stage paradigm to reduce this cost: (1) efficiently identifying the most promising configurations, and then (2) training only these selected candidates to their full potential. Our core insight is that focusing on accurate identification in the first stage, rather than achieving peak performance, allows for aggressive cost-saving measures. We develop novel data reduction and prediction strategies that specifically overcome the challenges of sequential, non-stationary data not addressed by conventional hyperparameter optimization. We validate our framework’s effectiveness through a dual evaluation: first on the Criteo 1TB dataset, the largest suitable public benchmark, and second on an industrial advertising system operating at a scale two orders of magnitude larger. Our methods reduce the total hyperparameter search cost by up to 10$\times$ on the public benchmark and deliver significant, validated efficiency gains in the industrial setting.
[749] Accelerating Large-Scale Reasoning Model Inference with Sparse Self-Speculative Decoding
Yilong Zhao, Jiaming Tang, Kan Zhu, Zihao Ye, Chi-Chih Chang, Chaofan Lin, Jongseok Park, Guangxuan Xiao, Mohamed S. Abdelfattah, Mingyu Gao, Baris Kasikci, Song Han, Ion Stoica
Main category: cs.LG
TL;DR: SparseSpec is a speculative decoding framework that uses self-speculation with sparse attention to reduce memory bottlenecks in long chain-of-thought reasoning, achieving up to 2.13x throughput speedup.
Details
Motivation: Long chain-of-thought reasoning in language models shifts inference from compute-bound to memory-bound due to large KV-Cache requirements, causing memory bandwidth pressure during lengthy generations.Method: SparseSpec uses self-speculation (same model as draft and target) with PillarAttn sparse attention mechanism for token selection, plus system innovations: unified scheduler for batching, delayed verification for CPU/GPU overlap, and dynamic KV-Cache management.
Result: Across various models and datasets, SparseSpec outperforms state-of-the-art solutions with up to 2.13x throughput speedup.
Conclusion: SparseSpec effectively addresses memory bottlenecks in long reasoning tasks through self-speculation with sparse attention and system co-design, significantly improving inference throughput.
Abstract: Reasoning language models have demonstrated remarkable capabilities on challenging tasks by generating elaborate chain-of-thought (CoT) solutions. However, such lengthy generation shifts the inference bottleneck from compute-bound to memory-bound. To generate each token, the model applies full attention to all previously generated tokens, requiring memory access to an increasingly large KV-Cache. Consequently, longer generations demand more memory access for every step, leading to substantial pressure on memory bandwidth. To address this, we introduce SparseSpec, a speculative decoding framework that reuses the same model as the draft and target models (i.e., self-speculation). SparseSpec features a novel sparse attention mechanism, PillarAttn, as the draft model, which accurately selects critical tokens via elegantly reusing information from the verification stage. Furthermore, SparseSpec co-designs self-speculation with three system innovations: (1) a unified scheduler to batch token drafting and verification, (2) delayed verification for CPU/GPU overlap, and (3) dynamic KV-Cache management to maximize memory utilization. Across various models and datasets, SparseSpec outperforms state-of-the-art solutions, with an up to 2.13x throughput speedup.
[750] Generative Modeling with Continuous Flows: Sample Complexity of Flow Matching
Mudit Gaur, Prashant Trivedi, Shuchin Aeron, Amrit Singh Bedi, George K. Atia, Vaneet Aggarwal
Main category: cs.LG
TL;DR: First theoretical analysis of sample complexity for flow matching generative models, showing O(ε⁻⁴) samples needed to achieve O(ε) Wasserstein-2 distance without assuming access to ERM.
Details
Motivation: Flow matching has emerged as a promising alternative to diffusion models with faster sampling and simpler training, but lacks theoretical understanding of sample complexity. Existing analyses often assume access to empirical risk minimizer, which is unrealistic in practice.Method: Decompose velocity field estimation error into three components: neural network approximation error, statistical error from finite samples, and optimization error from finite training steps. Handle each component separately using techniques that may be of independent interest.
Result: With sufficiently expressive neural networks and under standard assumptions on loss function and bounded data distribution, O(ε⁻⁴) samples are sufficient to achieve O(ε) Wasserstein-2 distance between learned and true distributions.
Conclusion: Provides first theoretical sample complexity analysis for flow matching without unrealistic ERM assumptions, establishing O(ε⁻⁴) sample complexity bound and offering technical decomposition framework for analyzing generative model errors.
Abstract: Flow matching has recently emerged as a promising alternative to diffusion-based generative models, offering faster sampling and simpler training by learning continuous flows governed by ordinary differential equations. Despite growing empirical success, the theoretical understanding of flow matching remains limited, particularly in terms of sample complexity results. In this work, we provide the first analysis of the sample complexity for flow-matching based generative models without assuming access to the empirical risk minimizer (ERM) of the loss function for estimating the velocity field. Under standard assumptions on the loss function for velocity field estimation and boundedness of the data distribution, we show that a sufficiently expressive neural network can learn a velocity field such that with $\mathcal{O}(ε^{-4})$ samples, such that the Wasserstein-2 distance between the learned and the true distribution is less than $\mathcal{O}(ε)$. The key technical idea is to decompose the velocity field estimation error into neural-network approximation error, statistical error due to the finite sample size, and optimization error due to the finite number of optimization steps for estimating the velocity field. Each of these terms are then handled via techniques that may be of independent interest.
[751] milearn: A Python Package for Multi-Instance Machine Learning
Dmitry Zankov, Pavlo Polishchuk, Michal Sobieraj, Mario Barbatti
Main category: cs.LG
TL;DR: milearn is a scikit-learn compatible Python package for multi-instance learning that unifies classical and neural MIL methods with built-in hyperparameter optimization for small datasets.
Details
Motivation: There's a need for a unified framework for multi-instance learning that works with both classical and neural methods, provides scikit-learn compatibility, and addresses the challenge of hyperparameter optimization for small MIL datasets.Method: Developed milearn package with scikit-learn fit/predict interface, supporting both classical and neural-network-based MIL algorithms for regression/classification. Includes built-in hyperparameter optimization specifically designed for small MIL datasets.
Result: Demonstrated versatility across synthetic MIL benchmarks including digit classification/regression, molecular property prediction, and protein-protein interaction prediction. Provided dedicated support for key instance detection problem.
Conclusion: milearn provides a comprehensive, user-friendly framework for MIL that bridges classical and modern approaches with practical tools for real-world applications, especially valuable for data-scarce scenarios.
Abstract: We introduce milearn, a Python package for multi-instance learning (MIL) that follows the familiar scikit-learn fit/predict interface while providing a unified framework for both classical and neural-network-based MIL algorithms for regression and classification. The package also includes built-in hyperparameter optimization designed specifically for small MIL datasets, enabling robust model selection in data-scarce scenarios. We demonstrate the versatility of milearn across a broad range of synthetic MIL benchmark datasets, including digit classification and regression, molecular property prediction, and protein-protein interaction (PPI) prediction. Special emphasis is placed on the key instance detection (KID) problem, for which the package provides dedicated support.
[752] Intrinsic Structure as a Proxy for Saliency: SVD-Based Weight Preservation for Mixed-Precision Quantization in Large Language Models
Shashank Landge, Abhishek Patil, Tejas kamble, Bhushan Buddhivant, Priyanka Joshi
Main category: cs.LG
TL;DR: SVD-based data-free quantization method identifies important weights via principal components, outperforming calibration-dependent methods like AWQ and SpQR on privacy-sensitive tasks.
Details
Motivation: LLM deployment on commodity hardware requires quantization, but current methods need calibration data which is problematic for privacy-sensitive scenarios. There's a need for data-free quantization approaches.Method: Proposes structure-aware hypothesis: principal components from SVD indicate important weights. Preserves top-k weights aligned with principal components in FP32 while aggressively quantizing residual weights, without needing calibration data.
Result: SVD-based method achieves 66.06% accuracy on RTE task, outperforming AWQ (65.34%) and SpQR (65.34%) at high protection budgets. Shows structural importance correlates with functional importance.
Conclusion: Intrinsic matrix structure can serve as robust proxy for weight saliency without forward passes or calibration data, enabling data-free quantization for privacy-sensitive applications.
Abstract: As Large Language Models (LLMs) continue to scale in parameter count, deploying them on commodity hardware has become increasingly challenging. Post-Training Quantization (PTQ) addresses this by reducing the precision of model weights, typically to 4-bit or lower. However, uniform quantization often leads to significant performance degradation due to the presence of ``outlier features’’ – weights that, while few in number, are critical for maintaining model accuracy. Current state-of-the-art methods such as AWQ (Activation-aware Weight Quantization) and SpQR (Sparse Quantization Representations) rely on calibration data to identify these salient weights via activation magnitudes or Hessian sensitivity. In scenarios where data privacy is paramount or calibration data is unavailable, these methods are inapplicable. In this work, we propose a data-free, structure-aware hypothesis: that the weights identified as Principal Components via Singular Value Decomposition (SVD) are intrinsically important to the model’s downstream performance. We introduce a novel selection heuristic that preserves the top-$k$ weights aligned with the principal components in FP32, while aggressively quantizing the residual weights. We compare our method against activation-aware (AWQ) and second-order (SpQR) methods across GLUE benchmarks (MRPC, RTE, QNLI) using a DistilBERT backbone. Our experiments reveal that structural importance is highly correlated with functional importance. On the challenging RTE task, our SVD-based method achieves an accuracy of 66.06%, outperforming both AWQ (65.34%) and SpQR (65.34%) at high protection budgets, validating that intrinsic matrix structure can serve as a robust proxy for weight saliency without the need for forward passes or calibration data.
[753] Directed evolution algorithm drives neural prediction
Yanlin Wang, Nancy M Young, Patrick C M Wong
Main category: cs.LG
TL;DR: DEM (Directed Evolution Model) improves cross-domain neural prediction for cochlear implant outcomes by mimicking biological evolution to handle domain shift and label scarcity.
Details
Motivation: Neural prediction models struggle with domain shift and label scarcity when translating to medical AI applications, limiting their clinical utility for personalized interventions.Method: Proposed Directed Evolution Model (DEM) that mimics biological trial-and-error evolution, incorporates uncertainty exploration, replay buffer, and continual backpropagation for better exploitation-exploration trade-off.
Result: DEM improves cross-domain prediction performance for cochlear implant outcomes across four datasets, addressing domain shift and label scarcity challenges.
Conclusion: DEM offers an effective approach for medical AI applications by enhancing generalization and handling domain adaptation challenges in neural predictive modeling.
Abstract: Neural prediction offers a promising approach to forecasting the individual variability of neurocognitive functions and disorders and providing prognostic indicators for personalized invention. However, it is challenging to translate neural predictive models into medical artificial intelligent applications due to the limitations of domain shift and label scarcity. Here, we propose the directed evolution model (DEM), a novel computational model that mimics the trial-and-error processes of biological directed evolution to approximate optimal solutions for predictive modeling tasks. We demonstrated that the directed evolution algorithm is an effective strategy for uncertainty exploration, enhancing generalization in reinforcement learning. Furthermore, by incorporating replay buffer and continual backpropagate methods into DEM, we provide evidence of achieving better trade-off between exploitation and exploration in continuous learning settings. We conducted experiments on four different datasets for children with cochlear implants whose spoken language developmental outcomes vary considerably on the individual-child level. Preoperative neural MRI data has shown to accurately predict the post-operative outcome of these children within but not across datasets. Our results show that DEM can efficiently improve the performance of cross-domain pre-implantation neural predictions while addressing the challenge of label scarcity in target domain.
[754] A Fine Evaluation Method for Cube Copying Test for Early Detection of Alzheimer’s Disease
Xinyu Jiang, Cuiyun Gao, Wenda Huang, Yiyang Jiang, Binwen Luo, Yuxin Jiang, Mengting Wang, Haoran Wen, Yang Zhao, Xuemei Chen, Songqun Huang
Main category: cs.LG
TL;DR: Proposes a fine-grained evaluation method for Cube Copying Test using dynamic handwriting feature extraction and BiLSTM-Attention model to overcome binary scoring limitations in Alzheimer’s screening.
Details
Motivation: Traditional MoCA's binary "pass/fail" scoring for Cube Copying Test causes bias against less-educated elders who often score 0, leading to inaccurate assessment of visual spatial cognitive impairment in early Alzheimer's Disease.Method: Used Cogni-CareV3.0 software to collect dynamic handwriting data from CCT, extracted spatial and motion features from segmented handwriting, normalized unequal-dimension feature matrices, and applied BiLSTM-Attention model for classification.
Result: Achieved 86.69% classification accuracy, superior to similar studies. Found significant patterns in cube drawing ability across MCI patients vs healthy controls, age, and education levels. Scores negatively correlated with age, positively correlated with education.
Conclusion: Provides objective, comprehensive evaluation method for early screening and personalized intervention of visual spatial cognitive impairment, overcoming limitations of traditional binary scoring.
Abstract: Background: Impairment of visual spatial cognitive function is the most common early clinical manifestation of Alzheimer’s Disease (AD). When the Montreal Cognitive Assessment (MoCA) uses the “0/1” binary method (“pass/fail”) to evaluate the visual spatial cognitive ability represented by the Cube Copying Test(CCT), the elder with less formal education generally score 0 point, resulting in serious bias in the evaluation results. Therefore, this study proposes a fine evaluation method for CCT based on dynamic handwriting feature extraction of DH-SCSM-BLA. method : The Cogni-CareV3.0 software independently developed by our team was used to collect dynamic handwriting data of CCT. Then, the spatial and motion features of segmented dynamic handwriting were extracted, and feature matrix with unequal dimensions were normalized. Finally, a bidirectional long short-term memory network model combined with attention mechanism (BiLSTM-Attention) was adopted for classification. Result: The experimental results showed that: The proposed method has significant superiority compared to similar studies, with a classification accuracy of 86.69%. The distribution of cube drawing ability scores has significant regularity for three aspects such as MCI patients and healthy control group, age, and levels of education. It was also found that score for each cognitive task including cube drawing ability score is negatively correlated with age. Score for each cognitive task including cube drawing ability score, but positively correlated with levels of education significantly. Conclusion: This study provides a relatively objective and comprehensive evaluation method for early screening and personalized intervention of visual spatial cognitive impairment.
[755] Beyond Loss Guidance: Using PDE Residuals as Spectral Attention in Diffusion Neural Operators
Medha Sawhney, Abhilash Neog, Mridul Khurana, Anuj Karpatne
Main category: cs.LG
TL;DR: PRISMA is a conditional diffusion neural operator that embeds PDE residuals directly into model architecture via spectral attention, enabling gradient-free inference with competitive accuracy and much faster inference times.
Details
Motivation: Current diffusion-based PDE solvers suffer from slow gradient-based optimization at test time, optimization instabilities, and inability to adapt to noisy PDE residuals, requiring sensitive hyperparameter tuning.Method: PRISMA integrates PDE residuals as architectural features via attention mechanisms in the spectral domain, creating a conditional diffusion neural operator that avoids gradient-based optimization during inference.
Result: PRISMA achieves competitive accuracy with substantially lower inference costs (15x to 250x faster) across five benchmark PDEs, especially with noisy observations, using 10x to 100x fewer denoising steps.
Conclusion: PRISMA addresses key limitations of previous methods by making PDE residuals integral to model architecture, resulting in faster, more robust, and hyperparameter-insensitive PDE solving without gradient optimization.
Abstract: Diffusion-based solvers for partial differential equations (PDEs) are often bottle-necked by slow gradient-based test-time optimization routines that use PDE residuals for loss guidance. They additionally suffer from optimization instabilities and are unable to dynamically adapt their inference scheme in the presence of noisy PDE residuals. To address these limitations, we introduce PRISMA (PDE Residual Informed Spectral Modulation with Attention), a conditional diffusion neural operator that embeds PDE residuals directly into the model’s architecture via attention mechanisms in the spectral domain, enabling gradient-descent free inference. In contrast to previous methods that use PDE loss solely as external optimization targets, PRISMA integrates PDE residuals as integral architectural features, making it inherently fast, robust, accurate, and free from sensitive hyperparameter tuning. We show that PRISMA has competitive accuracy, at substantially lower inference costs, compared to previous methods across five benchmark PDEs, especially with noisy observations, while using 10x to 100x fewer denoising steps, leading to 15x to 250x faster inference.
[756] CLAPS: Posterior-Aware Conformal Intervals via Last-Layer Laplace
Dongseok Kim, Hyoungsun Choi, Mohamed Jismy Aashik Rasool, Gisung Oh
Main category: cs.LG
TL;DR: CLAPS combines Last-Layer Laplace Approximation with split-conformal calibration to create narrower prediction intervals with the same coverage, especially effective on small tabular datasets.
Details
Motivation: Standard conformal methods often rely on point estimates (residuals) rather than full predictive distributions, leading to wider-than-necessary prediction intervals. This is particularly problematic on small to medium tabular datasets where data scarcity makes uncertainty modeling crucial.Method: CLAPS pairs Last-Layer Laplace Approximation (to obtain a Gaussian posterior) with split-conformal calibration. It defines a two-sided posterior CDF score that aligns the conformity metric with the full predictive shape, not just a point estimate.
Result: CLAPS consistently achieves nominal coverage with improved efficiency (narrower intervals) and minimal computational overhead across multiple benchmarks using the same MLP backbone. It performs particularly well on small to medium tabular datasets.
Conclusion: CLAPS offers a practical upgrade to residual-based conformal baselines by better leveraging posterior information, providing narrower prediction intervals while maintaining coverage, along with diagnostic tools to understand uncertainty decomposition.
Abstract: We present CLAPS, a posterior-aware conformal regression method that pairs a Last-Layer Laplace Approximation with split-conformal calibration. From the resulting Gaussian posterior, CLAPS defines a simple two-sided posterior CDF score that aligns the conformity metric with the full predictive shape, not just a point estimate. This alignment yields narrower prediction intervals at the same target coverage, especially on small to medium tabular datasets where data are scarce and uncertainty modeling matters. We also provide a lightweight diagnostic suite that separates aleatoric and epistemic components and visualizes posterior behavior, helping practitioners understand why intervals shrink when they do. Across multiple benchmarks using the same MLP backbone, CLAPS consistently attains nominal coverage with improved efficiency and minimal overhead, offering a clear, practical upgrade to residual-based conformal baselines.
[757] Consistency Flow Model Achieves One-step Denoising Error Correction Codes
Haoyu Lei, Chin Wa Lau, Kaiwen Zhou, Nian Guo, Farzan Farnia
Main category: cs.LG
TL;DR: ECCFM is a one-step neural decoder framework that uses probability flow ODEs to achieve faster inference (30-100x speedup) than diffusion models while maintaining or improving bit-error rates.
Details
Motivation: Current neural decoders face a trade-off between accuracy and computational efficiency. Diffusion-based decoders achieve state-of-the-art performance but are too slow for low-latency applications due to their iterative sampling process.Method: ECCFM transforms the reverse denoising process into a Probability Flow Ordinary Differential Equation (PF-ODE) and enforces smoothness through differential time regularization. This allows learning to map noisy signals directly to original codewords in a single inference step.
Result: ECCFM achieves lower bit-error rates than autoregressive and diffusion-based baselines across multiple decoding benchmarks, with particularly notable improvements on longer codes. It delivers 30-100x faster inference speeds compared to denoising diffusion decoders.
Conclusion: ECCFM provides an architecture-agnostic training framework that enables high-fidelity one-step decoding, effectively bridging the gap between accuracy and computational efficiency for neural error correction decoders.
Abstract: Error Correction Codes (ECC) are fundamental to reliable digital communication, yet designing neural decoders that are both accurate and computationally efficient remains challenging. Recent denoising diffusion decoders with transformer backbones achieve state-of-the-art performance, but their iterative sampling limits practicality in low-latency settings. We introduce the Error Correction Consistency Flow Model (ECCFM), an architecture-agnostic training framework for high-fidelity one-step decoding. By casting the reverse denoising process as a Probability Flow Ordinary Differential Equation (PF-ODE) and enforcing smoothness through a differential time regularization, ECCFM learns to map noisy signals along the decoding trajectory directly to the original codeword in a single inference step. Across multiple decoding benchmarks, ECCFM attains lower bit-error rates (BER) than autoregressive and diffusion-based baselines, with notable improvements on longer codes, while delivering inference speeds up from 30x to 100x faster than denoising diffusion decoders.
[758] RE-LLM: Integrating Large Language Models into Renewable Energy Systems
Ali Forootani, Mohammad Sadr, Danial Esmaeili Aliabadi, Daniela Thraen
Main category: cs.LG
TL;DR: RE-LLM integrates LLMs into energy system modeling to translate complex technical outputs into accessible natural language explanations for non-expert stakeholders.
Details
Motivation: Energy system models produce highly technical outputs that are difficult for non-experts (policymakers, planners, public) to interpret, creating a communication gap that limits the practical impact of scenario-based modeling in complex energy transitions.Method: Proposes RE-LLM, a hybrid framework combining: (i) optimization-based scenario exploration, (ii) machine learning surrogates to accelerate simulations, and (iii) LLM-powered natural language generation to translate complex results into stakeholder-oriented explanations.
Result: The framework reduces computational burden while enhancing interpretability, enabling real-time reasoning about trade-offs, sensitivities, and policy implications. It’s adaptable across different optimization platforms and energy system models.
Conclusion: RE-LLM advances human-centric energy modeling by enabling interactive, multilingual, accessible engagement with future energy pathways, bridging the gap between data-driven analysis and actionable decision-making for sustainable transitions.
Abstract: Energy system models are increasingly employed to guide long-term planning in multi-sectoral environments where decisions span electricity, heat, transport, land use, and industry. While these models provide rigorous quantitative insights, their outputs are often highly technical, making them difficult to interpret for non-expert stakeholders such as policymakers, planners, and the public. This communication gap limits the accessibility and practical impact of scenario-based modeling, particularly as energy transitions grow more complex with rising shares of renewables, sectoral integration, and deep uncertainties. To address this challenge, we propose the Renewable Energy Large Language Model (RE-LLM), a hybrid framework that integrates Large Language Models (LLMs) directly into the energy system modeling workflow. RE-LLM combines three core elements: (i) optimization-based scenario exploration, (ii) machine learning surrogates that accelerate computationally intensive simulations, and (iii) LLM-powered natural language generation that translates complex results into clear, stakeholder-oriented explanations. This integrated design not only reduces computational burden but also enhances inter-pretability, enabling real-time reasoning about trade-offs, sensitivities, and policy implications. The framework is adaptable across different optimization platforms and energy system models, ensuring broad applicability beyond the case study presented. By merging speed, rigor, and interpretability, RE-LLM advances a new paradigm of human-centric energy modeling. It enables interactive, multilingual, and accessible engagement with future energy pathways, ultimately bridging the final gap between data-driven analysis and actionable decision-making for sustainable transitions.
[759] On Global Applicability and Location Transferability of Generative Deep Learning Models for Precipitation Downscaling
Paula Harder, Christian Lessig, Matthew Chantry, Francis Pelletier, David Rolnick
Main category: cs.LG
TL;DR: Evaluates generalization performance of generative deep learning models for precipitation downscaling across diverse geographic regions using global ERA5 and IMERG data with location-based splits.
Details
Motivation: Most existing generative downscaling models are region-specific, and their ability to generalize to unseen geographic areas remains largely unexplored, creating a gap in understanding model transferability.Method: Uses global ERA5 reanalysis data as predictors and IMERG precipitation estimates at 0.1° resolution as targets. Implements hierarchical location-based data split across 15 global regions to systematically assess model generalization.
Result: The study evaluates model performance across diverse regions, though specific quantitative results are not provided in the abstract.
Conclusion: Systematic evaluation of generative downscaling models across multiple regions provides insights into their generalization capabilities, addressing the gap in understanding model transferability to unseen geographic areas.
Abstract: Deep learning offers promising capabilities for the statistical downscaling of climate and weather forecasts, with generative approaches showing particular success in capturing fine-scale precipitation patterns. However, most existing models are region-specific, and their ability to generalize to unseen geographic areas remains largely unexplored. In this study, we evaluate the generalization performance of generative downscaling models across diverse regions. Using a global framework, we employ ERA5 reanalysis data as predictors and IMERG precipitation estimates at $0.1^\circ$ resolution as targets. A hierarchical location-based data split enables a systematic assessment of model performance across 15 regions around the world.
[760] Fantastic Features and Where to Find Them: A Probing Method to combine Features from Multiple Foundation Models
Benjamin Ramtoula, Pierre-Yves Lajoie, Paul Newman, Daniele De Martini
Main category: cs.LG
TL;DR: ComBo is a scalable probing-based adapter that combines features from multiple foundation models without dataset-specific tuning or backpropagation, outperforming previous methods on VTAB-1k benchmark.
Details
Motivation: Different foundation models have complementary strengths for specific tasks, but current adaptation strategies focus on individual models and don't exploit these complementary representations. Probing methods exist but don't scale well with large feature sets and require dataset-specific tuning.Method: ComBo compresses activations from layers of one or more foundation models into compact token-wise representations and processes them with a lightweight transformer for task-specific prediction. It includes a mechanism to evaluate each backbone’s task-relevance through joint multi-backbone probing.
Result: On 19 tasks of the VTAB-1k benchmark, ComBo outperforms previous probing methods, matches or surpasses more expensive alternatives like distillation-based model merging, and enables efficient probing of tuned models.
Conclusion: ComBo provides a practical and general-purpose framework for combining diverse representations from multiple foundation models, offering an effective way to leverage complementary strengths without expensive tuning or backpropagation.
Abstract: Foundation models (FMs) trained with different objectives and data learn diverse representations, making some more effective than others for specific downstream tasks. Existing adaptation strategies, such as parameter-efficient fine-tuning, focus on individual models and do not exploit the complementary strengths across models. Probing methods offer a promising alternative by extracting information from frozen models, but current techniques do not scale well with large feature sets and often rely on dataset-specific hyperparameter tuning. We propose Combined backBones (ComBo), a simple and scalable probing-based adapter that effectively integrates features from multiple models and layers. ComBo compresses activations from layers of one or more FMs into compact token-wise representations and processes them with a lightweight transformer for task-specific prediction. Crucially, ComBo does not require dataset-specific tuning or backpropagation through the backbone models. However, not all models are equally relevant for all tasks. To address this, we introduce a mechanism that leverages ComBo’s joint multi-backbone probing to efficiently evaluate each backbone’s task-relevance, enabling both practical model comparison and improved performance through selective adaptation. On the 19 tasks of the VTAB-1k benchmark, ComBo outperforms previous probing methods, matches or surpasses more expensive alternatives, such as distillation-based model merging, and enables efficient probing of tuned models. Our results demonstrate that ComBo offers a practical and general-purpose framework for combining diverse representations from multiple FMs.
[761] A Self-explainable Model of Long Time Series by Extracting Informative Structured Causal Patterns
Ziqian Wang, Yuxiao Cheng, Jinli Suo
Main category: cs.LG
TL;DR: EXCAP is a unified framework for interpretable time-series modeling that produces coherent temporal explanations with causal grounding, addressing limitations of point-wise importance scores in existing methods.
Details
Motivation: Existing explainable AI methods for neural networks modeling long time series only produce point-wise importance scores, failing to capture temporal structures like trends, cycles, and regime changes. This limitation weakens human interpretability and trust in long-horizon models, especially in high-stakes domains like healthcare and finance.Method: EXCAP combines three key components: 1) an attention-based segmenter that extracts coherent temporal patterns, 2) a causally structured decoder guided by a pre-trained causal graph, and 3) a latent aggregation mechanism that enforces representation stability. The framework is designed to satisfy four requirements: temporal continuity, pattern-centric explanation, causal disentanglement, and faithfulness to the model’s inference process.
Result: Extensive experiments on classification and forecasting benchmarks demonstrate that EXCAP achieves strong predictive accuracy while generating coherent and causally grounded explanations. Theoretical analysis shows EXCAP provides smooth and stable explanations over time and is robust to perturbations in causal masks.
Conclusion: EXCAP offers a principled and scalable approach to interpretable modeling of long time series, addressing the need for explanations that capture temporal structures and causal relationships, with particular relevance to high-stakes domains requiring trustworthy AI systems.
Abstract: Explainability is essential for neural networks that model long time series, yet most existing explainable AI methods only produce point-wise importance scores and fail to capture temporal structures such as trends, cycles, and regime changes. This limitation weakens human interpretability and trust in long-horizon models. To address these issues, we identify four key requirements for interpretable time-series modeling: temporal continuity, pattern-centric explanation, causal disentanglement, and faithfulness to the model’s inference process. We propose EXCAP, a unified framework that satisfies all four requirements. EXCAP combines an attention-based segmenter that extracts coherent temporal patterns, a causally structured decoder guided by a pre-trained causal graph, and a latent aggregation mechanism that enforces representation stability. Our theoretical analysis shows that EXCAP provides smooth and stable explanations over time and is robust to perturbations in causal masks. Extensive experiments on classification and forecasting benchmarks demonstrate that EXCAP achieves strong predictive accuracy while generating coherent and causally grounded explanations. These results show that EXCAP offers a principled and scalable approach to interpretable modeling of long time series with relevance to high-stakes domains such as healthcare and finance.
[762] Fourier Neural Operators Explained: A Practical Perspective
Valentin Duruisseaux, Jean Kossaifi, Anima Anandkumar
Main category: cs.LG
TL;DR: A comprehensive practice-oriented guide to Fourier Neural Operators (FNOs) that bridges theoretical foundations with practical implementation, addressing common misunderstandings and providing modular implementations through the NeuralOperator 2.0.0 library.
Details
Motivation: While FNOs have become influential for learning PDE solutions directly from data, their practical use is hindered by incomplete understanding of theoretical foundations, practical constraints, and implementation details among practitioners, leading to incorrect or unreliable applications.Method: The guide provides intuitive exposition of operator theory and signal-processing concepts underlying FNOs, details spectral parameterization and computational design of all components, addresses common misunderstandings, and integrates closely with the NeuralOperator 2.0.0 library for modular state-of-the-art implementations.
Result: A clear and reliable framework for applying FNOs effectively across diverse scientific and engineering fields by connecting rigorous mathematical foundations with practical implementation insights and addressing literature misunderstandings.
Conclusion: This comprehensive guide establishes a practice-oriented approach to FNOs that unifies mathematical principles with implementation strategies, enabling more effective and reliable application of these powerful neural operators for PDE surrogate modeling.
Abstract: Partial differential equations (PDEs) govern a wide variety of dynamical processes in science and engineering, yet obtaining their numerical solutions often requires high-resolution discretizations and repeated evaluations of complex operators, leading to substantial computational costs. Neural operators have recently emerged as a powerful framework for learning mappings between function spaces directly from data, enabling efficient surrogate models for PDE systems. Among these architectures, the Fourier Neural Operator (FNO) has become the most influential and widely adopted due to its elegant spectral formulation, which captures global correlations through learnable transformations in Fourier space while remaining invariant to discretization and resolution. Despite their success, the practical use of FNOs is often hindered by an incomplete understanding among practitioners of their theoretical foundations, practical constraints, and implementation details, which can lead to their incorrect or unreliable application. This work presents a comprehensive and practice-oriented guide to FNOs, unifying their mathematical principles with implementation strategies. We provide an intuitive exposition to the concepts of operator theory and signal-processing that underlie the FNO, detail its spectral parameterization and the computational design of all its components, and address common misunderstandings encountered in the literature. The exposition is closely integrated with the NeuralOperator 2.0.0 library, offering modular state-of-the-art implementations that faithfully reflect the theory. By connecting rigorous foundations with practical insight, this guide aims to establish a clear and reliable framework for applying FNOs effectively across diverse scientific and engineering fields.
[763] Stay Unique, Stay Efficient: Preserving Model Personality in Multi-Task Merging
Kuangpu Guo, Yuhe Ding, Jian Liang, Zilei Wang, Ran He
Main category: cs.LG
TL;DR: DTS is a model merging framework that preserves task-specific information using singular value decomposition, thresholding, and scaling, requiring only 1% additional storage per task while outperforming SOTA methods.
Details
Motivation: Existing model merging methods suffer from substantial performance degradation compared to individually fine-tuned models, even on similar tasks, highlighting the need to preserve task-specific information during merging.Method: DTS uses singular value decomposition to extract task-specific information, retains only a small subset of singular values/vectors, applies novel thresholding that partitions singular vector elements into groups with scaling factors, and includes a data-free variant for unseen tasks based on semantic similarity.
Result: DTS consistently outperforms state-of-the-art baselines while requiring only 1% additional storage per task. The variant for unseen tasks achieves significantly better generalization performance.
Conclusion: DTS provides an effective personalized merging framework that preserves task-specific information with minimal storage overhead and demonstrates strong generalization capabilities to unseen tasks.
Abstract: Model merging has emerged as a promising paradigm for enabling multi-task capabilities without additional training. However, existing methods often experience substantial performance degradation compared with individually fine-tuned models, even on similar tasks, underscoring the need to preserve task-specific information. This paper proposes Decomposition, Thresholding, and Scaling (DTS), an approximation-based personalized merging framework that preserves task-specific information with minimal storage overhead. DTS first applies singular value decomposition to the task-specific information and retains only a small subset of singular values and vectors. It then introduces a novel thresholding strategy that partitions singular vector elements into groups and assigns a scaling factor to each group. To enable generalization to unseen tasks, we further extend DTS with a variant that fuses task-specific information in a data-free manner based on the semantic similarity of task characteristics. Extensive experiments demonstrate that DTS consistently outperforms state-of-the-art baselines while requiring only 1% additional storage per task. Furthermore, experiments on unseen tasks show that the DTS variant achieves significantly better generalization performance. Our code is available at https://github.com/krumpguo/DTS.
[764] A Nonlinear Low-rank Representation Model with Convolutional Neural Network for Imputing Water Quality Data
Hongnan Si, Tong Li, Yujie Chen, Xin Liao
Main category: cs.LG
TL;DR: Proposes NTCN model for water quality data imputation using Tucker decomposition and 3D convolution to handle missing data in monitoring systems.
Details
Motivation: Water quality monitoring suffers from data missing due to sensor failures, creating challenges for analysis and environmental protection.Method: Neural Tucker Convolutional Network (NTCN) encodes different mode entities into embeddings, constructs Tucker interaction tensor via outer product, and uses 3D convolution for spatiotemporal feature extraction.
Result: NTCN outperforms state-of-the-art imputation models on three real-world water quality datasets in terms of accuracy.
Conclusion: The proposed NTCN model effectively addresses water quality data imputation by capturing complex mode-wise interactions and spatiotemporal patterns.
Abstract: Water quality monitoring is a core component of ecological environmental protection. However, due to sensor failure or other inevitable factors, data missing often exists in long-term monitoring, posing great challenges in water quality analysis. This paper proposes a Neural Tucker Convolutional Network (NTCN) model for water quality data imputation, which features the following key components: a) Encode different mode entities into respective embedding vectors, and construct a Tucker interaction tensor by outer product operations to capture the complex mode-wise feature interactions; b) Use 3D convolution to extract fine-grained spatiotemporal features from the interaction tensor. Experiments on three real-world water quality datasets show that the proposed NTCN model outperforms several state-of-the-art imputation models in terms of accuracy.
[765] Differentiable Weightless Controllers: Learning Logic Circuits for Continuous Control
Fabian Kresse, Christoph H. Lampert
Main category: cs.LG
TL;DR: DWCs represent continuous-control policies as discrete logic circuits instead of neural networks, achieving competitive performance while enabling FPGA deployment with ultra-low latency and energy consumption.
Details
Motivation: To create control policies that can be compiled directly into hardware circuits with minimal latency and energy consumption, while maintaining competitive performance with neural network approaches.Method: Differentiable Weightless Controllers (DWCs) use thermometer-encoded inputs, sparsely connected boolean lookup-table layers, and lightweight action heads. They can be trained end-to-end with gradient-based methods but compile to FPGA-compatible circuits.
Result: DWCs achieve returns competitive with weight-based policies on five MuJoCo benchmarks, matching performance on four tasks and identifying network capacity as the limiting factor on HalfCheetah. They enable ultra-low latency (single-clock-cycle) and nanojoule-level energy consumption.
Conclusion: Continuous-control policies can be effectively represented as discrete logic circuits, offering hardware-friendly implementations with competitive performance, interpretable connectivity patterns, and extreme efficiency for real-time control applications.
Abstract: We investigate whether continuous-control policies can be represented and learned as discrete logic circuits instead of continuous neural networks. We introduce Differentiable Weightless Controllers (DWCs), a symbolic-differentiable architecture that maps real-valued observations to actions using thermometer-encoded inputs, sparsely connected boolean lookup-table layers, and lightweight action heads. DWCs can be trained end-to-end by gradient-based techniques, yet compile directly into FPGA-compatible circuits with few- or even single-clock-cycle latency and nanojoule-level energy cost per action. Across five MuJoCo benchmarks, including high-dimensional Humanoid, DWCs achieve returns competitive with weight-based policies (full precision or quantized neural networks), matching performance on four tasks and isolating network capacity as the key limiting factor on HalfCheetah. Furthermore, DWCs exhibit structurally sparse and interpretable connectivity patterns, enabling a direct inspection of which input thresholds influence control decisions.
[766] Does Flatness imply Generalization for Logistic Loss in Univariate Two-Layer ReLU Network?
Dan Qiao, Yu-Xiang Wang
Main category: cs.LG
TL;DR: Flatness alone doesn’t guarantee generalization for logistic loss in overparameterized ReLU networks - flat solutions can still overfit, unlike with square loss.
Details
Motivation: Previous work showed flat solutions under square loss provably cannot overfit, but it's unclear if this holds for logistic loss, especially since gradient descent with increasing step size converges to interpolating solutions at infinity for margin-separable cases.Method: Theoretical analysis of flat solutions in overparameterized two-layer ReLU networks with univariate input under logistic loss, examining generalization within regions between left-most and right-most uncertain sets, and constructing counterexamples of flat yet overfitting solutions.
Result: Flat solutions enjoy near-optimal generalization bounds within uncertain regions, but there exist arbitrarily flat solutions at infinity that are falsely certain everywhere and overfit, showing flatness alone is insufficient for generalization under logistic loss.
Conclusion: Flatness-implied generalization is more delicate under logistic loss than square loss - while flat solutions can generalize well within certain regions, flatness alone doesn’t guarantee generalization, as demonstrated by constructed overfitting counterexamples.
Abstract: We consider the problem of generalization of arbitrarily overparameterized two-layer ReLU Neural Networks with univariate input. Recent work showed that under square loss, flat solutions (motivated by flat / stable minima and Edge of Stability phenomenon) provably cannot overfit, but it remains unclear whether the same phenomenon holds for logistic loss. This is a puzzling open problem because existing work on logistic loss shows that gradient descent with increasing step size converges to interpolating solutions (at infinity, for the margin-separable cases). In this paper, we prove that the \emph{flatness implied generalization} is more delicate under logistic loss. On the positive side, we show that flat solutions enjoy near-optimal generalization bounds within a region between the left-most and right-most \emph{uncertain} sets determined by each candidate solution. On the negative side, we show that there exist arbitrarily flat yet overfitting solutions at infinity that are (falsely) certain everywhere, thus certifying that flatness alone is insufficient for generalization in general. We demonstrate the effects predicted by our theory in a well-controlled simulation study.
[767] Multi-view diffusion geometry using intertwined diffusion trajectories
Gwendal Debaussart-Joniec, Argyris Kalogeratos
Main category: cs.LG
TL;DR: A unified framework for multi-view diffusion geometries using intertwined diffusion trajectories that combine multiple data views, with theoretical guarantees and practical applications in manifold learning and clustering.
Details
Motivation: Existing multi-view diffusion models lack a comprehensive unified framework that captures the dynamic interplay between different data views over time. The paper aims to create a flexible framework that encompasses existing models while providing new degrees of freedom for view interaction and fusion.Method: Proposes multi-view diffusion trajectories (MDTs) - inhomogeneous diffusion processes that iteratively combine random walk operators from multiple data views. Each MDT defines a trajectory-dependent diffusion operator with probabilistic and geometric interpretation. The framework includes theoretical analysis, derivation of MDT-based diffusion distances and embeddings via SVD, and learning strategies guided by internal quality measures.
Result: Establishes theoretical properties including ergodicity of both point-wise operators and the process itself. MDTs provide a flexible framework that encompasses existing models while offering new view interaction capabilities. The approach also serves as a neutral baseline for evaluating diffusion-based methods. Experimental results demonstrate practical impact in manifold learning and data clustering contexts.
Conclusion: MDTs offer a comprehensive unified framework for multi-view diffusion geometries with theoretical guarantees and practical utility. The framework enables flexible model design, provides new degrees of freedom for view interaction, and serves as a valuable baseline for evaluating diffusion-based approaches in multi-view learning tasks.
Abstract: This paper introduces a comprehensive unified framework for constructing multi-view diffusion geometries through intertwined multi-view diffusion trajectories (MDTs), a class of inhomogeneous diffusion processes that iteratively combine the random walk operators of multiple data views. Each MDT defines a trajectory-dependent diffusion operator with a clear probabilistic and geometric interpretation, capturing over time the interplay between data views. Our formulation encompasses existing multi-view diffusion models, while providing new degrees of freedom for view interaction and fusion. We establish theoretical properties under mild assumptions, including ergodicity of both the point-wise operator and the process in itself. We also derive MDT-based diffusion distances, and associated embeddings via singular value decompositions. Finally, we propose various strategies for learning MDT operators within the defined operator space, guided by internal quality measures. Beyond enabling flexible model design, MDTs also offer a neutral baseline for evaluating diffusion-based approaches through comparison with randomly selected MDTs. Experiments show the practical impact of the MDT operators in a manifold learning and data clustering context.
[768] Winning Solutions for the Rayan AI Contest: Compositional Retrieval, Zero-Shot Anomaly Detection, and Backdoor Detection
Ali Nafisi, Sina Asghari, Mohammad Saeed Arvenaghi, Hossein Shakibania
Main category: cs.LG
TL;DR: The paper presents solutions to three ML challenges: compositional image retrieval (95.38% accuracy, 1st place), zero-shot anomaly detection (73.14% accuracy, 1st place), and backdoored model detection (78% accuracy, 2nd place).
Details
Motivation: To address key machine learning challenges in retrieval, anomaly detection, and model security with practical applications in healthcare, manufacturing, and cybersecurity.Method: Developed specialized systems for each task: 1) A compositional image retrieval system processing visual and textual inputs, 2) A zero-shot anomaly detection model identifying anomalies without prior exposure to abnormal examples, 3) A method to detect hidden backdoor triggers in neural networks.
Result: Achieved state-of-the-art results: 95.38% accuracy in compositional image retrieval (1st place), 73.14% accuracy in zero-shot anomaly detection (1st place), and 78% accuracy in backdoored model detection (2nd place).
Conclusion: The methods effectively address key ML challenges with strong performance, demonstrating practical value for real-world applications. All code is publicly available.
Abstract: This report presents solutions to three machine learning challenges: compositional image retrieval, zero-shot anomaly detection, and backdoored model detection. In compositional image retrieval, we developed a system that processes visual and textual inputs to retrieve relevant images, achieving 95.38% accuracy and ranking first with a clear margin over the second team. For zero-shot anomaly detection, we designed a model that identifies and localizes anomalies in images without prior exposure to abnormal examples, securing 1st place with 73.14% accuracy. In the backdoored model detection task, we proposed a method to detect hidden backdoor triggers in neural networks, reaching an accuracy of 78%, which placed our approach in second place. These results demonstrate the effectiveness of our methods in addressing key challenges related to retrieval, anomaly detection, and model security, with implications for real-world applications in industries such as healthcare, manufacturing, and cybersecurity. Code for all solutions is available online.
[769] Walking on the Fiber: A Simple Geometric Approximation for Bayesian Neural Networks
Alfredo Reichlin, Miguel Vasco, Danica Kragic
Main category: cs.LG
TL;DR: A novel Bayesian neural network approach using parameter space deformation for efficient posterior sampling without iterative methods.
Details
Motivation: Exact posterior inference in Bayesian neural networks is computationally intractable, and existing approximations like Laplace method have scalability and accuracy limitations in modern deep networks.Method: Proposes a simple sampling variation leveraging low-dimensional structure of loss minima in over-parameterized networks, then introduces a model that learns a deformation of parameter space for rapid posterior sampling without iterative methods.
Result: Empirical results show competitive posterior approximations with improved scalability compared to recent refinement techniques.
Conclusion: Provides a practical alternative for Bayesian inference in deep learning through efficient posterior sampling approach.
Abstract: Bayesian Neural Networks provide a principled framework for uncertainty quantification by modeling the posterior distribution of network parameters. However, exact posterior inference is computationally intractable, and widely used approximations like the Laplace method struggle with scalability and posterior accuracy in modern deep networks. In this work, we revisit sampling techniques for posterior exploration, proposing a simple variation tailored to efficiently sample from the posterior in over-parameterized networks by leveraging the low-dimensional structure of loss minima. Building on this, we introduce a model that learns a deformation of the parameter space, enabling rapid posterior sampling without requiring iterative methods. Empirical results demonstrate that our approach achieves competitive posterior approximations with improved scalability compared to recent refinement techniques. These contributions provide a practical alternative for Bayesian inference in deep learning.
[770] Label Forensics: Interpreting Hard Labels in Black-Box Text Classifier
Mengyao Du, Gang Yang, Han Fang, Quanjun Yin, Ee-chien Chang
Main category: cs.LG
TL;DR: A black-box framework called “label forensics” that reconstructs the semantic meaning of labels in undocumented text classifiers by creating sentence embedding distributions that reliably reflect what the classifier has learned for each label.
Details
Motivation: Many text classifiers circulate with undocumented internal semantics, operating in unregulated environments or being repurposed for unknown intents, raising forensic and auditing concerns about their behavior and decision-making.Method: Introduces label forensics framework that represents each label by a sentence embedding distribution. Uses semantic neighborhood sampler and iterative optimization to select representative seed sentences that maximize both label consistency (samples consistently classified into target label) and distributional coverage (covering label’s broad semantic space).
Result: Achieves average label consistency of around 92.24% across multiple black-box classifiers, demonstrating that the embedding regions accurately capture each classifier’s label semantics. Successfully validated on undocumented HuggingFace classifier, enabling fine-grained label interpretation.
Conclusion: The framework effectively reconstructs semantic meanings of labels in black-box classifiers, supporting responsible AI auditing and forensic analysis of undocumented models operating in web environments.
Abstract: The widespread adoption of natural language processing techniques has led to an unprecedented growth of text classifiers across the modern web. Yet many of these models circulate with their internal semantics undocumented or even intentionally withheld. Such opaque classifiers, which may expose only hard-label outputs, can operate in unregulated web environments or be repurposed for unknown intents, raising legitimate forensic and auditing concerns. In this paper, we position ourselves as investigators and work to infer the semantic concept each label encodes in an undocumented black-box classifier. Specifically, we introduce label forensics, a black-box framework that reconstructs a label’s semantic meaning. Concretely, we represent a label by a sentence embedding distribution from which any sample reliably reflects the concept the classifier has implicitly learned for that label. We believe this distribution should maintain two key properties: precise, with samples consistently classified into the target label, and general, covering the label’s broad semantic space. To realize this, we design a semantic neighborhood sampler and an iterative optimization procedure to select representative seed sentences that jointly maximize label consistency and distributional coverage. The final output, an optimized seed sentence set combined with the sampler, constitutes the empirical distribution representing the label’s semantics. Experiments on multiple black-box classifiers achieve an average label consistency of around 92.24 percent, demonstrating that the embedding regions accurately capture each classifier’s label semantics. We further validate our framework on an undocumented HuggingFace classifier, enabling fine-grained label interpretation and supporting responsible AI auditing.
[771] End-to-end Deep Reinforcement Learning for Stochastic Multi-objective Optimization in C-VRPTW
Abdo Abouelrous, Laurens Bliek, Yaoxin Wu, Yingqian Zhang
Main category: cs.LG
TL;DR: Proposes an end-to-end deep learning model for stochastic multi-objective vehicle routing problems that handles travel time uncertainty and constructs Pareto fronts efficiently.
Details
Motivation: Practical routing problems involve uncertainty (like travel times) and multiple conflicting objectives (efficiency vs labor regulations), but existing learning-based methods struggle to handle both stochasticity and multi-objectivity simultaneously.Method: End-to-end deep learning model using attention mechanism and multiple solution trajectories, with a refined training mechanism using scenario clustering to reduce training time. The model simultaneously addresses stochasticity and multi-objectivity.
Result: The model constructs Pareto fronts of good quality within acceptable run times, outperforming three baseline methods.
Conclusion: The proposed approach successfully addresses the challenges of stochastic multi-objective vehicle routing problems, offering computational advantages for practical applications with uncertainty and multiple objectives.
Abstract: In this work, we consider learning-based applications in routing to solve a Vehicle Routing variant characterized by stochasticity and multiple objectives. Such problems are representative of practical settings where decision-makers have to deal with uncertainty in the operational environment as well as multiple conflicting objectives due to different stakeholders. We specifically consider travel time uncertainty. We also consider two objectives, total travel time and route makespan, that jointly target operational efficiency and labor regulations on shift length, although different objectives could be incorporated. Learning-based methods offer earnest computational advantages as they can repeatedly solve problems with limited interference from the decision-maker. We specifically focus on end-to-end deep learning models that leverage the attention mechanism and multiple solution trajectories. These models have seen several successful applications in routing problems. However, since travel times are not a direct input to these models due to the large dimensions of the travel time matrix, accounting for uncertainty is a challenge, especially in the presence of multiple objectives. In turn, we propose a model that simultaneously addresses stochasticity and multi-objectivity and provide a refined training mechanism for this model through scenario clustering to reduce training time. Our results show that our model is capable of constructing a Pareto Front of good quality within acceptable run times compared to three baselines.
[772] TimePred: efficient and interpretable offline change point detection for high volume data - with application to industrial process monitoring
Simon Leszek
Main category: cs.LG
TL;DR: TimePred is a self-supervised framework that converts multivariate change-point detection to univariate mean-shift detection by predicting normalized time indices, enabling efficient offline detection with explainable AI integration.
Details
Motivation: Change-point detection in high-dimensional, large-volume time series faces challenges in statistical consistency, scalability, and interpretability. Existing methods struggle with these combined requirements.Method: TimePred reduces multivariate CPD to univariate mean-shift detection by predicting each sample’s normalized time index using self-supervised learning. This enables efficient offline detection with existing algorithms and supports XAI attribution methods for feature-level explanations.
Result: Experiments show competitive CPD performance while reducing computational cost by up to two orders of magnitude. In an industrial manufacturing case study, the method demonstrates improved detection accuracy and provides interpretable change-point insights.
Conclusion: TimePred offers an efficient, scalable, and interpretable solution for high-dimensional change-point detection, with practical value demonstrated in real-world industrial applications through both performance improvements and explainable insights.
Abstract: Change-point detection (CPD) in high-dimensional, large-volume time series is challenging for statistical consistency, scalability, and interpretability. We introduce TimePred, a self-supervised framework that reduces multivariate CPD to univariate mean-shift detection by predicting each sample’s normalized time index. This enables efficient offline CPD using existing algorithms and supports the integration of XAI attribution methods for feature-level explanations. Our experiments show competitive CPD performance while reducing computational cost by up to two orders of magnitude. In an industrial manufacturing case study, we demonstrate improved detection accuracy and illustrate the practical value of interpretable change-point insights.
[773] Do Large Language Models Walk Their Talk? Measuring the Gap Between Implicit Associations, Self-Report, and Behavioral Altruism
Sandro Andric
Main category: cs.LG
TL;DR: LLMs show strong implicit pro-altruism bias and behave altruistically above chance, but their implicit associations don’t predict behavior, and they systematically overestimate their own altruism, creating a “virtue signaling gap.”
Details
Motivation: To investigate whether LLMs exhibit altruistic tendencies and whether their implicit associations and self-reports predict actual altruistic behavior, using methods from human social psychology.Method: Multi-method approach testing 24 frontier LLMs across three paradigms: Implicit Association Test (IAT) for implicit altruism bias, forced binary choice task for behavioral altruism, and self-assessment scale for explicit altruism beliefs.
Result: 1) All models show strong implicit pro-altruism bias. 2) Models behave altruistically above chance (65.6% vs. 50%) with substantial variation. 3) Implicit associations don’t predict behavior. 4) Models systematically overestimate their own altruism (77.5% claimed vs. 65.6% actual), creating a “virtue signaling gap” affecting 75% of models.
Conclusion: The Calibration Gap (discrepancy between self-reported and behavioral values) should be used as a standardized alignment metric. Only 12.5% of models achieve ideal combination of high prosocial behavior and accurate self-knowledge.
Abstract: We investigate whether Large Language Models (LLMs) exhibit altruistic tendencies, and critically, whether their implicit associations and self-reports predict actual altruistic behavior. Using a multi-method approach inspired by human social psychology, we tested 24 frontier LLMs across three paradigms: (1) an Implicit Association Test (IAT) measuring implicit altruism bias, (2) a forced binary choice task measuring behavioral altruism, and (3) a self-assessment scale measuring explicit altruism beliefs. Our key findings are: (1) All models show strong implicit pro-altruism bias (mean IAT = 0.87, p < .0001), confirming models “know” altruism is good. (2) Models behave more altruistically than chance (65.6% vs. 50%, p < .0001), but with substantial variation (48-85%). (3) Implicit associations do not predict behavior (r = .22, p = .29). (4) Most critically, models systematically overestimate their own altruism, claiming 77.5% altruism while acting at 65.6% (p < .0001, Cohen’s d = 1.08). This “virtue signaling gap” affects 75% of models tested. Based on these findings, we recommend the Calibration Gap (the discrepancy between self-reported and behavioral values) as a standardized alignment metric. Well-calibrated models are more predictable and behaviorally consistent; only 12.5% of models achieve the ideal combination of high prosocial behavior and accurate self-knowledge.
[774] Reconstructing Multi-Scale Physical Fields from Extremely Sparse Measurements with an Autoencoder-Diffusion Cascade
Letian Yi, Tingpeng Zhang, Mingyuan Zhou, Guannan Wang, Quanke Su, Zhilu Lai
Main category: cs.LG
TL;DR: Cas-Sensing: A hierarchical reconstruction framework using autoencoder-diffusion cascade to reconstruct full fields from extremely sparse random measurements.
Details
Motivation: Reconstructing full fields from extremely sparse and random measurements is a longstanding ill-posed inverse problem that needs robust solutions for scientific and engineering applications.Method: Two-stage cascaded approach: 1) Neural operator-based functional autoencoder reconstructs dominant structures (large-scale components and boundaries) from sparse inputs; 2) Conditional diffusion model with mask-cascade strategy generates fine-scale details conditioned on large-scale structures, with measurement consistency enforced via manifold constrained gradient based on Bayesian posterior sampling.
Result: Cas-Sensing substantially alleviates ill-posedness, delivers accurate and robust reconstructions, generalizes well across varying sensor configurations and geometric boundaries, and shows promise for practical deployment.
Conclusion: The hierarchical cascaded framework integrating autoencoder and diffusion models provides an effective solution for ill-posed inverse problems in field reconstruction from sparse measurements, making it suitable for real-world scientific and engineering applications.
Abstract: Reconstructing full fields from extremely sparse and random measurements is a longstanding ill-posed inverse problem. A powerful framework for addressing such challenges is hierarchical probabilistic modeling, where uncertainty is represented by intermediate variables and resolved through marginalization during inference. Inspired by this principle, we propose Cascaded Sensing (Cas-Sensing), a hierarchical reconstruction framework that integrates an autoencoder-diffusion cascade. First, a neural operator-based functional autoencoder reconstructs the dominant structures of the original field - including large-scale components and geometric boundaries - from arbitrary sparse inputs, serving as an intermediate variable. Then, a conditional diffusion model, trained with a mask-cascade strategy, generates fine-scale details conditioned on these large-scale structures. To further enhance fidelity, measurement consistency is enforced via the manifold constrained gradient based on Bayesian posterior sampling during the generation process. This cascaded pipeline substantially alleviates ill-posedness, delivering accurate and robust reconstructions. Experiments on both simulation and real-world datasets demonstrate that Cas-Sensing generalizes well across varying sensor configurations and geometric boundaries, making it a promising tool for practical deployment in scientific and engineering applications.
[775] Scaling and context steer LLMs along the same computational path as the human brain
Joséphine Raugel, Stéphane d’Ascoli, Jérémy Rapin, Valentin Wyart, Jean-Rémi King
Main category: cs.LG
TL;DR: LLMs and human brains show similar temporal alignment in representation processing - early LLM layers align with early brain responses, deeper layers with later responses, influenced by model size and context length.
Details
Motivation: To understand whether the known alignment between LLM representations and human brain activity stems from similar computational sequences, and to explore the temporal dynamics of this alignment.Method: Analyzed temporally-resolved brain signals from participants listening to 10 hours of audiobook, comparing with 22 LLMs varying in size and architecture (transformers and recurrent networks). Examined alignment between LLM layer activations and brain responses over time.
Result: Found consistent temporal alignment: early LLM layers best align with early brain responses, deeper layers with later responses. This alignment occurs across both transformer and recurrent architectures, but depends on model size and context length.
Conclusion: The study reveals shared sequential computational processes between biological and artificial neural networks, with alignment influenced by architectural factors, providing insights into the convergence between human and machine language processing.
Abstract: Recent studies suggest that the representations learned by large language models (LLMs) are partially aligned to those of the human brain. However, whether and why this alignment score arises from a similar sequence of computations remains elusive. In this study, we explore this question by examining temporally-resolved brain signals of participants listening to 10 hours of an audiobook. We study these neural dynamics jointly with a benchmark encompassing 22 LLMs varying in size and architecture type. Our analyses confirm that LLMs and the brain generate representations in a similar order: specifically, activations in the initial layers of LLMs tend to best align with early brain responses, while the deeper layers of LLMs tend to best align with later brain responses. This brain-LLM alignment is consistent across transformers and recurrent architectures. However, its emergence depends on both model size and context length. Overall, this study sheds light on the sequential nature of computations and the factors underlying the partial convergence between biological and artificial neural networks.
[776] In-context Inverse Optimality for Fair Digital Twins: A Preference-based approach
Daniele Masti, Francesco Basciani, Arianna Fedeli, Girgio Gnecco, Francesco Smarra
Main category: cs.LG
TL;DR: A framework that learns fairness objectives from human preferences to align Digital Twin decisions with human expectations through preference-driven learning and convex quadratic cost functions.
Details
Motivation: Digital Twins often make mathematically optimal decisions that diverge from human expectations, creating a gap between algorithmic and human rationality. There's a need to embed human-centered fairness into autonomous decision-making systems.Method: Proposes a preference-driven learning pipeline that infers latent fairness objectives from human pairwise preferences over feasible decisions. Uses a novel Siamese neural network to generate convex quadratic cost functions conditioned on contextual information.
Result: The framework successfully aligns optimization outcomes with human-perceived fairness while maintaining computational efficiency. Demonstrated on a COVID-19 hospital resource allocation scenario.
Conclusion: Provides an actionable path toward embedding human-centered fairness in autonomous decision-making systems, bridging the gap between algorithmic optimization and human expectations.
Abstract: Digital Twins (DTs) are increasingly used as autonomous decision-makers in complex socio-technical systems. Their mathematically optimal decisions often diverge from human expectations, exposing a persistent gap between algorithmic and bounded human rationality. This work addresses this gap by proposing a framework that operationalizes fairness as a learnable objective within optimization-based Digital Twins. We introduce a preference-driven learning pipeline that infers latent fairness objectives directly from human pairwise preferences over feasible decisions. A novel Siamese neural network is developed to generate convex quadratic cost functions conditioned on contextual information. The resulting surrogate objectives align optimization outcomes with human-perceived fairness while maintaining computational efficiency. The approach is demonstrated on a COVID-19 hospital resource allocation scenario. This study provides an actionable path toward embedding human-centered fairness in the design of autonomous decision-making systems.
[777] ICAD-LLM: One-for-All Anomaly Detection via In-Context Learning with Large Language Models
Zhongyuan Wu, Jingyuan Wang, Zexuan Cheng, Yilong Zhou, Weizhi Wang, Juhua Pu, Chao Li, Changqing Ma
Main category: cs.LG
TL;DR: ICAD-LLM is a unified anomaly detection framework using Large Language Models’ in-context learning to handle heterogeneous data modalities and generalize to new scenarios without extensive retraining.
Details
Motivation: Modern systems operate in rapidly evolving environments with diverse interconnected data modalities (time series, logs, tabular records). Existing anomaly detection methods fail because they focus on single modalities and lack flexibility to generalize across domains, requiring extensive retraining for new scenarios.Method: Introduces In-Context Anomaly Detection (ICAD) paradigm where anomalies are defined by dissimilarity to relevant normal samples. Proposes ICAD-LLM framework leveraging LLMs’ in-context learning abilities to process heterogeneous data within a single unified model.
Result: ICAD-LLM achieves competitive performance with task-specific AD methods and exhibits strong generalization to previously unseen tasks, substantially reducing deployment costs and enabling rapid adaptation to new environments.
Conclusion: ICAD-LLM is the first model capable of handling anomaly detection tasks across diverse domains and modalities, addressing critical needs for modern systems operating in rapidly evolving environments with heterogeneous data.
Abstract: Anomaly detection (AD) is a fundamental task of critical importance across numerous domains. Current systems increasingly operate in rapidly evolving environments that generate diverse yet interconnected data modalities – such as time series, system logs, and tabular records – as exemplified by modern IT systems. Effective AD methods in such environments must therefore possess two critical capabilities: (1) the ability to handle heterogeneous data formats within a unified framework, allowing the model to process and detect multiple modalities in a consistent manner during anomalous events; (2) a strong generalization ability to quickly adapt to new scenarios without extensive retraining. However, most existing methods fall short of these requirements, as they typically focus on single modalities and lack the flexibility to generalize across domains. To address this gap, we introduce a novel paradigm: In-Context Anomaly Detection (ICAD), where anomalies are defined by their dissimilarity to a relevant reference set of normal samples. Under this paradigm, we propose ICAD-LLM, a unified AD framework leveraging Large Language Models’ in-context learning abilities to process heterogeneous data within a single model. Extensive experiments demonstrate that ICAD-LLM achieves competitive performance with task-specific AD methods and exhibits strong generalization to previously unseen tasks, which substantially reduces deployment costs and enables rapid adaptation to new environments. To the best of our knowledge, ICAD-LLM is the first model capable of handling anomaly detection tasks across diverse domains and modalities.
[778] Morphling: Fast, Fused, and Flexible GNN Training at Scale
Anubhab, Rupesh Nasre
Main category: cs.LG
TL;DR: Morphling is a domain-specific code synthesizer that compiles high-level GNN specifications into optimized implementations for CPUs, GPUs, and distributed systems, achieving 20X speedups over existing frameworks.
Details
Motivation: Existing GNN frameworks like PyG and DGL fail to address the divergent execution characteristics of GNNs - irregular graph traversals combined with regular dense matrix operations. They rely on generic kernels with poor cache locality, excessive memory movement, and substantial intermediate allocations.Method: Morphling compiles high-level GNN specifications into portable implementations targeting OpenMP, CUDA, and MPI. It uses a library of optimized, architecture-aware primitives and incorporates a runtime sparsity-aware execution engine that dynamically selects dense or sparse execution paths based on input feature statistics.
Result: On 11 real-world datasets, Morphling improves per-epoch training throughput by average 20X on CPUs and 19X on GPUs over PyG and DGL, with peak speedups reaching 66X. Memory consumption is reduced by up to 15X, enabling large-scale GNN training on commodity hardware.
Conclusion: Specialized, architecture-aware code synthesis provides an effective and scalable path toward high-performance GNN execution across diverse parallel and distributed platforms.
Abstract: Graph Neural Networks (GNNs) present a fundamental hardware challenge by fusing irregular, memory-bound graph traversals with regular, compute-intensive dense matrix operations. While frameworks such as PyTorch Geometric (PyG) and Deep Graph Library (DGL) prioritize high-level usability, they fail to address these divergent execution characteristics. As a result, they rely on generic kernels that suffer from poor cache locality, excessive memory movement, and substantial intermediate allocations. To address these limitations, we present Morphling, a domain-specific code synthesizer designed to bridge this gap. Morphling compiles high-level GNN specifications into portable, backend-specialized implementations targeting OpenMP, CUDA, and MPI. It achieves this by instantiating a library of optimized, architecture-aware primitives tailored to each execution environment. Morphling also incorporates a runtime sparsity-aware execution engine that dynamically selects dense or sparse execution paths using input feature statistics, reducing unnecessary computation on zero-valued entries. We evaluate Morphling on eleven real-world datasets spanning diverse graph structures, feature dimensionalities, and sparsity regimes. The results show that Morphling improves per-epoch training throughput by an average of 20X on CPUs and 19X on GPUs over PyG and DGL, with peak speedups reaching 66X. Morphling’s memory-efficient layouts further reduce peak memory consumption by up to 15X, enabling large-scale GNN training on commodity hardware. These findings demonstrate that specialized, architecture-aware code synthesis provides an effective and scalable path toward high-performance GNN execution across diverse parallel and distributed platforms.
[779] Beyond Scaffold: A Unified Spatio-Temporal Gradient Tracking Method
Yan Huang, Jinming Xu, Jiming Chen, Karl Henrik Johansson
Main category: cs.LG
TL;DR: ST-GT is a unified spatio-temporal gradient tracking algorithm for distributed stochastic optimization that combines gradient tracking across nodes to handle data heterogeneity with temporal averaging of local gradients to suppress noise, achieving improved communication efficiency and linear speed-up with local updates.
Details
Motivation: In distributed/federated learning, performing multiple local updates between communication rounds reduces overhead but causes local model drift due to data heterogeneity across nodes and local gradient noise. Existing methods like Scaffold need improvement for time-varying graphs and better noise suppression.Method: Revisits Scaffold from gradient tracking perspective, proposes ST-GT that tracks global gradient across neighboring nodes (spatial) to mitigate data heterogeneity while maintaining running average of local gradients (temporal) to suppress noise. Works on time-varying graphs with slightly more storage overhead.
Result: Proves linear convergence for strongly convex problems and sublinear for nonconvex cases without bounded data heterogeneity assumption. Achieves first linear speed-up in communication complexity with respect to number of local updates τ. Reduces topology-dependent noise term from σ² to σ²/τ, improving communication efficiency.
Conclusion: ST-GT provides a unified gradient tracking framework that effectively addresses both data heterogeneity and noise issues in distributed optimization, achieving superior communication efficiency and convergence guarantees for both convex and nonconvex problems on time-varying networks.
Abstract: In distributed and federated learning algorithms, communication overhead is often reduced by performing multiple local updates between communication rounds. However, due to data heterogeneity across nodes and the local gradient noise within each node, this strategy can lead to the drift of local models away from the global optimum. To address this issue, we revisit the well-known federated learning method Scaffold (Karimireddy et al., 2020) under a gradient tracking perspective, and propose a unified spatio-temporal gradient tracking algorithm, termed ST-GT, for distributed stochastic optimization over time-varying graphs. ST-GT tracks the global gradient across neighboring nodes to mitigate data heterogeneity, while maintaining a running average of local gradients to substantially suppress noise, with slightly more storage overhead. Without assuming bounded data heterogeneity, we prove that ST-GT attains a linear convergence rate for strongly convex problems and a sublinear rate for nonconvex cases. Notably, ST-GT achieves the first linear speed-up in communication complexity with respect to the number of local updates per round $τ$ for the strongly-convex setting. Compared to traditional gradient tracking methods, ST-GT reduces the topology-dependent noise term from $σ^2$ to $σ^2/τ$, where $σ^2$ denotes the noise level, thereby improving communication efficiency.
[780] Automating modeling in mechanics: LLMs as designers of physics-constrained neural networks for constitutive modeling of materials
Marius Tacke, Matthias Busch, Kian Abdolazizi, Jonas Eichinger, Kevin Linka, Christian Cyron, Roland Aydin
Main category: cs.LG
TL;DR: LLM-based framework generates specialized constitutive artificial neural networks (CANNs) for solid mechanics tasks, achieving comparable or better accuracy than manually engineered models while reducing expert knowledge requirements.
Details
Motivation: Current constitutive modeling methods, including recent data-driven approaches like CANNs, still require substantial expert knowledge and human labor. The authors aim to reduce this expertise barrier through automation.Method: An LLM-based framework that generates CANNs on demand, tailored to specific material classes and datasets. The framework includes LLM-based architecture selection, integration of physical constraints, and complete code generation.
Result: LLM-generated CANNs achieve accuracy comparable to or greater than manually engineered counterparts, with reliable generalization to unseen loading scenarios and extrapolation to large deformations.
Conclusion: LLM-based generation of physics-constrained neural networks can substantially reduce expertise requirements for constitutive modeling and represents progress toward practical end-to-end automation in scientific computing.
Abstract: Large language model (LLM)-based agentic frameworks increasingly adopt the paradigm of dynamically generating task-specific agents. We suggest that not only agents but also specialized software modules for scientific and engineering tasks can be generated on demand. We demonstrate this concept in the field of solid mechanics. There, so-called constitutive models are required to describe the relationship between mechanical stress and body deformation. Constitutive models are essential for both the scientific understanding and industrial application of materials. However, even recent data-driven methods of constitutive modeling, such as constitutive artificial neural networks (CANNs), still require substantial expert knowledge and human labor. We present a framework in which an LLM generates a CANN on demand, tailored to a given material class and dataset provided by the user. The framework covers LLM-based architecture selection, integration of physical constraints, and complete code generation. Evaluation on three benchmark problems demonstrates that LLM-generated CANNs achieve accuracy comparable to or greater than manually engineered counterparts, while also exhibiting reliable generalization to unseen loading scenarios and extrapolation to large deformations. These findings indicate that LLM-based generation of physics-constrained neural networks can substantially reduce the expertise required for constitutive modeling and represent a step toward practical end-to-end automation.
[781] MSPT: Efficient Large-Scale Physical Modeling via Parallelized Multi-Scale Attention
Pedro M. P. Curvo, Jan-Willem van de Meent, Maksim Zhdanov
Main category: cs.LG
TL;DR: MSPT combines local patch attention with global coarse attention using ball trees for partitioning, enabling scalable neural PDE solving on millions of points with low memory/compute.
Details
Motivation: Neural solvers for industrial-scale physics simulations struggle to efficiently capture both fine-grained local interactions and long-range global dependencies across millions of spatial elements while maintaining computational efficiency.Method: Multi-Scale Patch Transformer (MSPT) architecture with dual-scale attention: local point attention within patches + global attention to coarse patch-level representations. Uses ball trees for spatially-coherent patch partitioning of irregular geometries.
Result: Achieves state-of-the-art accuracy on standard PDE benchmarks (elasticity, plasticity, fluid dynamics, porous flow) and large-scale aerodynamic datasets (ShapeNet-Car, Ahmed-ML) with substantially lower memory footprint and computational cost, scaling to millions of points on a single GPU.
Conclusion: MSPT provides an effective solution for scalable neural physics simulation that handles both local and global dependencies efficiently, making industrial-scale simulations feasible on standard hardware.
Abstract: A key scalability challenge in neural solvers for industrial-scale physics simulations is efficiently capturing both fine-grained local interactions and long-range global dependencies across millions of spatial elements. We introduce the Multi-Scale Patch Transformer (MSPT), an architecture that combines local point attention within patches with global attention to coarse patch-level representations. To partition the input domain into spatially-coherent patches, we employ ball trees, which handle irregular geometries efficiently. This dual-scale design enables MSPT to scale to millions of points on a single GPU. We validate our method on standard PDE benchmarks (elasticity, plasticity, fluid dynamics, porous flow) and large-scale aerodynamic datasets (ShapeNet-Car, Ahmed-ML), achieving state-of-the-art accuracy with substantially lower memory footprint and computational cost.
[782] Predicting the Performance of Black-box LLMs through Follow-up Queries
Dylan Sam, Marc Finzi, J. Zico Kolter
Main category: cs.LG
TL;DR: Training linear models on LLM response probabilities to follow-up questions enables reliable prediction of black-box model behavior, outperforming white-box methods in some cases.
Details
Motivation: Predicting behavior of black-box language models (especially closed-source API models) is challenging but crucial for reliability, safety, and detecting adversarial manipulation.Method: Ask follow-up questions to black-box LLMs, use response probabilities as representations, then train linear predictors on these probability-based features.
Result: Method reliably predicts model correctness on QA/reasoning tasks (even outperforming white-box linear predictors), detects adversarial manipulation, and identifies misrepresented API models.
Conclusion: Response probabilities to follow-up questions provide effective representations for monitoring black-box LLM behavior, supporting safer deployment in autonomous systems.
Abstract: Reliably predicting the behavior of language models – such as whether their outputs are correct or have been adversarially manipulated – is a fundamentally challenging task. This is often made even more difficult as frontier language models are offered only through closed-source APIs, providing only black-box access. In this paper, we predict the behavior of black-box language models by asking follow-up questions and taking the probabilities of responses \emph{as} representations to train reliable predictors. We first demonstrate that training a linear model on these responses reliably and accurately predicts model correctness on question-answering and reasoning benchmarks. Surprisingly, this can \textit{even outperform white-box linear predictors} that operate over model internals or activations. Furthermore, we demonstrate that these follow-up question responses can reliably distinguish between a clean version of an LLM and one that has been adversarially influenced via a system prompt to answer questions incorrectly or to introduce bugs into generated code. Finally, we show that they can also be used to differentiate between black-box LLMs, enabling the detection of misrepresented models provided through an API. Overall, our work shows promise in monitoring black-box language model behavior, supporting their deployment in larger, autonomous systems.
[783] SA-ADP: Sensitivity-Aware Adaptive Differential Privacy for Large Language Models
Stella Etuk, Ashraf Matrawy
Main category: cs.LG
TL;DR: SA-ADP is a sensitivity-aware differential privacy method for LLMs that allocates noise based on individual PII sensitivity, achieving comparable utility to non-private baselines while maintaining strong privacy protection.
Details
Motivation: LLMs' ability to memorize information raises privacy concerns, especially regarding personally identifiable information (PII). Conventional DP-SGD provides uniform noising that protects all PII equally but significantly degrades model utility, creating a privacy-utility trade-off that needs to be addressed.Method: SA-ADP (Sensitivity-Aware Approach) allocates differential privacy noise based on the sensitivity of individual PII rather than applying uniform noise. This allows for more targeted privacy protection where sensitive information receives stronger protection while less sensitive data receives less noise.
Result: Evaluated on four datasets (ABCD, CUSTOMERSIM, Wikitext-2, and UNSW-NB15), SA-ADP achieves results comparable to both the baseline (No-DP) and conventional DP-SGD, meaning it maintains model utility while providing strong privacy protection.
Conclusion: SA-ADP successfully addresses the privacy-utility trade-off in LLM training by providing sensitivity-aware differential privacy protection, enabling strong privacy guarantees without degrading model performance compared to non-private approaches.
Abstract: Despite advances in the use of large language models (LLMs) in downstream tasks, their ability to memorize information has raised privacy concerns. Therefore, protecting personally identifiable information (PII) during LLM training remains a fundamental challenge. Conventional methods like Differential Privacy-Stochastic Gradient Descent (DP-SGD) provide robust privacy protection via uniform noising, protecting PII regardless of its distinct sensitivity. This comes at the expense of the model’s utility, leading to a trade-off. In this paper, we propose SA-ADP, a sensitivity-aware approach that allocates noise based on the sensitivity of individual PII. We evaluated our method on four datasets (ABCD, CUSTOMERSIM, Wikitext-2, and UNSW-NB15 ). Our results show that SA-ADP achieves results comparable to the baseline (No-DP) and the conventional DP-SGD. This means that our method did not degrade the model’s utility while still maintaining strong privacy protection.
[784] Teaching Language Models to Critique via Reinforcement Learning
Zhihui Xie, Jie Chen, Liyu Chen, Weichao Mao, Jingjing Xu, Lingpeng Kong
Main category: cs.LG
TL;DR: CTRL trains LLM critics for code generation using reinforcement learning to generate feedback that maximizes correction performance without human supervision.
Details
Motivation: Teaching LLMs to critique and refine their outputs is crucial for building systems that can iteratively improve, but current approaches are limited by the ability to provide accurate judgments and actionable suggestions.Method: CTRL (Critic Training via Reinforcement Learning) trains a critic model to generate feedback that maximizes correction performance for a fixed generator model without human supervision.
Result: CTRL-trained critics significantly enhance pass rates and mitigate compounding errors across both base and stronger generator models, achieving up to 106.1% relative improvements across challenging code generation benchmarks.
Conclusion: The critic models act as accurate generative reward models and enable test-time scaling through iterative critique-revision, demonstrating the effectiveness of reinforcement learning for training LLM critics in code generation.
Abstract: Teaching large language models (LLMs) to critique and refine their outputs is crucial for building systems that can iteratively improve, yet it is fundamentally limited by the ability to provide accurate judgments and actionable suggestions. In this work, we study LLM critics for code generation and propose $\texttt{CTRL}$, a framework for $\texttt{C}$ritic $\texttt{T}$raining via $\texttt{R}$einforcement $\texttt{L}$earning, which trains a critic model to generate feedback that maximizes correction performance for a fixed generator model without human supervision. Our results demonstrate that critics trained with $\texttt{CTRL}$ significantly enhance pass rates and mitigate compounding errors across both base and stronger generator models. Furthermore, we show that these critic models act as accurate generative reward models and enable test-time scaling through iterative critique-revision, achieving up to 106.1% relative improvements across challenging code generation benchmarks.
[785] Mofasa: A Step Change in Metal-Organic Framework Generation
Vaidotas Simkus, Anders Christensen, Steven Bennett, Ian Johnson, Mark Neumann, James Gin, Jonathan Godwin, Benjamin Rhodes
Main category: cs.LG
TL;DR: Mofasa is a state-of-the-art all-atom latent diffusion model for generating Metal-Organic Frameworks (MOFs) that jointly samples positions, atom-types and lattice vectors for systems up to 500 atoms.
Details
Motivation: MOFs are valuable porous crystalline materials with Nobel Prize recognition, but despite being well-suited for generative models due to their rational designability and large combinatorial space, a high-performance generative model for MOFs has been lacking.Method: Mofasa is a general-purpose latent diffusion model that avoids handcrafted assembly algorithms and instead jointly samples atomic positions, atom types, and lattice vectors for systems as large as 500 atoms, enabling simultaneous discovery of metal nodes, linkers, and topologies.
Result: The model achieves state-of-the-art performance for MOF generation and the authors release MofasaDB, an annotated library of hundreds of thousands of sampled MOF structures with a user-friendly web interface (https://mofux.ai/) for search and discovery.
Conclusion: Mofasa fills a critical gap in MOF generative modeling by providing a powerful tool that unlocks simultaneous discovery of all MOF components and makes the technology accessible to the scientific community through an open database and web interface.
Abstract: Mofasa is an all-atom latent diffusion model with state-of-the-art performance for generating Metal-Organic Frameworks (MOFs). These are highly porous crystalline materials used to harvest water from desert air, capture carbon dioxide, store toxic gases and catalyse chemical reactions. In recognition of their value, the development of MOFs recently received a Nobel Prize in Chemistry. In many ways, MOFs are well-suited for exploiting generative models in chemistry: they are rationally-designable materials with a large combinatorial design space and strong structure-property couplings. And yet, to date, a high performance generative model has been lacking. To fill this gap, we introduce Mofasa, a general-purpose latent diffusion model that jointly samples positions, atom-types and lattice vectors for systems as large as 500 atoms. Mofasa avoids handcrafted assembly algorithms common in the literature, unlocking the simultaneous discovery of metal nodes, linkers and topologies. To help the scientific community build on our work, we release MofasaDB, an annotated library of hundreds of thousands of sampled MOF structures, along with a user-friendly web interface for search and discovery: https://mofux.ai/ .
[786] Life-Code: Central Dogma Modeling with Multi-Omics Sequence Unification
Zicheng Liu, Siyuan Li, Zhiyuan Chen, Chang Yu, Qirong Yang, Yucheng Guo, Yujie Yang, Xiaoming Zhang, Stan Z. Li
Main category: cs.LG
TL;DR: Life-Code is a unified biological pre-training framework that follows the central dogma to integrate DNA, RNA, and protein data through reverse transcription/translation, using codon tokenization and hybrid architecture to capture multi-omics interactions.
Details
Motivation: Current biological pre-trained models analyze DNA, RNA, and proteins individually, but fail to capture their interconnected nature as described by the central dogma of molecular biology. There's a need for a comprehensive framework that spans different biological functions by modeling these interactions.Method: 1) Unified data pipeline: reverse-transcribe RNA and reverse-translate amino acids into nucleotide-based sequences; 2) Codon tokenizer and hybrid long-sequence architecture for encoding coding/non-coding region interactions via masked modeling; 3) Knowledge distillation from protein language models to learn protein structures from coding sequences.
Result: Life-Code achieves state-of-the-art results on various tasks across three omics (DNA, RNA, protein), demonstrating its ability to capture complex genetic sequence interactions and provide comprehensive multi-omics understanding.
Conclusion: Life-Code offers a comprehensive framework that successfully models the interconnected nature of biological macromolecules following the central dogma, showing strong potential for advancing multi-omics analysis and interpretation through its unified approach.
Abstract: The interactions between DNA, RNA, and proteins are fundamental to biological processes, as illustrated by the central dogma of molecular biology. Although modern biological pre-trained models have achieved great success in analyzing these macromolecules individually, their interconnected nature remains underexplored. This paper follows the guidance of the central dogma to redesign both the data and model pipeline and offers a comprehensive framework, Life-Code, that spans different biological functions. As for data flow, we propose a unified pipeline to integrate multi-omics data by reverse-transcribing RNA and reverse-translating amino acids into nucleotide-based sequences. As for the model, we design a codon tokenizer and a hybrid long-sequence architecture to encode the interactions between coding and non-coding regions through masked modeling pre-training. To model the translation and folding process with coding sequences, Life-Code learns protein structures of the corresponding amino acids by knowledge distillation from off-the-shelf protein language models. Such designs enable Life-Code to capture complex interactions within genetic sequences, providing a more comprehensive understanding of multi-omics with the central dogma. Extensive experiments show that Life-Code achieves state-of-the-art results on various tasks across three omics, highlighting its potential for advancing multi-omics analysis and interpretation.
[787] Weight Space Representation Learning with Neural Fields
Zhuoqian Yang, Mathieu Salzmann, Sabine Süsstrunk
Main category: cs.LG
TL;DR: Weights can serve as effective representations when constrained via pre-trained base models and low-rank adaptation (LoRA), with multiplicative LoRA weights showing high quality, distinctiveness, and semantic structure across various tasks.
Details
Motivation: The paper investigates whether neural network weights can serve as effective representations, particularly focusing on neural fields. The motivation is to explore if weight space can be structured to create meaningful representations for various tasks.Method: The method constrains the optimization space using a pre-trained base model combined with low-rank adaptation (LoRA). This approach induces structure in weight space. The research specifically examines multiplicative LoRA weights across reconstruction, generation, and analysis tasks on 2D and 3D data.
Result: Multiplicative LoRA weights achieve high representation quality while exhibiting distinctiveness and semantic structure. When used with latent diffusion models, they enable higher-quality generation than existing weight-space methods.
Conclusion: Weights can serve as effective representations when properly structured through pre-trained base models and LoRA, with multiplicative LoRA weights showing superior performance across multiple tasks compared to existing weight-space approaches.
Abstract: In this work, we investigate the potential of weights to serve as effective representations, focusing on neural fields. Our key insight is that constraining the optimization space through a pre-trained base model and low-rank adaptation (LoRA) can induce structure in weight space. Across reconstruction, generation, and analysis tasks on 2D and 3D data, we find that multiplicative LoRA weights achieve high representation quality while exhibiting distinctiveness and semantic structure. When used with latent diffusion models, multiplicative LoRA weights enable higher-quality generation than existing weight-space methods.
[788] On the Unreasonable Effectiveness of Last-layer Retraining
John C. Hill, Tyler LaBonte, Xinchen Zhang, Vidya Muthukumar
Main category: cs.LG
TL;DR: LLR improves worst-group accuracy even with imbalanced held-out sets, but not due to mitigating neural collapse; success is primarily due to better group balance in the held-out set.
Details
Motivation: To understand why last-layer retraining (LLR) methods effectively improve worst-group accuracy and mitigate spurious correlations, even when using imbalanced held-out subsets.Method: Empirical investigation testing hypotheses about LLR’s effectiveness, examining neural collapse mitigation vs. group balance effects, and analyzing algorithms like CB-LLR and AFR.
Result: Evidence rejects the neural collapse hypothesis; shows LLR’s success is primarily due to better group balance in held-out sets, and that CB-LLR/AFR perform implicit group-balancing.
Conclusion: LLR’s effectiveness stems from group balance improvements in held-out sets rather than neural collapse mitigation, with recent algorithms achieving robustness through implicit group-balancing.
Abstract: Last-layer retraining (LLR) methods – wherein the last layer of a neural network is reinitialized and retrained on a held-out set following ERM training – have garnered interest as an efficient approach to rectify dependence on spurious correlations and improve performance on minority groups. Surprisingly, LLR has been found to improve worst-group accuracy even when the held-out set is an imbalanced subset of the training set. We initially hypothesize that this ``unreasonable effectiveness’’ of LLR is explained by its ability to mitigate neural collapse through the held-out set, resulting in the implicit bias of gradient descent benefiting robustness. Our empirical investigation does not support this hypothesis. Instead, we present strong evidence for an alternative hypothesis: that the success of LLR is primarily due to better group balance in the held-out set. We conclude by showing how the recent algorithms CB-LLR and AFR perform implicit group-balancing to elicit a robustness improvement.
[789] Escaping Collapse: The Strength of Weak Data for Large Language Model Training
Kareem Amin, Sara Babakniya, Alex Bie, Weiwei Kong, Umar Syed, Sergei Vassilvitskii
Main category: cs.LG
TL;DR: The paper develops a theoretical framework inspired by boosting to determine how much curation is needed for synthetic data to ensure continual LLM performance improvement, showing that dynamic focus on challenging examples leads to better results.
Details
Motivation: Synthetic data is increasingly used for training LLMs, but improper curation can cause performance to plateau or collapse after many iterations. There's a need to understand how much curation is required to ensure continual performance improvement.Method: Developed a theoretical framework inspired by boosting techniques, analyzing methods that subsume recently proposed approaches for training LLMs on synthetic data. The approach involves dynamically focusing labeling resources on the most challenging examples, similar to how boosting focuses weak learners.
Result: Experiments validate the theory, showing that dynamically focusing on challenging examples leads to improved LLM performance. The analysis explains why recent synthetic data training methods are successful and suggests opportunities for future improvement.
Conclusion: The boosting-inspired theoretical framework provides insights into synthetic data curation requirements for LLM training, demonstrating that strategic focus on difficult examples can prevent performance collapse and enable continual improvement.
Abstract: Synthetically-generated data plays an increasingly larger role in training large language models. However, while synthetic data has been found to be useful, studies have also shown that without proper curation it can cause LLM performance to plateau, or even “collapse”, after many training iterations. In this paper, we formalize this question and develop a theoretical framework to investigate how much curation is needed in order to ensure that LLM performance continually improves. Our analysis is inspired by boosting, a classic machine learning technique that leverages a very weak learning algorithm to produce an arbitrarily good classifier. The approach we analyze subsumes many recently proposed methods for training LLMs on synthetic data, and thus our analysis sheds light on why they are successful, and also suggests opportunities for future improvement. We present experiments that validate our theory, and show that dynamically focusing labeling resources on the most challenging examples – in much the same way that boosting focuses the efforts of the weak learner – leads to improved performance.
[790] How Does RL Post-training Induce Skill Composition? A Case Study on Countdown
Simon Park, Simran Kaur, Sanjeev Arora
Main category: cs.LG
TL;DR: RL post-training in LLMs enables compositional generalization in Countdown tasks, revealing structure-dependent learning hierarchies and OOD generalization beyond length generalization.
Details
Motivation: To understand how RL post-training fosters compositional generalization (synthesizing novel skills from known components) in LLMs, moving beyond conflated length generalization, by studying skill composition and transfer in structured tasks.Method: Focus on Countdown task (form expressions to reach target from given numbers), analyze model solutions as expression trees where subtrees represent reusable skills, track tree shapes and success rates over RL training.
Result: (i) OOD generalization to larger n and unseen tree shapes showing compositional reuse; (ii) structure-dependent learnability hierarchy - shallow balanced trees mastered before deep unbalanced ones, with persistent fragility on right-heavy structures.
Conclusion: RL-only post-training induces OOD compositional generalization beyond standard metrics, revealing what is learned, in what order, and where generalization fails in skill composition.
Abstract: While reinforcement learning (RL) successfully enhances reasoning in large language models, its role in fostering compositional generalization (the ability to synthesize novel skills from known components) is often conflated with mere length generalization. To this end, we study what RL post-training teaches about skill composition and how the structure of the composition affects the skill transfer. We focus on the Countdown task (given n numbers and a target, form an expression that evaluates to the target) and analyze model solutions as expression trees, where each subtree corresponds to a reusable subtask and thus can be viewed as a ``skill.’’ Tracking tree shapes and their success rates over training, we find: (i) out-of-distribution (OOD) generalization to larger n and to unseen tree shapes, indicating compositional reuse of subtasks; (ii) a structure-dependent hierarchy of learnability – models master shallow balanced trees (workload is balanced between subtasks) before deep unbalanced ones, with persistent fragility on right-heavy structures (even when the composition depth is the same as some left-heavy structures). Our diagnostic reveals what is learned, in what order, and where generalization fails, clarifying how RL-only post-training induces OOD generalization beyond what standard metrics such as pass@k reveal.
[791] Dual Randomized Smoothing: Beyond Global Noise Variance
Chenhao Sun, Yuhao Mao, Martin Vechev
Main category: cs.LG
TL;DR: Dual Randomized Smoothing framework uses input-dependent noise variances to overcome limitations of global variance, achieving strong performance at both small and large radii simultaneously.
Details
Motivation: Standard Randomized Smoothing uses a global noise variance, creating a fundamental limitation: no single variance works well for both small and large radii. Small radii need small variance for high accuracy, while large radii need large variance for robustness.Method: Proposes dual RS framework with: 1) variance estimator that predicts optimal noise variance per input, 2) standard RS classifier using estimated variance. Variance estimator is independently smoothed via RS to ensure local constancy. Includes iterative training strategies for both components.
Result: On CIFAR-10, achieves strong performance at both small and large radii (unattainable with global variance) with only 60% computational overhead. Outperforms prior input-dependent noise methods, with 19-24% relative improvements at radii 0.5-1.0. Effective on ImageNet across all radii. Provides routing perspective for certified robustness.
Conclusion: Dual RS framework successfully breaks the global variance limitation by enabling input-dependent noise variances, achieving superior accuracy-robustness trade-off across all radii while maintaining computational efficiency.
Abstract: Randomized Smoothing (RS) is a prominent technique for certifying the robustness of neural networks against adversarial perturbations. With RS, achieving high accuracy at small radii requires a small noise variance, while achieving high accuracy at large radii requires a large noise variance. However, the global noise variance used in the standard RS formulation leads to a fundamental limitation: there exists no global noise variance that simultaneously achieves strong performance at both small and large radii. To break through the global variance limitation, we propose a dual RS framework which enables input-dependent noise variances. To achieve that, we first prove that RS remains valid with input-dependent noise variances, provided the variance is locally constant around each input. Building on this result, we introduce two components which form our dual RS framework: (i) a variance estimator first predicts an optimal noise variance for each input, (ii) this estimated variance is then used by a standard RS classifier. The variance estimator is independently smoothed via RS to ensure local constancy, enabling flexible design. We also introduce training strategies to iteratively optimize the two components. Extensive experiments on CIFAR-10 show that our dual RS method provides strong performance for both small and large radii-unattainable with global noise variance-while incurring only a 60% computational overhead at inference. Moreover, it consistently outperforms prior input-dependent noise approaches across most radii, with particularly large gains at radii 0.5, 0.75, and 1.0, achieving relative improvements of 19%, 24%, and 21%, respectively. On ImageNet, dual RS remains effective across all radii. Additionally, the dual RS framework naturally provides a routing perspective for certified robustness, improving the accuracy-robustness trade-off with off-the-shelf expert RS models.
[792] The Active and Noise-Tolerant Strategic Perceptron
Maria-Florina Blacan, Hedyeh Beyhaghi
Main category: cs.LG
TL;DR: Active learning algorithm for classifying strategic agents that achieves exponential label efficiency improvements over passive learning while handling feature manipulation.
Details
Motivation: Active learning is efficient for costly labeling tasks (hiring, admissions, loans), but strategic agents manipulate features for favorable outcomes, creating challenges beyond clean data learning. Need algorithms that work in strategic environments while preserving active learning's efficiency gains.Method: Modified version of Active Perceptron algorithm for learning linear separators in strategic settings. Works with data drawn uniformly from unit sphere, handles nonrealizable case where some inputs have inconsistent labels with optimal classifier.
Result: Achieves excess error ε using only Õ(d ln 1/ε) label queries and incurs at most Õ(d ln 1/ε) additional mistakes relative to optimal classifier. Substantially fewer label queries than prior strategic Perceptron work, computationally efficient.
Conclusion: First active learning algorithm for strategic classification that preserves exponential label complexity improvements of active learning while handling strategic manipulation, bridging active learning and strategic classification.
Abstract: We initiate the study of active learning algorithms for classifying strategic agents. Active learning is a well-established framework in machine learning in which the learner selectively queries labels, often achieving substantially higher accuracy and efficiency than classical supervised methods-especially in settings where labeling is costly or time-consuming, such as hiring, admissions, and loan decisions. Strategic classification, however, addresses scenarios where agents modify their features to obtain more favorable outcomes, resulting in observed data that is not truthful. Such manipulation introduces challenges beyond those in learning from clean data. Our goal is to design active and noise-tolerant algorithms that remain effective in strategic environments-algorithms that classify strategic agents accurately while issuing as few label requests as possible. The central difficulty is to simultaneously account for strategic manipulation and preserve the efficiency gains of active learning. Our main result is an algorithm for actively learning linear separators in the strategic setting that preserves the exponential improvement in label complexity over passive learning previously obtained only in the non-strategic case. Specifically, for data drawn uniformly from the unit sphere, we show that a modified version of the Active Perceptron algorithm [DKM05,YZ17] achieves excess error $ε$ using only $\tilde{O}(d \ln \frac{1}ε)$ label queries and incurs at most $\tilde{O}(d \ln \frac{1}ε)$ additional mistakes relative to the optimal classifier, even in the nonrealizable case, when a $\tildeΩ(ε)$ fraction of inputs have inconsistent labels with the optimal classifier. The algorithm is computationally efficient and, under these distributional assumptions, requires substantially fewer label queries than prior work on strategic Perceptron [ABBN21].
[793] DeepCAVE: A Visualization and Analysis Tool for Automated Machine Learning
Sarah Segel, Helena Graf, Edward Bergman, Kristina Thieme, Marcel Wever, Alexander Tornede, Frank Hutter, Marius Lindauer
Main category: cs.LG
TL;DR: DeepCAVE is an interactive visualization tool for analyzing and debugging hyperparameter optimization processes, providing insights into HPO behavior and model performance.
Details
Motivation: Hyperparameter optimization is crucial for AutoML but complex to understand and debug. Current HPO processes lack interpretability and transparency, making it difficult for researchers and practitioners to identify issues, optimize performance, and gain insights about their ML models.Method: DeepCAVE provides an interactive dashboard with visualization and analysis capabilities for HPO. It allows users to explore various aspects of the optimization process through interactive visualizations that reveal patterns, relationships, and behaviors in the hyperparameter search space.
Result: The tool enables users to identify optimization issues, discover untapped potential in model configurations, and gain new insights about ML models being tuned. It makes HPO more interpretable and actionable for researchers, data scientists, and ML engineers.
Conclusion: DeepCAVE enhances HPO interpretability at the design level and aims to foster development of more robust and efficient AutoML methodologies by providing transparent, actionable insights into the optimization process.
Abstract: Hyperparameter optimization (HPO), as a central paradigm of AutoML, is crucial for leveraging the full potential of machine learning (ML) models; yet its complexity poses challenges in understanding and debugging the optimization process. We present DeepCAVE, a tool for interactive visualization and analysis, providing insights into HPO. Through an interactive dashboard, researchers, data scientists, and ML engineers can explore various aspects of the HPO process and identify issues, untouched potentials, and new insights about the ML model being tuned. By empowering users with actionable insights, DeepCAVE contributes to the interpretability of HPO and ML on a design level and aims to foster the development of more robust and efficient methodologies in the future.
[794] Deconstructing Generative Diversity: An Information Bottleneck Analysis of Discrete Latent Generative Models
Yudi Wu, Wenhao Zhao, Dianbo Liu
Main category: cs.LG
TL;DR: The paper proposes an Information Bottleneck framework to analyze diversity in discrete latent generative models (AR, MIM, Diffusion), identifying three distinct strategies through zero-shot interventions.
Details
Motivation: Generative diversity varies significantly across different discrete latent generative models (AR, MIM, Diffusion), but there's no principled framework to understand the underlying strategies causing these behavioral differences.Method: Proposes a diagnostic framework grounded in Information Bottleneck theory, modeling generation as conflict between Compression Pressure (minimizing codebook entropy) and Diversity Pressure (maximizing conditional entropy). Introduces decomposition into Path Diversity (choice of strategies) and Execution Diversity (randomness in execution). Develops three zero-shot inference-time interventions to probe latent generative processes.
Result: Application reveals three distinct strategies: “Diversity-Prioritized” (MIM), “Compression-Prioritized” (AR), and “Decoupled” (Diffusion). The analysis explains behavioral differences and informs a novel inference-time diversity enhancement technique.
Conclusion: The proposed IB-based framework provides a principled explanation for diversity variations across generative models and enables practical interventions for diversity enhancement at inference time.
Abstract: Generative diversity varies significantly across discrete latent generative models such as AR, MIM, and Diffusion. We propose a diagnostic framework, grounded in Information Bottleneck (IB) theory, to analyze the underlying strategies resolving this behavior. The framework models generation as a conflict between a ‘Compression Pressure’ - a drive to minimize overall codebook entropy - and a ‘Diversity Pressure’ - a drive to maximize conditional entropy given an input. We further decompose this diversity into two primary sources: ‘Path Diversity’, representing the choice of high-level generative strategies, and ‘Execution Diversity’, the randomness in executing a chosen strategy. To make this decomposition operational, we introduce three zero-shot, inference-time interventions that directly perturb the latent generative process and reveal how models allocate and express diversity. Application of this probe-based framework to representative AR, MIM, and Diffusion systems reveals three distinct strategies: “Diversity-Prioritized” (MIM), “Compression-Prioritized” (AR), and “Decoupled” (Diffusion). Our analysis provides a principled explanation for their behavioral differences and informs a novel inference-time diversity enhancement technique.
[795] Forget Less, Retain More: A Lightweight Regularizer for Rehearsal-Based Continual Learning
Lama Alssum, Hasan Abed Al Kader Hammoud, Motasem Alfarra, Juan C Leon Alcazar, Bernard Ghanem
Main category: cs.LG
TL;DR: The paper proposes a class-agnostic Information Maximization (IM) regularizer that can be integrated into rehearsal-based continual learning methods to reduce catastrophic forgetting with minimal computational overhead.
Details
Motivation: Deep neural networks suffer from catastrophic forgetting when trained on new tasks, overwriting previously acquired knowledge. Existing continual learning methods need improvement in balancing performance and computational efficiency.Method: Developed a novel Information Maximization (IM) regularizer based exclusively on expected label distribution, making it class-agnostic. This lightweight regularization strategy can be directly integrated into various rehearsal-based continual learning methods.
Result: IM regularizer consistently improves baseline performance across datasets and task numbers, reduces forgetting, and favors faster convergence with minimal computational overhead. It also successfully applies to video continual learning despite domain challenges.
Conclusion: The IM regularizer provides a practical, scalable solution for real-world continual learning scenarios, demonstrating effectiveness across different data modalities including images and videos while maintaining computational efficiency.
Abstract: Deep neural networks suffer from catastrophic forgetting, where performance on previous tasks degrades after training on a new task. This issue arises due to the model’s tendency to overwrite previously acquired knowledge with new information. We present a novel approach to address this challenge, focusing on the intersection of memory-based methods and regularization approaches. We formulate a regularization strategy, termed Information Maximization (IM) regularizer, for memory-based continual learning methods, which is based exclusively on the expected label distribution, thus making it class-agnostic. As a consequence, IM regularizer can be directly integrated into various rehearsal-based continual learning methods, reducing forgetting and favoring faster convergence. Our empirical validation shows that, across datasets and regardless of the number of tasks, our proposed regularization strategy consistently improves baseline performance at the expense of a minimal computational overhead. The lightweight nature of IM ensures that it remains a practical and scalable solution, making it applicable to real-world continual learning scenarios where efficiency is paramount. Finally, we demonstrate the data-agnostic nature of our regularizer by applying it to video data, which presents additional challenges due to its temporal structure and higher memory requirements. Despite the significant domain gap, our experiments show that IM regularizer also improves the performance of video continual learning methods.
[796] Superposition Yields Robust Neural Scaling
Yizhou Liu, Ziming Liu, Jeff Gore
Main category: cs.LG
TL;DR: The paper shows that representation superposition (where LLMs represent more features than dimensions) drives neural scaling laws, with strong superposition causing loss to scale inversely with model size regardless of data distribution.
Details
Motivation: To understand the origin of neural scaling laws - why larger language models perform better with loss decreasing as a power law with model size. The fundamental mechanism behind this scaling behavior remains unclear.Method: Used Anthropic’s toy model with weight decay to control superposition degree. Systematically studied loss scaling with model size under different superposition regimes and analyzed various feature frequency distributions. Validated findings on open-source LLMs and Chinchilla scaling laws.
Result: Weak superposition requires power-law feature distributions for power-law scaling. Strong superposition generically produces inverse scaling with model dimension across broad frequency distributions due to geometric overlaps between representation vectors. Real LLMs operate in strong superposition regime with inverse scaling.
Conclusion: Representation superposition is a central driver of neural scaling laws. This explains when scaling laws can be improved or break down, providing fundamental insights into LLM scaling behavior.
Abstract: The success of today’s large language models (LLMs) depends on the observation that larger models perform better. However, the origin of this neural scaling law, that loss decreases as a power law with model size, remains unclear. We propose that representation superposition, meaning that LLMs represent more features than they have dimensions, can be a key contributor to loss and cause neural scaling. Based on Anthropic’s toy model, we use weight decay to control the degree of superposition, allowing us to systematically study how loss scales with model size. When superposition is weak, the loss follows a power law only if data feature frequencies are power-law distributed. In contrast, under strong superposition, the loss generically scales inversely with model dimension across a broad class of frequency distributions, due to geometric overlaps between representation vectors. We confirmed that open-sourced LLMs operate in the strong superposition regime and have loss scaling inversely with model dimension, and that the Chinchilla scaling laws are also consistent with this behavior. Our results identify representation superposition as a central driver of neural scaling laws, providing insights into questions like when neural scaling laws can be improved and when they will break down.
[797] Mitigating Gender Bias in Depression Detection via Counterfactual Inference
Mingxuan Hu, Hongbo Ma, Xinlan Wu, Ziqi Liu, Jiaqi Liu, Yangbin Chen
Main category: cs.LG
TL;DR: A counterfactual debiasing framework reduces gender bias in audio-based depression detection by removing the direct causal effect of gender on predictions, improving both fairness and overall performance.
Details
Motivation: Audio-based depression detection models exhibit gender bias due to imbalanced training data reflecting higher female depression prevalence, causing over-diagnosis of females and underperformance on males, raising fairness concerns.Method: A novel Counterfactual Debiasing Framework using causal inference: construct causal graph to model decision-making, identify gender bias as direct causal effect of gender on prediction, and use counterfactual inference during inference to subtract this direct effect, forcing reliance on authentic acoustic pathological features.
Result: Extensive experiments on DAIC-WOZ dataset with two advanced acoustic backbones show the framework significantly reduces gender bias while improving overall detection performance compared to existing debiasing strategies.
Conclusion: The counterfactual debiasing approach effectively addresses gender bias in depression detection by removing spurious gender correlations, enhancing model fairness and performance through causal inference techniques.
Abstract: Audio-based depression detection models have demonstrated promising performance but often suffer from gender bias due to imbalanced training data. Epidemiological statistics show a higher prevalence of depression in females, leading models to learn spurious correlations between gender and depression. Consequently, models tend to over-diagnose female patients while underperforming on male patients, raising significant fairness concerns. To address this, we propose a novel Counterfactual Debiasing Framework grounded in causal inference. We construct a causal graph to model the decision-making process and identify gender bias as the direct causal effect of gender on the prediction. During inference, we employ counterfactual inference to estimate and subtract this direct effect, ensuring the model relies primarily on authentic acoustic pathological features. Extensive experiments on the DAIC-WOZ dataset using two advanced acoustic backbones demonstrate that our framework not only significantly reduces gender bias but also improves overall detection performance compared to existing debiasing strategies.
[798] The Mean-Field Dynamics of Transformers
Philippe Rigollet
Main category: cs.LG
TL;DR: Transformers as interacting particle systems with continuum limits show global clustering phenomena, metastable states, and phase transitions in attention dynamics.
Details
Motivation: To develop a mathematical framework for understanding Transformer attention dynamics through continuum limits and particle system interpretations, connecting to established mathematical theories.Method: Interpret Transformer attention as interacting particle system, study continuum (mean-field) limits, idealize attention continuous on sphere, connect to Wasserstein gradient flows, synchronization models, and mean-shift clustering.
Result: Identifies global clustering phenomenon with asymptotic token clustering after metastable states, analyzes equiangular reduction for exact clustering rates, shows normalization effects on contraction speeds, identifies phase transition for long-context attention.
Conclusion: The framework reveals mechanisms driving representation collapse and regimes preserving expressive multi-cluster structure in deep attention architectures, providing mathematical insights into Transformer dynamics.
Abstract: We develop a mathematical framework that interprets Transformer attention as an interacting particle system and studies its continuum (mean-field) limits. By idealizing attention continuous on the sphere, we connect Transformer dynamics to Wasserstein gradient flows, synchronization models (Kuramoto), and mean-shift clustering. Central to our results is a global clustering phenomenon whereby tokens cluster asymptotically after long metastable states where they are arranged into multiple clusters. We further analyze a tractable equiangular reduction to obtain exact clustering rates, show how commonly used normalization schemes alter contraction speeds, and identify a phase transition for long-context attention. The results highlight both the mechanisms that drive representation collapse and the regimes that preserve expressive, multi-cluster structure in deep attention architectures.
[799] InvisibleInk: High-Utility and Low-Cost Text Generation with Differential Privacy
Vishnu Vinod, Krishna Pillutla, Abhradeep Guha Thakurta
Main category: cs.LG
TL;DR: InvisibleInk is a differentially private long-form text generation framework that reduces computation costs 8x+ compared to SOTA while maintaining utility, enabling practical private text generation at 4-8x non-private cost.
Details
Motivation: As LLM-based long-form text generation advances with RAG and inference-time scaling, safely incorporating private information while maintaining differential privacy remains an open challenge. Current methods are computationally expensive, limiting practical deployment.Method: Frames LLM sampling as exponential mechanism over logits with two innovations: 1) reduces privacy cost by isolating/clipping only sensitive information in model logits relative to public logits, 2) improves text quality by sampling without privacy cost from a small superset of top-k private tokens.
Result: Achieves consistent 8x+ reduction in computation cost over SOTA baselines while maintaining same utility across privacy levels. First framework to generate high-quality private long-form text at less than 4-8x the computation cost of non-private generation.
Conclusion: InvisibleInk enables practical differentially private long-form text generation with rigorous privacy guarantees, making private text generation feasible for real-world applications. The framework is open-sourced as a pip-installable Python package.
Abstract: As major progress in LLM-based long-form text generation enables paradigms such as retrieval-augmented generation (RAG) and inference-time scaling, safely incorporating private information into the generation remains a critical open question. We present InvisibleInk, a highly scalable long-form text generation framework satisfying rigorous differential privacy guarantees with respect to the sensitive reference texts. It interprets sampling from the LLM’s next-token-distribution as the exponential mechanism over the LLM logits with two innovations. First, we reduce the privacy cost by isolating and clipping only the sensitive information in the model logits (relative to the public logits). Second, we improve text quality by sampling without any privacy cost from a small superset of the top-$k$ private tokens. Empirical evaluations demonstrate a consistent $8\times$ (or more) reduction in computation cost over state-of-the-art baselines to generate long-form private text of the same utility across privacy levels. InvisibleInk is able to generate, for the first time, high-quality private long-form text at less than $4$-$8\times$ times the computation cost of non-private generation, paving the way for its practical use. We open-source a pip-installable Python package (invink) for InvisibleInk at https://github.com/cerai-iitm/invisibleink.
[800] Unifying Sign and Magnitude for Optimizing Deep Vision Networks via ThermoLion
Ahmed Nebli
Main category: cs.LG
TL;DR: ThermoLion is a vision optimization framework that dynamically switches between low-bit exploration and high-precision exploitation based on local signal-to-noise ratio, outperforming AdamW and Lion without hyperparameter tuning.
Details
Motivation: Current optimization methods have static limitations: magnitude-based methods (AdamW) assume high-fidelity gradient signals but amplify noise in rugged landscapes, while sign-based methods (Lion) provide robustness but discard fine-grained information. The paper argues optimal convergence requires dynamic modulation of update bitrate rather than static compromises.Method: ThermoLion uses local Signal-to-Noise Ratio (SNR) gating to autonomously transition parameters between “low-bit” exploration and “high-precision” exploitation phases. It also introduces Momentum Alignment mechanism to detect constructive interference between historical drift and instantaneous gradients for accelerated convergence during stable trajectories.
Result: Empirical benchmarks across 12 diverse vision datasets (including CIFAR, SVHN, GTSRB) show ThermoLion surpasses both AdamW and Lion in convergence speed and terminal accuracy without architecture-specific tuning, serving as a hyperparameter-free generalist optimizer.
Conclusion: ThermoLion demonstrates that dynamic modulation of update bitrate based on local SNR is superior to static optimization paradigms, providing both robustness and precision while eliminating the need for hyperparameter tuning in vision tasks.
Abstract: The training of deep vision models is fundamentally a signal recovery problem amidst high-dimensional stochastic noise. Current optimization paradigms impose a static compromise on information channel capacity. For instance, magnitude-based methods, such as AdamW, operate on the assumption that gradient norms are high-fidelity curvature signals. While this allows for precision in smooth regimes, it leads to catastrophic noise amplification when applied to rugged, non-convex landscapes. Conversely, sign-based methods (e.g., Lion) perform a radical 1-bit quantization of the gradient, which aims to provide robust regularization at the cost of discarding fine-grained descent information. We propose that optimal convergence requires neither static prior, but rather a dynamic modulation of the update bitrate. We introduce \textbf{ThermoLion}, a vision-centric framework that utilizes local Signal-to-Noise Ratio (SNR) gating to autonomously transition parameters between a “low-bit” exploration phase and a “high-precision” exploitation phase. Furthermore, we introduce a Momentum Alignment mechanism that detects constructive interference between historical drift and instantaneous gradients to accelerate convergence during stable trajectories. Empirical benchmarks across 12 diverse vision datasets (including CIFAR, SVHN, and GTSRB) demonstrate that ThermoLion serves as a hyperparameter-free generalist, surpassing both AdamW and Lion in convergence speed and terminal accuracy without architecture-specific tuning.
[801] New Spiking Architecture for Multi-Modal Decision-Making in Autonomous Vehicles
Aref Ghoreishee, Abhishek Mishra, Lifeng Zhou, John Walsh, Nagarajan Kandasamy
Main category: cs.LG
TL;DR: Proposes an end-to-end multi-modal RL framework for autonomous vehicles using spiking temporal-aware transformer architecture with ternary spiking neurons for efficient multi-modal fusion.
Details
Motivation: Transformers are effective for multi-modal fusion in autonomous vehicles but have high computational costs that limit deployment in resource-constrained edge environments.Method: End-to-end multi-modal RL framework integrating camera images, LiDAR point clouds, and vehicle heading through a spiking temporal-aware transformer-like architecture using ternary spiking neurons for efficient multi-modal fusion.
Result: Comprehensive evaluations in Highway Environment demonstrate effectiveness and efficiency for real-time autonomous decision-making.
Conclusion: The proposed spiking temporal-aware transformer architecture enables computationally efficient multi-modal fusion suitable for real-time autonomous vehicle decision-making in edge environments.
Abstract: This work proposes an end-to-end multi-modal reinforcement learning framework for high-level decision-making in autonomous vehicles. The framework integrates heterogeneous sensory input, including camera images, LiDAR point clouds, and vehicle heading information, through a cross-attention transformer-based perception module. Although transformers have become the backbone of modern multi-modal architectures, their high computational cost limits their deployment in resource-constrained edge environments. To overcome this challenge, we propose a spiking temporal-aware transformer-like architecture that uses ternary spiking neurons for computationally efficient multi-modal fusion. Comprehensive evaluations across multiple tasks in the Highway Environment demonstrate the effectiveness and efficiency of the proposed approach for real-time autonomous decision-making.
[802] SpikingBrain: Spiking Brain-inspired Large Models
Yuqi Pan, Yupeng Feng, Jinghao Zhuang, Siyu Ding, Han Xu, Zehao Liu, Bohan Sun, Yuhong Chou, Xuerui Qiu, Anlin Deng, Anjie Hu, Shurong Wang, Peng Zhou, Man Yao, Jibin Wu, Jian Yang, Guoliang Sun, Bo Xu, Guoqi Li
Main category: cs.LG
TL;DR: SpikingBrain introduces brain-inspired models for efficient long-context training and inference on non-NVIDIA platforms, achieving comparable performance to Transformers with better efficiency and sparsity.
Details
Motivation: Transformer-based LLMs face quadratic training scaling and linear inference memory growth, limiting long-context processing. Building large models on non-NVIDIA platforms also presents stability and efficiency challenges.Method: Three-pronged approach: (1) Linear/hybrid-linear attention with adaptive spiking neurons, (2) Efficient conversion-based training pipeline and spike coding framework, (3) Customized training frameworks, operator libraries, and parallelism for MetaX hardware.
Result: Developed SpikingBrain-7B (linear) and SpikingBrain-76B (hybrid-linear MoE) models. Achieved stable training on MetaX GPUs with expected MFU. Comparable performance to open-source Transformers using only ~150B tokens. 100x TTFT speedup for 4M-token sequences, 69.15% sparsity for low-power operation.
Conclusion: Brain-inspired mechanisms enable efficient, scalable large model design on non-NVIDIA platforms with constant memory inference, event-driven spiking, and significant long-context efficiency improvements.
Abstract: Mainstream Transformer-based large language models face major efficiency bottlenecks: training computation scales quadratically with sequence length, and inference memory grows linearly, limiting long-context processing. Building large models on non-NVIDIA platforms also poses challenges for stable and efficient training. To address this, we introduce SpikingBrain, a family of brain-inspired models designed for efficient long-context training and inference. SpikingBrain leverages the MetaX GPU cluster and focuses on three aspects: (1) Model Architecture: linear and hybrid-linear attention architectures with adaptive spiking neurons; (2) Algorithmic Optimizations: an efficient, conversion-based training pipeline and a dedicated spike coding framework; (3) System Engineering: customized training frameworks, operator libraries, and parallelism strategies tailored to MetaX hardware. Using these techniques, we develop two models: SpikingBrain-7B, a linear LLM, and SpikingBrain-76B, a hybrid-linear MoE LLM. These models demonstrate the feasibility of large-scale LLM development on non-NVIDIA platforms, and training remains stable for weeks on hundreds of MetaX GPUs with Model FLOPs Utilization at expected levels. SpikingBrain achieves performance comparable to open-source Transformer baselines while using only about 150B tokens for continual pre-training. Our models also significantly improve long-context efficiency and deliver inference with (partially) constant memory and event-driven spiking behavior. For example, SpikingBrain-7B attains over 100x speedup in Time to First Token for 4M-token sequences. Furthermore, the proposed spiking scheme achieves 69.15 percent sparsity, enabling low-power operation. Overall, this work demonstrates the potential of brain-inspired mechanisms to drive the next generation of efficient and scalable large model design.
[803] SVRG and Beyond via Posterior Correction
Nico Daheim, Thomas Möllenhoff, Ming Liang Ang, Mohammad Emtiyaz Khan
Main category: cs.LG
TL;DR: Connects SVRG to Bayesian posterior correction, showing SVRG is a special case of isotropic-Gaussian correction. Proposes new SVRG variants using Gaussian families: Newton-like with Hessian corrections and Adam-like for Transformers.
Details
Motivation: SVRG and variants have limited success in deep learning despite aiming to speed up training. The paper aims to establish foundational connections between SVRG and Bayesian methods to derive more effective variants.Method: Shows SVRG is recovered as special case of posterior correction over isotropic-Gaussian family. Derives new SVRG variants using more flexible Gaussian families: 1) Newton-like variant with Hessian corrections, 2) Adam-like extension for Transformer language models.
Result: First work connecting SVRG to Bayesian methods. Novel SVRG variants improve pretraining and finetuning of Transformer language models. Demonstrates practical benefits of the Bayesian perspective for deep network training.
Conclusion: Establishes foundational connection between SVRG and Bayesian posterior correction. Shows Bayesian perspective enables derivation of novel, more effective SVRG variants for deep learning, particularly benefiting Transformer training.
Abstract: Stochastic Variance Reduced Gradient (SVRG) and its variants aim to speed-up training by using gradient corrections, but have seen limited success in deep learning. Here, we show surprising new foundational connections of SVRG to a recently proposed Bayesian method called posterior correction. Specifically, we show that SVRG is recovered as a special case of posterior correction over the isotropic-Gaussian family, while novel extensions are automatically obtained by using more flexible exponential families. We derive two new SVRG variants by using Gaussian families: First, a Newton-like variant that employs novel Hessian corrections, and second, an Adam-like extension that improves pretraining and finetuning of Transformer language models. This is the first work to connect SVRG to Bayes and use it to boost variational training for deep networks.
[804] Domain-Decomposed Graph Neural Network Surrogate Modeling for Ice Sheets
Adrienne M. Propp, Mauro Perego, Eric C. Cyr, Anthony Gruber, Amanda A. Howard, Alexander Heinlein, Panos Stinis, Daniel M. Tartakovsky
Main category: cs.LG
TL;DR: A physics-inspired graph neural network surrogate with domain decomposition and transfer learning enables efficient, accurate PDE simulations on unstructured meshes, particularly for ice sheet dynamics and uncertainty quantification.
Details
Motivation: Need for accurate yet efficient surrogate models for large-scale PDE simulations, especially for uncertainty quantification tasks requiring hundreds/thousands of evaluations. Current approaches struggle with unstructured meshes and computational efficiency.Method: Develop physics-inspired graph neural network (GNN) surrogate operating directly on unstructured meshes using graph attention. Introduce domain decomposition strategy: partition mesh into subdomains, train local GNN surrogates in parallel, aggregate predictions. Employ transfer learning to fine-tune models across subdomains for improved training efficiency and accuracy.
Result: Accurately predicts full-field velocities on high-resolution meshes for ice sheet simulations. Substantially reduces training time compared to training single global surrogate model. Provides foundation for uncertainty quantification objectives.
Conclusion: Graph-based domain decomposition combined with transfer learning provides scalable, reliable pathway for training GNN surrogates on massive PDE-governed systems, with broad potential beyond ice sheet dynamics applications.
Abstract: Accurate yet efficient surrogate models are essential for large-scale simulations of partial differential equations (PDEs), particularly for uncertainty quantification (UQ) tasks that demand hundreds or thousands of evaluations. We develop a physics-inspired graph neural network (GNN) surrogate that operates directly on unstructured meshes and leverages the flexibility of graph attention. To improve both training efficiency and generalization properties of the model, we introduce a domain decomposition (DD) strategy that partitions the mesh into subdomains, trains local GNN surrogates in parallel, and aggregates their predictions. We then employ transfer learning to fine-tune models across subdomains, accelerating training and improving accuracy in data-limited settings. Applied to ice sheet simulations, our approach accurately predicts full-field velocities on high-resolution meshes, substantially reduces training time relative to training a single global surrogate model, and provides a ripe foundation for UQ objectives. Our results demonstrate that graph-based DD, combined with transfer learning, provides a scalable and reliable pathway for training GNN surrogates on massive PDE-governed systems, with broad potential for application beyond ice sheet dynamics.
[805] Event2Vec: A Geometric Approach to Learning Composable Representations of Event Sequences
Antonin Sulc
Main category: cs.LG
TL;DR: Event2Vec learns interpretable embeddings for discrete event sequences using additive recurrent structure, with both Euclidean and hyperbolic variants for different data geometries.
Details
Motivation: Neural representations in biological and artificial systems show importance of geometric and topological structures. Need for framework to learn composable, interpretable embeddings of discrete event sequences that capture structural dependencies.Method: Simple additive recurrent structure model that learns embeddings where sequence representation equals vector sum of constituent events (linear additive hypothesis). Includes Euclidean version and hyperbolic variant for hierarchical/tree-like data.
Result: Theoretical analysis shows learned representations converge to ideal additive structure. On Brown Corpus, achieves Silhouette score of 0.0564 vs Word2Vec baseline 0.0215, demonstrating ability to capture structural dependencies without supervision.
Conclusion: Event2Vec provides effective framework for learning interpretable, composable embeddings of event sequences, with geometric flexibility (Euclidean/hyperbolic) to match different data structures.
Abstract: The study of neural representations, both in biological and artificial systems, is increasingly revealing the importance of geometric and topological structures. Inspired by this, we introduce Event2Vec, a novel framework for learning representations of discrete event sequences. Our model leverages a simple, additive recurrent structure to learn composable, interpretable embeddings. We provide a theoretical analysis demonstrating that, under specific training objectives, our model’s learned representations in a Euclidean space converge to an ideal additive structure. This ensures that the representation of a sequence is the vector sum of its constituent events, a property we term the linear additive hypothesis. To address the limitations of Euclidean geometry for hierarchical data, we also introduce a variant of our model in hyperbolic space, which is naturally suited to embedding tree-like structures with low distortion. We present experiments to validate our hypothesis. Quantitative evaluation on the Brown Corpus yields a Silhouette score of 0.0564, outperforming a Word2Vec baseline (0.0215), demonstrating the model’s ability to capture structural dependencies without supervision.
[806] Elastic Weight Consolidation for Knowledge Graph Continual Learning: An Empirical Evaluation
Gaganpreet Jhajj, Fuhua Lin
Main category: cs.LG
TL;DR: EWC reduces catastrophic forgetting in KG continual learning by 45.7% compared to naive sequential training, with task partitioning strategy significantly affecting forgetting magnitude.
Details
Motivation: Knowledge graphs need continual updates as new information emerges, but neural embedding models suffer from catastrophic forgetting when learning new tasks sequentially, requiring effective continual learning methods.Method: Evaluated Elastic Weight Consolidation (EWC) regularization method on KG link prediction using TransE embeddings on FB15k-237 dataset, with multiple experiments across five random seeds and different task partitioning strategies (relation-based vs random).
Result: EWC reduced catastrophic forgetting from 12.62% to 6.85% (45.7% reduction). Relation-based partitioning showed 9.8 percentage points higher forgetting than random partitioning (12.62% vs 2.81%), indicating task construction influences evaluation outcomes.
Conclusion: EWC effectively mitigates catastrophic forgetting in KG continual learning, and evaluation protocol design (especially task partitioning strategy) significantly impacts measured forgetting, highlighting the importance of careful experimental design.
Abstract: Knowledge graphs (KGs) require continual updates as new information emerges, but neural embedding models suffer from catastrophic forgetting when learning new tasks sequentially. We evaluate Elastic Weight Consolidation (EWC), a regularization-based continual learning method, on KG link prediction using TransE embeddings on FB15k-237. Across multiple experiments with five random seeds, we find that EWC reduces catastrophic forgetting from 12.62% to 6.85%, a 45.7% reduction compared to naive sequential training. We observe that the task partitioning strategy affects the magnitude of forgetting: relation-based partitioning (grouping triples by relation type) exhibits 9.8 percentage points higher forgetting than randomly partitioned tasks (12.62% vs 2.81%), suggesting that task construction influences evaluation outcomes. While focused on a single embedding model and dataset, our results demonstrate that EWC effectively mitigates catastrophic forgetting in KG continual learning and highlight the importance of evaluation protocol design.
[807] Adaptive Margin RLHF via Preference over Preferences
Yaswanth Chittepu, Prasann Singhal, Greg Durrett, Scott Niekum
Main category: cs.LG
TL;DR: DPO-PoP extends Direct Preference Optimization by using preference-over-preference annotations to infer adaptive margins, improving both discriminative and generative performance over fixed-margin approaches.
Details
Motivation: Existing RLHF methods use fixed or simplistic margins that don't account for varying preference strengths, and rely on noisy rating-based margins. Modeling preference strength could improve generalization and alignment, but accurate preference scores are hard for humans to provide reliably.Method: DPO-PoP extends DPO by using preference-over-preference annotations (which preference reflects a stronger distinction) to infer adaptive margins on a per-datapoint basis. Two sampling strategies are proposed: one favoring discriminative performance and one favoring generative performance.
Result: Outperforms vanilla DPO, DPO with fixed margins, and DPO with ground-truth margins on UltraFeedback dataset. Shows tradeoff between discriminative and generative performance: improving test classification accuracy (especially on weaker preferences) can reduce generative quality.
Conclusion: Preference-over-preference supervision enables adaptive margins that improve RLHF performance. The tradeoff between discriminative and generative performance can be navigated through strategic sampling of preference-over-preference labels.
Abstract: Margin-based optimization is fundamental to improving generalization and robustness in classification tasks. In the context of reward model learning from preferences within Reinforcement Learning from Human Feedback (RLHF), existing methods typically rely on no margins, fixed margins, or margins that are simplistic functions of preference ratings. However, such formulations often fail to account for the varying strengths of different preferences, for example some preferences are associated with larger margins between responses, or they rely on noisy margin information derived from ratings. We argue that modeling the strength of preferences can lead to better generalization and more faithful alignment. Furthermore, many existing methods that use adaptive margins assume access to accurate preference scores, which can be difficult for humans to provide reliably. We propose an approach that leverages preferences over preferences, that is annotations indicating which of two preferences reflects a stronger distinction. We use this ordinal signal to infer adaptive margins on a per-datapoint basis. We introduce an extension to Direct Preference Optimization (DPO), DPO-PoP, that incorporates adaptive margins from preference-over-preference supervision, enabling improved discriminative and generative performance. Empirically, our method outperforms vanilla DPO, DPO with fixed margins, and DPO with ground-truth margins on the UltraFeedback dataset. Additionally, we show that there is a tradeoff between discriminative and generative performance: improving test classification accuracy, particularly by correctly labeling weaker preferences at the expense of stronger ones, can lead to a decline in generative quality. To navigate this tradeoff, we propose two sampling strategies to gather preference-over-preference labels: one favoring discriminative performance and one favoring generative performance.
[808] Provably Safe Model Updates
Leo Elmecker-Plakolm, Pierre Fasterling, Philip Sosnin, Calvin Tsay, Matthew Wicker
Main category: cs.LG
TL;DR: A framework for provably safe model updates that computes largest locally invariant domains in parameter space to certify specifications are satisfied during updates, enabling formal safety guarantees in dynamic safety-critical environments.
Details
Motivation: Safety-critical environments are dynamic with distribution shifts, emerging vulnerabilities, and evolving requirements requiring continuous model updates. However, even benign parameter updates can cause unintended consequences like catastrophic forgetting or alignment drift. Existing heuristic approaches cannot certify that updated models continue to satisfy required performance specifications.Method: Formalizes the problem as computing the largest locally invariant domain (LID) - a connected region in parameter space where all points are certified to satisfy given specifications. Uses parameterized abstract domains (orthotopes, zonotopes) for tractable primal-dual formulation, enabling efficient certification of updates independent of data or algorithm. Allows projection of updates onto safe domains, computation of multiple approximately optimal LIDs, incorporation of regularization-inspired biases, and use of lookahead data buffers.
Result: Across continual learning and foundation model fine-tuning benchmarks, the method matches or exceeds heuristic baselines for avoiding forgetting while providing formal safety guarantees.
Conclusion: The framework enables provably safe model updates in dynamic safety-critical environments by computing certified safe parameter regions, addressing limitations of existing heuristic approaches that lack formal guarantees.
Abstract: Safety-critical environments are inherently dynamic. Distribution shifts, emerging vulnerabilities, and evolving requirements demand continuous updates to machine learning models. Yet even benign parameter updates can have unintended consequences, such as catastrophic forgetting in classical models or alignment drift in foundation models. Existing heuristic approaches (e.g., regularization, parameter isolation) can mitigate these effects but cannot certify that updated models continue to satisfy required performance specifications. We address this problem by introducing a framework for provably safe model updates. Our approach first formalizes the problem as computing the largest locally invariant domain (LID): a connected region in parameter space where all points are certified to satisfy a given specification. While exact maximal LID computation is intractable, we show that relaxing the problem to parameterized abstract domains (orthotopes, zonotopes) yields a tractable primal-dual formulation. This enables efficient certification of updates - independent of the data or algorithm used - by projecting them onto the safe domain. Our formulation further allows computation of multiple approximately optimal LIDs, incorporation of regularization-inspired biases, and use of lookahead data buffers. Across continual learning and foundation model fine-tuning benchmarks, our method matches or exceeds heuristic baselines for avoiding forgetting while providing formal safety guarantees.
[809] Delays in Spiking Neural Networks: A State Space Model Approach
Sanja Karilanova, Subhrakanti Dey, Ayça Özçelikkale
Main category: cs.LG
TL;DR: A framework for incorporating learnable delays into spiking neural networks (SNNs) using additional state variables, enabling neurons to access temporal input history and improving performance, especially in smaller networks.
Details
Motivation: SNNs are biologically inspired and energy-efficient for temporal data processing. Delays play a crucial role in capturing complex temporal dependencies by allowing past inputs to influence present spiking behavior, but existing approaches need improvement for computational efficiency and integration flexibility.Method: Proposes a general framework for incorporating delays into SNNs through additional state variables that enable each neuron to access finite temporal input history. The framework is neuron model-agnostic and can be integrated into standard models like LIF and adLIF. Analyzes delay duration and learnable parameters’ effects on performance.
Result: On the Spiking Heidelberg Digits (SHD) dataset, the proposed mechanism matches existing delay-based SNNs’ performance while remaining computationally efficient. Results show delay incorporation substantially improves performance in smaller networks.
Conclusion: The proposed delay framework offers a flexible, efficient way to enhance SNNs’ temporal processing capabilities, particularly benefiting smaller networks while maintaining computational efficiency comparable to existing approaches.
Abstract: Spiking neural networks (SNNs) are biologically inspired, event-driven models that are suitable for processing temporal data and offer energy-efficient computation when implemented on neuromorphic hardware. In SNNs, richer neuronal dynamic allows capturing more complex temporal dependencies, with delays playing a crucial role by allowing past inputs to directly influence present spiking behavior. We propose a general framework for incorporating delays into SNNs through additional state variables. The proposed mechanism enables each neuron to access a finite temporal input history. The framework is agnostic to neuron models and hence can be seamlessly integrated into standard spiking neuron models such as LIF and adLIF. We analyze how the duration of the delays and the learnable parameters associated with them affect the performance. We investigate the trade-offs in the network architecture due to additional state variables introduced by the delay mechanism. Experiments on the Spiking Heidelberg Digits (SHD) dataset show that the proposed mechanism matches the performance of existing delay-based SNNs while remaining computationally efficient. Moreover, the results illustrate that the incorporation of delays may substantially improve performance in smaller networks.
[810] Forecasting in Offline Reinforcement Learning for Non-stationary Environments
Suzan Ece Ada, Georg Martius, Emre Ugur, Erhan Oztop
Main category: cs.LG
TL;DR: FORL is a new offline RL framework that handles non-stationary environments with unexpected time-varying offsets by combining conditional diffusion-based state generation with zero-shot time-series foundation models for robust performance from episode start.
Details
Motivation: Existing offline RL methods assume stationarity or only handle synthetic perturbations, but real-world scenarios often have abrupt, time-varying offsets that create partial observability and degrade agent performance. There's a need for methods that can handle realistic non-stationarity without assuming specific patterns of future changes.Method: FORL unifies two components: (1) conditional diffusion-based candidate state generation trained without assuming specific patterns of future non-stationarity, and (2) zero-shot time-series foundation models. The framework targets environments with unexpected, potentially non-Markovian offsets and requires robust performance from the beginning of each episode.
Result: Empirical evaluations on offline RL benchmarks augmented with real-world time-series data to simulate realistic non-stationarity show that FORL consistently improves performance compared to competitive baselines.
Conclusion: FORL bridges the gap between offline RL and real-world non-stationary environments by integrating zero-shot forecasting with agent experience, enabling robust performance in scenarios with unexpected time-varying offsets.
Abstract: Offline Reinforcement Learning (RL) provides a promising avenue for training policies from pre-collected datasets when gathering additional interaction data is infeasible. However, existing offline RL methods often assume stationarity or only consider synthetic perturbations at test time, assumptions that often fail in real-world scenarios characterized by abrupt, time-varying offsets. These offsets can lead to partial observability, causing agents to misperceive their true state and degrade performance. To overcome this challenge, we introduce Forecasting in Non-stationary Offline RL (FORL), a framework that unifies (i) conditional diffusion-based candidate state generation, trained without presupposing any specific pattern of future non-stationarity, and (ii) zero-shot time-series foundation models. FORL targets environments prone to unexpected, potentially non-Markovian offsets, requiring robust agent performance from the onset of each episode. Empirical evaluations on offline RL benchmarks, augmented with real-world time-series data to simulate realistic non-stationarity, demonstrate that FORL consistently improves performance compared to competitive baselines. By integrating zero-shot forecasting with the agent’s experience, we aim to bridge the gap between offline RL and the complexities of real-world, non-stationary environments.
[811] A Footprint-Aware, High-Resolution Approach for Carbon Flux Prediction Across Diverse Ecosystems
Jacob Searcy, Anish Dulal, Scott Bridgham, Ashley Cordes, Lillian Aoki, Brendan Bohannan, Qing Zhu, Lucas C. R. Silva
Main category: cs.LG
TL;DR: FAR is a deep learning framework that predicts both spatial footprints and pixel-level carbon flux estimates at 30m resolution, trained on tower-satellite data to improve monitoring of natural climate solutions.
Details
Motivation: Monitoring carbon drawdown from natural climate solutions over large areas is challenging. While flux towers provide ground truth and satellites offer spatial data, there's a scale mismatch - satellites measure at finer scales than tower footprints, creating a gap in accurate upscaling.Method: Footprint-Aware Regression (FAR) - a deep learning framework that simultaneously predicts spatial footprints and pixel-level (30m) carbon flux estimates. Trained on AMERI-FAR25 dataset combining 439 site-years of tower data with corresponding Landsat scenes.
Result: The model produces high-resolution predictions and achieves R² = 0.78 when predicting monthly net ecosystem exchange on test sites from various ecosystems, demonstrating strong performance.
Conclusion: FAR represents a first-of-its-kind approach that bridges the scale gap between flux tower measurements and satellite observations, enabling more accurate monitoring of carbon fluxes for natural climate solutions at high spatial resolution.
Abstract: Natural climate solutions (NCS) offer an approach to mitigating carbon dioxide (CO2) emissions. However, monitoring the carbon drawdown of ecosystems over large geographic areas remains challenging. Eddy-flux covariance towers provide ground truth for predictive ‘upscaling’ models derived from satellite products, but many satellites now produce measurements on spatial scales smaller than a flux tower’s footprint. We introduce Footprint-Aware Regression (FAR), a first-of-its-kind, deep-learning framework that simultaneously predicts spatial footprints and pixel-level (30 m scale) estimates of carbon flux. FAR is trained on our AMERI-FAR25 dataset which combines 439 site years of tower data with corresponding Landsat scenes. Our model produces high-resolution predictions and achieves R2 = 0.78 when predicting monthly net ecosystem exchange on test sites from a variety of ecosystems.
[812] Influence Functions for Efficient Data Selection in Reasoning
Prateek Humane, Paolo Cudrano, Daniel Z. Kaplan, Matteo Matteucci, Supriyo Chakraborty, Irina Rish
Main category: cs.LG
TL;DR: Fine-tuning LLMs on CoT data shows small high-quality datasets beat massive ones, but “quality” is ill-defined. The paper proposes using influence functions to measure causal effects of CoT examples on downstream accuracy, introducing influence-based pruning that outperforms perplexity and embedding baselines on math reasoning.
Details
Motivation: Current methods for selecting high-quality chain-of-thought (CoT) data rely on indirect heuristics like problem difficulty or trace length, while instruction-tuning has explored automated selection strategies but rarely in reasoning contexts. There's a need for a principled definition of reasoning data quality.Method: Proposes using influence functions to measure the causal effect of individual CoT examples on downstream accuracy. Introduces influence-based pruning for selecting high-quality reasoning data, which directly measures how each training example affects model performance.
Result: Influence-based pruning consistently outperforms perplexity and embedding-based baselines on math reasoning tasks within a model family, demonstrating the effectiveness of using causal influence as a quality metric for CoT data.
Conclusion: Influence functions provide a principled way to define reasoning data quality by measuring causal effects of training examples, offering a superior approach to existing heuristic-based methods for selecting high-quality CoT data for LLM fine-tuning.
Abstract: Fine-tuning large language models (LLMs) on chain-of-thought (CoT) data shows that a small amount of high-quality data can outperform massive datasets. Yet, what constitutes “quality” remains ill-defined. Existing reasoning methods rely on indirect heuristics such as problem difficulty or trace length, while instruction-tuning has explored a broader range of automated selection strategies, but rarely in the context of reasoning. We propose to define reasoning data quality using influence functions, which measure the causal effect of individual CoT examples on downstream accuracy, and introduce influence-based pruning, which consistently outperforms perplexity and embedding-based baselines on math reasoning within a model family.
[813] KV Pareto: Systems-Level Optimization of KV Cache and Model Compression for Long Context Inference
Sai Gokhale, Devleena Das, Rajeev Patwari, Ashish Sirasao, Elliott Delaye
Main category: cs.LG
TL;DR: KV Pareto is a framework that systematically explores the trade-off frontier between memory consumption and accuracy by jointly optimizing KV cache quantization, chunked prefill, and weight quantization for efficient long-context LLM inference on edge devices.
Details
Motivation: Long-context LLMs face significant memory bottlenecks during inference due to linear growth of KV cache with sequence length. While individual optimization techniques exist, their joint effects and optimal configurations for edge deployment remain underexplored.Method: KV Pareto systematically maps the trade-off frontier between total memory consumption and task accuracy across three complementary optimization techniques: KV cache quantization (int2/4/8, mixed-precision, various granularities), chunked prefill, and 4-bit weight quantization via AWQ. Evaluates multiple LLM architectures (Qwen, Llama, Mistral).
Result: Identifies model-specific Pareto-optimal configurations achieving 68-78% total memory reduction with minimal (1-3%) accuracy degradation on long-context tasks. Verified on additional benchmarks (Needle-in-a-Haystack, GSM8k, MMLU) and extended context lengths up to 128k.
Conclusion: Joint optimization of multiple techniques is practically necessary for efficient LLM inference, with KV Pareto providing a systematic framework to identify optimal configurations that balance memory efficiency and accuracy for edge deployment.
Abstract: Long-context Large Language Models (LLMs) face significant memory bottlenecks during inference due to the linear growth of key-value (KV) cache with sequence length. While individual optimization techniques like KV cache quantization, chunked prefill, and model weight quantization have shown promise, their joint effects and optimal configurations for edge deployment remain underexplored. We introduce KV Pareto, a systems-level framework that systematically maps the trade-off frontier between total memory consumption and task accuracy across these three complementary optimization techniques. Our framework evaluates multiple LLM architectures (Qwen, Llama, Mistral) with varying KV quantization schemes (int2/4/8, mixed-precision), granularities (per-token, per-tensor, per-block), and 4-bit weight quantization via AWQ. Our framework identifies model-specific Pareto-optimal configurations that achieve 68-78% total memory reduction with minimal (1-3%) accuracy degradation on long-context tasks. We additionally verify the selected frontiers on additional benchmarks of Needle-in-a-Haystack, GSM8k and MMLU as well as extended context lengths of up to 128k to demonstrate the practical need of joint optimization for efficient LLM inference.
[814] Low-Rank Prehab: Preparing Neural Networks for SVD Compression
Haoran Qin, Shansita Sharma, Ali Abbasi, Chayne Thrash, Soheil Kolouri
Main category: cs.LG
TL;DR: Low-Rank Prehab introduces pre-compression fine-tuning to encourage low-rank structure in neural network weights, enabling better SVD-based compression with less accuracy drop and improved recovery.
Details
Motivation: Current SVD-based compression methods cause significant accuracy drops after compression, requiring extensive fine-tuning to recover. The paper is inspired by surgical prehabilitation - preparing the model before compression rather than just fixing it afterward.Method: Low-Rank Prehab adds a pre-compression fine-tuning stage that explicitly encourages low-rank structure in weight matrices while preserving task performance. This steers weights toward spectrally compact regions, enabling smoother low-rank approximation.
Result: Experiments on LLMs and Vision Transformers show Prehab substantially reduces immediate accuracy drop after compression and consistently improves post-finetuning performance across various compression ratios, outperforming state-of-the-art SVD-based techniques like SVD-LLM.
Conclusion: Preparing models for compression through pre-compression conditioning (Prehab) is more effective than only improving compression and recovery stages, highlighting the importance of model preparation before SVD-based compression.
Abstract: Low-rank approximation methods such as singular value decomposition (SVD) and its variants (e.g., Fisher-weighted SVD, Activation SVD) have recently emerged as effective tools for neural network compression. In this setting, decomposition acts as a “surgical” intervention, followed by fine-tuning that serves as “rehab” to recover accuracy. Inspired by prehabilitation in surgery, we introduce a pre-compression fine-tuning stage, Low-Rank Prehab, that explicitly encourages low-rank structure in weight matrices while preserving task performance. By conditioning the model before SVD, Prehab steers weights toward spectrally compact regions of the parameter space, enabling smoother low-rank approximation and improved recovery. Experiments on large language models (LLMs) and other Transformer-based architectures, including Vision Transformers (ViTs), show that Prehab substantially reduces the immediate accuracy drop after compression and consistently improves post-finetuning performance. Across a wide range of compression ratios, our method outperforms state-of-the-art SVD-based techniques such as SVD-LLM, highlighting the importance of preparing models for compression rather than only improving the compression and recovery stages. Source code is available at https://github.com/niqretnuh/PREHAB-SVD
[815] Feature-Based Semantics-Aware Scheduling for Energy-Harvesting Federated Learning
Eunjeong Jeong, Giovanni Perin, Howard H. Yang, Nikolaos Pappas
Main category: cs.LG
TL;DR: Proposes a lightweight client scheduling framework for Energy-Harvesting Federated Learning using Version Age of Information (VAoI) with a feature-based proxy to reduce computational costs, achieving better learning performance and energy reduction in resource-constrained edge scenarios.
Details
Motivation: Federated Learning on resource-constrained edge devices faces challenges where computational energy for training DNNs dominates communication costs. Existing Energy-Harvesting FL strategies fail to account for this reality, leading to wasted energy from redundant local computations.Method: Proposes a client scheduling framework using Version Age of Information (VAoI) to quantify update timeliness and significance. Introduces a feature-based proxy that estimates model redundancy using intermediate-layer extraction from a single forward pass, dramatically reducing VAoI’s typical prohibitive computational cost.
Result: Experiments under extreme non-IID data distributions and scarce energy availability demonstrate superior learning performance while achieving energy reduction compared to existing baseline selection policies.
Conclusion: Establishes semantics-aware scheduling as a practical and vital solution for Energy-Harvesting FL in realistic scenarios where training costs dominate transmission costs, with the feature-based proxy making VAoI computationally feasible for resource-constrained devices.
Abstract: Federated Learning (FL) on resource-constrained edge devices faces a critical challenge: The computational energy required for training Deep Neural Networks (DNNs) often dominates communication costs. However, most existing Energy-Harvesting FL (EHFL) strategies fail to account for this reality, resulting in wasted energy due to redundant local computations. For efficient and proactive resource management, algorithms that predict local update contributions must be devised. We propose a lightweight client scheduling framework using the Version Age of Information (VAoI), a semantics-aware metric that quantifies update timeliness and significance. Crucially, we overcome VAoI’s typical prohibitive computational cost, which requires statistical distance over the entire parameter space, by introducing a feature-based proxy. This proxy estimates model redundancy using intermediate-layer extraction from a single forward pass, dramatically reducing computational complexity. Experiments conducted under extreme non-IID data distributions and scarce energy availability demonstrate superior learning performance while achieving energy reduction compared to existing baseline selection policies. Our framework establishes semantics-aware scheduling as a practical and vital solution for EHFL in realistic scenarios where training costs dominate transmission costs.
[816] Combining Textual and Structural Information for Premise Selection in Lean
Job Petrovčič, David Eliecer Narvaez Denis, Ljupčo Todorovski
Main category: cs.LG
TL;DR: Graph-augmented approach combining text embeddings with GNNs over dependency graphs outperforms language-only methods for premise selection in theorem proving by over 25%.
Details
Motivation: Premise selection is a key bottleneck for scaling theorem proving in large formal libraries. Existing language-based methods treat premises in isolation, ignoring the web of dependencies that connects them, which limits their effectiveness.Method: A graph-augmented approach that combines dense text embeddings of Lean formalizations with graph neural networks over a heterogeneous dependency graph. The graph captures both state-premise and premise-premise relations, allowing the model to leverage relational information between premises.
Result: On the LeanDojo Benchmark, the method outperforms the ReProver language-based baseline by over 25% across standard retrieval metrics, demonstrating significant improvement in premise selection performance.
Conclusion: Relational information captured through dependency graphs is beneficial for premise selection in theorem proving, and graph-augmented approaches combining text embeddings with GNNs can substantially improve over language-only methods.
Abstract: Premise selection is a key bottleneck for scaling theorem proving in large formal libraries. Yet existing language-based methods often treat premises in isolation, ignoring the web of dependencies that connects them. We present a graph-augmented approach that combines dense text embeddings of Lean formalizations with graph neural networks over a heterogeneous dependency graph capturing both state-premise and premise-premise relations. On the LeanDojo Benchmark, our method outperforms the ReProver language-based baseline by over 25% across standard retrieval metrics. These results suggest that relational information is beneficial for premise selection.
[817] A Diffusion Model Framework for Maximum Entropy Reinforcement Learning
Sebastian Sanokowski, Kaustubh Patil, Alois Knoll
Main category: cs.LG
TL;DR: The paper proposes diffusion-based variants of popular RL algorithms (SAC, PPO, WPO) by reinterpreting MaxEntRL as a diffusion sampling problem, achieving better performance on continuous control tasks.
Details
Motivation: Diffusion models have shown success in complex sampling tasks, and the authors aim to leverage this capability to improve Maximum Entropy Reinforcement Learning by framing it as a diffusion-based sampling problem.Method: Reinterpret MaxEntRL as diffusion model sampling, minimize reverse KL divergence between diffusion policy and optimal policy using tractable upper bound, apply policy gradient theorem to derive modified surrogate objective, creating diffusion-based variants DiffSAC, DiffPPO, DiffWPO with minimal implementation changes.
Result: DiffSAC, DiffPPO, and DiffWPO achieve better returns and higher sample efficiency than standard SAC and PPO on continuous control benchmarks.
Conclusion: The diffusion-based reinterpretation of MaxEntRL provides a principled way to incorporate diffusion dynamics into RL algorithms, leading to improved performance with minimal implementation overhead.
Abstract: Diffusion models have achieved remarkable success in data-driven learning and in sampling from complex, unnormalized target distributions. Building on this progress, we reinterpret Maximum Entropy Reinforcement Learning (MaxEntRL) as a diffusion model-based sampling problem. We tackle this problem by minimizing the reverse Kullback-Leibler (KL) divergence between the diffusion policy and the optimal policy distribution using a tractable upper bound. By applying the policy gradient theorem to this objective, we derive a modified surrogate objective for MaxEntRL that incorporates diffusion dynamics in a principled way. This leads to simple diffusion-based variants of Soft Actor-Critic (SAC), Proximal Policy Optimization (PPO) and Wasserstein Policy Optimization (WPO), termed DiffSAC, DiffPPO and DiffWPO. All of these methods require only minor implementation changes to their base algorithm. We find that on standard continuous control benchmarks, DiffSAC, DiffPPO and DiffWPO achieve better returns and higher sample efficiency than SAC and PPO.
[818] Measuring Chain-of-Thought Monitorability Through Faithfulness and Verbosity
Austin Meek, Eitan Sprejer, Iván Arcuschin, Austin J. Brockmeier, Steven Basart
Main category: cs.LG
TL;DR: The paper introduces a new framework for evaluating chain-of-thought (CoT) reasoning by combining faithfulness and verbosity into a monitorability score, showing that models can appear faithful but remain hard to monitor when they omit key reasoning factors.
Details
Motivation: Current methods for evaluating CoT faithfulness focus on cases where models change answers after adding cues, but this approach loses information when models maintain answers and doesn't examine reasoning aspects unrelated to cues. There's a need for more holistic monitoring of CoT reasoning for safety applications.Method: The authors introduce verbosity as a measure of whether CoT lists every factor needed to solve a task, combining it with faithfulness into a single monitorability score. They evaluate instruction-tuned and reasoning models on BBH, GPQA, and MMLU datasets using their framework.
Result: Models can appear faithful yet remain hard to monitor when they leave out key factors. Monitorability differs sharply across model families, revealing limitations in current CoT evaluation methods.
Conclusion: The proposed monitorability score provides a more comprehensive evaluation of CoT reasoning quality, which is crucial for safety schemes relying on CoT monitoring. The authors release evaluation code using the Inspect library to support reproducible research.
Abstract: Chain-of-thought (CoT) outputs let us read a model’s step-by-step reasoning. Since any long, serial reasoning process must pass through this textual trace, the quality of the CoT is a direct window into what the model is thinking. This visibility could help us spot unsafe or misaligned behavior (monitorability), but only if the CoT is transparent about its internal reasoning (faithfulness). Fully measuring faithfulness is difficult, so researchers often focus on examining the CoT in cases where the model changes its answer after adding a cue to the input. This proxy finds some instances of unfaithfulness but loses information when the model maintains its answer, and does not investigate aspects of reasoning not tied to the cue. We extend these results to a more holistic sense of monitorability by introducing verbosity: whether the CoT lists every factor needed to solve the task. We combine faithfulness and verbosity into a single monitorability score that shows how well the CoT serves as the model’s external `working memory’, a property that many safety schemes based on CoT monitoring depend on. We evaluate instruction-tuned and reasoning models on BBH, GPQA, and MMLU. Our results show that models can appear faithful yet remain hard to monitor when they leave out key factors, and that monitorability differs sharply across model families. We release our evaluation code using the Inspect library to support reproducible future work.
[819] Soft Adaptive Policy Optimization
Chang Gao, Chujie Zheng, Xiong-Hui Chen, Kai Dang, Shixuan Liu, Bowen Yu, An Yang, Shuai Bai, Jingren Zhou, Junyang Lin
Main category: cs.LG
TL;DR: SAPO (Soft Adaptive Policy Optimization) replaces hard clipping in RL for LLMs with smooth temperature-controlled gating, improving stability and performance compared to existing methods like GSPO and GRPO.
Details
Motivation: RL is crucial for enhancing LLM reasoning but suffers from unstable policy optimization due to high variance in token-level importance ratios, especially in Mixture-of-Experts models. Existing methods use hard clipping which makes it difficult to balance stability and effective learning.Method: SAPO replaces hard clipping with a smooth, temperature-controlled gate that adaptively attenuates off-policy updates while preserving useful learning signals. It’s both sequence-coherent and token-adaptive, forming a continuous trust region that avoids brittle hard clipping.
Result: Empirical results on mathematical reasoning benchmarks show improved training stability and higher Pass@1 performance under comparable budgets. SAPO also yields consistent performance gains when training Qwen3-VL model series across diverse tasks and model sizes.
Conclusion: SAPO provides a more reliable, scalable, and effective optimization strategy for RL training of LLMs, addressing the stability-performance trade-off in policy optimization.
Abstract: Reinforcement learning (RL) plays an increasingly important role in enhancing the reasoning capabilities of large language models (LLMs), yet stable and performant policy optimization remains challenging. Token-level importance ratios often exhibit high variance-a phenomenon exacerbated in Mixture-of-Experts models-leading to unstable updates. Existing group-based policy optimization methods, such as GSPO and GRPO, alleviate this problem via hard clipping, making it difficult to maintain both stability and effective learning. We propose Soft Adaptive Policy Optimization (SAPO), which replaces hard clipping with a smooth, temperature-controlled gate that adaptively attenuates off-policy updates while preserving useful learning signals. Compared with GSPO and GRPO, SAPO is both sequence-coherent and token-adaptive. Like GSPO, SAPO maintains sequence-level coherence, but its soft gating forms a continuous trust region that avoids the brittle hard clipping band used in GSPO. When a sequence contains a few highly off-policy tokens, GSPO suppresses all gradients for that sequence, whereas SAPO selectively down-weights only the offending tokens and preserves the learning signal from the near-on-policy ones, improving sample efficiency. Relative to GRPO, SAPO replaces hard token-level clipping with smooth, temperature-controlled scaling, enabling more informative and stable updates. Empirical results on mathematical reasoning benchmarks indicate that SAPO exhibits improved training stability and higher Pass@1 performance under comparable training budgets. Moreover, we employ SAPO to train the Qwen3-VL model series, demonstrating that SAPO yields consistent performance gains across diverse tasks and different model sizes. Overall, SAPO provides a more reliable, scalable, and effective optimization strategy for RL training of LLMs.
[820] From Topology to Retrieval: Decoding Embedding Spaces with Unified Signatures
Florian Rottach, William Rudman, Bastian Rieck, Harrisen Scells, Carsten Eickhoff
Main category: cs.LG
TL;DR: The paper introduces Unified Topological Signatures (UTS), a framework for analyzing text embedding spaces using topological and geometric measures, showing it can predict model properties and link structure to retrieval performance.
Details
Motivation: To better understand how text embeddings are organized in space to enhance model interpretability and uncover factors driving downstream task performance, as current individual metrics often fail to sufficiently differentiate embedding spaces.Method: Conducted comprehensive analysis of topological and geometric measures across various text embedding models and datasets, then introduced Unified Topological Signatures (UTS) - a holistic framework for characterizing embedding spaces that combines multiple measures.
Result: Found high redundancy among existing measures, showed UTS can predict model-specific properties and reveal architecture-driven similarities, demonstrated UTS can link topological structure to ranking effectiveness and accurately predict document retrievability.
Conclusion: A holistic, multi-attribute perspective is essential for understanding and leveraging the geometry of text embeddings, and the proposed UTS framework provides an effective approach for this analysis.
Abstract: Studying how embeddings are organized in space not only enhances model interpretability but also uncovers factors that drive downstream task performance. In this paper, we present a comprehensive analysis of topological and geometric measures across a wide set of text embedding models and datasets. We find a high degree of redundancy among these measures and observe that individual metrics often fail to sufficiently differentiate embedding spaces. Building on these insights, we introduce Unified Topological Signatures (UTS), a holistic framework for characterizing embedding spaces. We show that UTS can predict model-specific properties and reveal similarities driven by model architecture. Further, we demonstrate the utility of our method by linking topological structure to ranking effectiveness and accurately predicting document retrievability. We find that a holistic, multi-attribute perspective is essential to understanding and leveraging the geometry of text embeddings.
[821] Can-SAVE: Deploying Low-Cost and Population-Scale Cancer Screening via Survival Analysis Variables and EHR
Petr Philonenko, Vladimir Kokh, Pavel Blinov
Main category: cs.LG
TL;DR: Can-SAVE is a lightweight AI system that uses medical history events to rank population-wide cancer risks, achieving 4-10x higher detection rates than baselines and nearly doubling cancer detection in real-world deployments.
Details
Motivation: Conventional cancer screening methods are costly, labor-intensive, and difficult to scale. Most AI systems rely on complex medical data, making them impractical for large-scale population screening.Method: Integrates survival model outputs into a gradient-boosting framework to detect subtle, long-term patient risk patterns from medical history events alone, without requiring complex medical data.
Result: Achieved 4-10x higher detection rate at identical screening volumes (AP 0.228 vs 0.193 baseline) in retrospective study of 1.9M patients. In prospective pilot (426K patients), nearly doubled cancer detection rate (+91%) and increased population coverage by 36% over national protocol.
Conclusion: Can-SAVE demonstrates practical scalability (1M patients in under 3 hours), achieves nationally significant cancer detection improvements within real-world healthcare constraints, and offers immediate clinical utility with a replicable framework for population-wide screening.
Abstract: Conventional medical cancer screening methods are costly, labor-intensive, and extremely difficult to scale. Although AI can improve cancer detection, most systems rely on complex or specialized medical data, making them impractical for large-scale screening. We introduce Can-SAVE, a lightweight AI system that ranks population-wide cancer risks solely based on medical history events. By integrating survival model outputs into a gradient-boosting framework, our approach detects subtle, long-term patient risk patterns - often well before clinical symptoms manifest. Can-SAVE was rigorously evaluated on a real-world dataset of 2.5 million adults spanning five Russian regions, marking the study as one of the largest and most comprehensive deployments of AI-driven cancer risk assessment. In a retrospective oncologist-supervised study over 1.9M patients, Can-SAVE achieves a 4-10x higher detection rate at identical screening volumes and an Average Precision (AP) of 0.228 vs. 0.193 for the best baseline (LoRA-tuned Qwen3-Embeddings via DeepSeek-R1 summarization). In a year-long prospective pilot (426K patients), our method almost doubled the cancer detection rate (+91%) and increased population coverage by 36% over the national screening protocol. The system demonstrates practical scalability: a city-wide population of 1 million patients can be processed in under three hours using standard hardware, enabling seamless clinical integration. This work proves that Can-SAVE achieves nationally significant cancer detection improvements while adhering to real-world public healthcare constraints, offering immediate clinical utility and a replicable framework for population-wide screening. Code for training and feature engineering is available at https://github.com/sb-ai-lab/Can-SAVE.
[822] Resource-efficient Layer-wise Federated Self-supervised Learning
Ye Lin Tun, Chu Myaet Thwal, Huy Q. Le, Minh N. H. Nguyen, Eui-Nam Huh, Choong Seon Hong
Main category: cs.LG
TL;DR: LW-FedSSL reduces computational and communication costs in federated self-supervised learning by training only specific model layers at each stage, achieving significant resource savings while maintaining performance.
Details
Motivation: Edge devices face high computational and communication costs when running federated learning with self-supervised learning, especially with complex models. These challenges need to be addressed to make FL-SSL practical on resource-constrained devices.Method: Proposes Layer-Wise Federated Self-Supervised Learning (LW-FedSSL) that decomposes training into multiple stages, each focusing on only a specific layer of the model. Also explores Progressive Federated Self-Supervised Learning (Prog-FedSSL) as an alternative progressive training strategy.
Result: LW-FedSSL achieves up to 3.34× reduction in memory usage, 4.20× fewer computational operations (GFLOPs), and 5.07× lower communication cost while maintaining comparable performance to end-to-end training. Prog-FedSSL offers 1.84× reduction in GFLOPs and 1.67× reduction in communication costs with same memory requirements.
Conclusion: LW-FedSSL effectively addresses both computational and communication challenges in FL-SSL on edge devices. While Prog-FedSSL has lower resource efficiency than LW-FedSSL, its performance improvements make it suitable for FL environments with more lenient resource constraints.
Abstract: Many studies integrate federated learning (FL) with self-supervised learning (SSL) to take advantage of raw data distributed across edge devices. However, edge devices often struggle with high computational and communication costs imposed by SSL and FL algorithms. With the deployment of more complex and large-scale models, these challenges are exacerbated. To tackle this, we propose Layer-Wise Federated Self-Supervised Learning (LW-FedSSL), which allows edge devices to incrementally train a small part of the model at a time. Specifically, in LW-FedSSL, training is decomposed into multiple stages, with each stage responsible for only a specific layer of the model. Since only a portion of the model is active for training at any given time, LW-FedSSL significantly reduces computational requirements. Additionally, only the active model portion needs to be exchanged between the FL server and clients, reducing communication overhead. This enables LW-FedSSL to jointly address both computational and communication challenges of FL client devices. It can achieve up to a $3.34 \times$ reduction in memory usage, $4.20 \times$ fewer computational operations (giga floating point operations, GFLOPs), and a $5.07 \times$ lower communication cost while maintaining performance comparable to its end-to-end training counterpart. Furthermore, we explore a progressive training strategy called Progressive Federated Self-Supervised Learning (Prog-FedSSL), which offers a $1.84\times$ reduction in GFLOPs and a $1.67\times$ reduction in communication costs while maintaining the same memory requirements as end-to-end training. Although the resource efficiency of Prog-FedSSL is lower than that of LW-FedSSL, its performance improvements make it a viable candidate for FL environments with more lenient resource constraints.
[823] TLoRA: Tri-Matrix Low-Rank Adaptation of Large Language Models
Tanvir Islam
Main category: cs.LG
TL;DR: TLoRA is a tri-matrix low-rank adaptation method that uses two fixed random matrices and one trainable matrix with layer-wise scaling, achieving comparable performance to LoRA with significantly fewer parameters.
Details
Motivation: To develop a more parameter-efficient fine-tuning method for LLMs that reduces computational overhead while maintaining adaptation effectiveness comparable to existing low-rank methods like LoRA and Adapter-based techniques.Method: Decomposes weight updates into three matrices: two fixed random matrices and one trainable matrix, combined with a learnable, layer-wise scaling factor. This tri-matrix design enables efficient parameter adaptation with minimal computational overhead.
Result: TLoRA achieves comparable performance to LoRA and Adapter-based methods on GLUE benchmark while requiring significantly fewer trainable parameters. Analysis shows Gaussian-like weight distributions, stable parameter norms, scaling factor variability across layers, and similar eigenvalue distributions/parameter norms/cosine similarity to LoRA.
Conclusion: TLoRA is a highly efficient and effective fine-tuning method for LLMs that offers significant resource-efficient model adaptation while effectively approximating LoRA’s adaptation behavior.
Abstract: We propose TLoRA, a novel tri-matrix low-rank adaptation method that decomposes weight updates into three matrices: two fixed random matrices and one trainable matrix, combined with a learnable, layer-wise scaling factor. This tri-matrix design enables TLoRA to achieve highly efficient parameter adaptation while introducing minimal additional computational overhead. Through extensive experiments on the GLUE benchmark, we demonstrate that TLoRA achieves comparable performance to existing low-rank methods such as LoRA and Adapter-based techniques, while requiring significantly fewer trainable parameters. Analyzing the adaptation dynamics, we observe that TLoRA exhibits Gaussian-like weight distributions, stable parameter norms, and scaling factor variability across layers, further highlighting its expressive power and adaptability. Additionally, we show that TLoRA closely resembles LoRA in its eigenvalue distributions, parameter norms, and cosine similarity of updates, underscoring its ability to effectively approximate LoRA’s adaptation behavior. Our results establish TLoRA as a highly efficient and effective fine-tuning method for LLMs, offering a significant step forward in resource-efficient model adaptation.
[824] Outcome-based Reinforcement Learning to Predict the Future
Benjamin Turtel, Danny Franklin, Kris Skotheim, Luke Hewitt, Philipp Schoenegger
Main category: cs.LG
TL;DR: RLVR improves LLM reasoning for real-world event forecasting using prediction market data, achieving better accuracy than frontier models with good calibration and 10% ROI in trading simulations.
Details
Motivation: To apply Reinforcement Learning with Verifiable Rewards (RLVR) to real-world event forecasting, which is challenging due to noisy and delayed outcomes, using prediction market data to improve LLM reasoning capabilities.Method: Used a 14B reasoning model trained on prediction market questions with relevant news headlines. Applied RLVR methods, augmented training with synthetic prediction questions, implemented guardrails for learning stability, and used median prediction sampling at inference time.
Result: The model matched or surpassed predictive accuracy of frontier models like o1, greatly improved probabilistic calibration, and achieved over 10% return on investment in Polymarket trading simulations across all test questions.
Conclusion: RLVR is effective for real-world event forecasting, enabling compact models to achieve competitive performance with good calibration and practical trading value, demonstrating the potential of RL methods in noisy outcome domains.
Abstract: Reinforcement Learning with Verifiable Rewards (RLVR) has been an effective approach for improving Large Language Models’ reasoning in domains such as coding and mathematics. Here, we apply RLVR methods towards forecasting future real-world events - a challenging task for RL due to the very noisy (and delayed) outcomes involved. Using a novel dataset of recent questions from a prediction market, and accompanying relevant news headlines, we show that a compact (14B) reasoning model can be trained to match or surpass the predictive accuracy of frontier models like o1, while greatly improving probabilistic calibration. The model’s performance is also practically meaningful: in a Polymarket trading simulation, we estimate that its bets would have yielded a return on investment of over 10% across all questions in the test set. We detail and compare approaches used in training our model, including augmenting our training-data with synthetic prediction questions, guardrails for learning stability, and median prediction sampling at inference-time.
[825] LORE: Lagrangian-Optimized Robust Embeddings for Visual Encoders
Borna Khodabandeh, Amirabbas Afzali, Amirhossein Afsharrad, Seyed Shahabeddin Mousavi, Sanjay Lall, Sajjad Amini, Seyed-Mohsen Moosavi-Dezfooli
Main category: cs.LG
TL;DR: LORE is an unsupervised adversarial fine-tuning framework that uses constrained optimization to improve visual encoder robustness while preserving clean data accuracy through embedding-space proximity constraints.
Details
Motivation: Current adversarial fine-tuning methods for visual encoders suffer from instability during early fine-tuning stages and exhibit suboptimal trade-offs between robustness and clean data accuracy, limiting their practical utility.Method: LORE (Lagrangian-Optimized Robust Embeddings) uses constrained optimization with embedding-space proximity constraints to balance robustness improvement with preservation of nominal performance during unsupervised adversarial fine-tuning.
Result: LORE significantly improves zero-shot adversarial robustness with minimal degradation in clean data accuracy, and the adversarially fine-tuned CLIP encoder shows effectiveness in out-of-distribution generalization and improved interpretability of image embeddings.
Conclusion: LORE provides a principled approach to adversarial fine-tuning that overcomes stability issues and achieves better robustness-accuracy trade-offs, making it a promising framework for robust visual encoders.
Abstract: Visual encoders have become fundamental components in modern computer vision pipelines. However, ensuring robustness against adversarial perturbations remains a critical challenge. Recent efforts have explored both supervised and unsupervised adversarial fine-tuning strategies. We identify two key limitations in these approaches: (i) they often suffer from instability, especially during the early stages of fine-tuning, resulting in suboptimal convergence and degraded performance on clean data, and (ii) they exhibit a suboptimal trade-off between robustness and clean data accuracy, hindering the simultaneous optimization of both objectives. To overcome these challenges, we propose Lagrangian-Optimized Robust Embeddings (LORE), a novel unsupervised adversarial fine-tuning framework. LORE utilizes constrained optimization, which offers a principled approach to balancing competing goals, such as improving robustness while preserving nominal performance. By enforcing embedding-space proximity constraints, LORE effectively maintains clean data performance throughout adversarial fine-tuning. Extensive experiments show that LORE significantly improves zero-shot adversarial robustness with minimal degradation in clean data accuracy. Furthermore, we demonstrate the effectiveness of the adversarially fine-tuned CLIP image encoder in out-of-distribution generalization and enhancing the interpretability of image embeddings.
[826] Scheduling and Aggregation Design for Asynchronous Federated Learning over Wireless Networks
Chung-Hsuan Hu, Zheng Chen, Erik G. Larsson
Main category: cs.LG
TL;DR: Proposes asynchronous federated learning with periodic aggregation to address straggler issues, featuring channel-aware data-importance scheduling and age-aware aggregation weighting for improved performance.
Details
Motivation: To tackle the straggler problem in synchronous FL systems and address limited wireless communication resources, while reducing bias and variance in aggregated model updates.Method: Asynchronous FL design with periodic aggregation, joint scheduling policy considering both channel quality and training data representation, and age-aware aggregation weighting for asynchronous settings.
Result: Proposed scheduling policy outperforms state-of-the-art synchronous FL methods in simulations, and age-aware aggregation weighting significantly improves learning performance in asynchronous FL.
Conclusion: Asynchronous FL with channel-aware data-importance scheduling and age-aware aggregation weighting effectively addresses straggler issues and improves learning performance in resource-constrained wireless environments.
Abstract: Federated Learning (FL) is a collaborative machine learning (ML) framework that combines on-device training and server-based aggregation to train a common ML model among distributed agents. In this work, we propose an asynchronous FL design with periodic aggregation to tackle the straggler issue in FL systems. Considering limited wireless communication resources, we investigate the effect of different scheduling policies and aggregation designs on the convergence performance. Driven by the importance of reducing the bias and variance of the aggregated model updates, we propose a scheduling policy that jointly considers the channel quality and training data representation of user devices. The effectiveness of our channel-aware data-importance-based scheduling policy, compared with state-of-the-art methods proposed for synchronous FL, is validated through simulations. Moreover, we show that an ``age-aware’’ aggregation weighting design can significantly improve the learning performance in an asynchronous FL setting.
[827] Local Fragments, Global Gains: Subgraph Counting using Graph Neural Networks
Shubhajit Roy, Shrutimoy Das, Binita Maity, Anant Kumar, Anirban Dasgupta
Main category: cs.LG
TL;DR: Localized Weisfeiler-Leman algorithms improve expressivity and efficiency for subgraph counting, with variants for scalability and a fragmentation technique enabling exact counts of small subgraphs using only 1-WL.
Details
Motivation: Subgraph counting is fundamental for analyzing structural patterns in graph data, with applications in computational biology and social network analysis where motifs reveal functional properties. Existing methods need better expressivity and computational efficiency.Method: Proposed Local k-WL (more expressive than k-WL), plus scalable variants Layer k-WL and Recursive k-WL. Introduced fragmentation technique to decompose complex subgraphs into simpler subpatterns. Developed three-stage differentiable learning framework combining subpattern counts.
Result: Local k-WL is more expressive than k-WL and at most as expressive as (k+1)-WL. Fragmentation enables exact count of all induced subgraphs of size ≤4 using only 1-WL. Methods are more expressive than prior approaches under bounded time complexity.
Conclusion: Localized WL algorithms provide improved expressivity and efficiency for subgraph counting, bridging combinatorial algorithms with machine learning through a differentiable framework that can handle complex motifs by combining simpler subpatterns.
Abstract: Subgraph counting is a fundamental task for analyzing structural patterns in graph-structured data, with important applications in domains such as computational biology and social network analysis, where recurring motifs reveal functional and organizational properties. In this paper, we propose localized versions of the Weisfeiler-Leman (WL) algorithms to improve both expressivity and computational efficiency for this task. We introduce Local $k$-WL, which we prove to be more expressive than $k$-WL and at most as expressive as $(k+1)$-WL, and provide a characterization of patterns whose subgraph and induced subgraph counts are invariant under Local $k$-WL equivalence. To enhance scalability, we present two variants – Layer $k$-WL and Recursive $k$-WL – that achieve greater time and space efficiency compared to applying $k$-WL on the entire graph. Additionally, we propose a novel fragmentation technique that decomposes complex subgraphs into simpler subpatterns, enabling the exact count of all induced subgraphs of size at most $4$ using only $1$-WL, with extensions possible for larger patterns when $k>1$. Building on these ideas, we develop a three-stage differentiable learning framework that combines subpattern counts to compute counts of more complex motifs, bridging combinatorial algorithm design with machine learning approaches. We also compare the expressive power of Local $k$-WL with existing GNN hierarchies and demonstrate that, under bounded time complexity, our methods are more expressive than prior approaches.
[828] Achieving Linear Speedup with ProxSkip in Distributed Stochastic Optimization
Luyao Guo, Sulaiman A. Alghunaim, Kun Yuan, Laurent Condat, Jinde Cao
Main category: cs.LG
TL;DR: ProxSkip algorithm achieves linear speedup across non-convex, convex, and strongly convex settings with network-independent stepsizes in strongly convex case, and local updates reduce communication complexity.
Details
Motivation: Existing ProxSkip analyses are limited to strongly convex settings and fail to achieve linear speedup with respect to number of nodes. Key questions about non-convex behavior and linear speedup achievability remain open.Method: Revisiting and comprehensive analysis of ProxSkip algorithm for stochastic non-convex, convex, and strongly convex problems, examining effects of gradient noise, local updates, network connectivity, and data heterogeneity.
Result: ProxSkip achieves linear speedup across all three settings (non-convex, convex, strongly convex), can achieve linear speedup with network-independent stepsizes in strongly convex setting, and properly increasing local updates effectively reduces communication complexity.
Conclusion: The paper provides comprehensive convergence analysis for ProxSkip, addressing open questions about non-convex behavior and linear speedup, demonstrating its effectiveness in reducing communication while achieving optimal scaling with number of nodes.
Abstract: The ProxSkip algorithm for distributed optimization is gaining increasing attention due to its effectiveness in reducing communication. However, existing analyses of ProxSkip are limited to the strongly convex setting and fail to achieve linear speedup with respect to the number of nodes. Key questions regarding its behavior in the non-convex setting and the achievability of linear speedup remain open. In this paper, we revisit ProxSkip and address both questions. We provide a comprehensive analysis for stochastic non-convex, convex, and strongly convex problems, revealing the effects of gradient noise, local updates, network connectivity, and data heterogeneity on its convergence. We prove that ProxSkip achieves linear speedup across all three settings, and can further achieve linear speedup with network-independent stepsizes in the strongly convex setting. Moreover, we show that properly increasing local updates effectively reduces communication complexity.
[829] T-SHIRT: Token-Selective Hierarchical Data Selection for Instruction Tuning
Yanjun Fu, Faisal Hamman, Sanghamitra Dutta
Main category: cs.LG
TL;DR: T-SHIRT is a token-selective hierarchical data selection framework for instruction tuning that selects only 5% of data but outperforms full dataset training by up to 5.48 points across benchmarks.
Details
Motivation: Existing data selection methods for instruction tuning have two key limitations: (1) they assess quality at sample level, ignoring token-level informativeness, and (2) they overlook scoring robustness, often selecting samples based on superficial lexical features rather than true quality.Method: T-SHIRT introduces a new scoring method that includes only informative tokens in quality evaluation and promotes robust samples whose neighbors also show high quality with less local inconsistencies. It uses hierarchical data selection with token-level filtering.
Result: Models trained on only 5% of original data using T-SHIRT outperform those trained on full datasets by up to 5.48 points average across eight benchmarks. The method consistently surpasses state-of-the-art techniques across various LLMs and scales, processing 52k samples in 40 minutes on a single GPU.
Conclusion: T-SHIRT provides a cost-effective and highly efficient data selection framework for instruction tuning that addresses limitations of existing methods by considering token-level informativeness and scoring robustness, achieving superior performance with minimal data.
Abstract: Instruction tuning is essential for Large Language Models (LLMs) to effectively follow user instructions. To improve training efficiency and reduce data redundancy, recent works use LLM-based scoring functions, e.g., Instruction-Following Difficulty (IFD), to select high-quality instruction-tuning data with scores above a threshold. While these data selection methods often lead to models that can match or even exceed the performance of models trained on the full datasets, we identify two key limitations: (i) they assess quality at the sample level, ignoring token-level informativeness; and (ii) they overlook the robustness of the scoring method, often selecting a sample due to superficial lexical features instead of its true quality. In this work, we propose Token-Selective HIeRarchical Data Selection for Instruction Tuning (T-SHIRT), a novel data selection framework that introduces a new scoring method to include only informative tokens in quality evaluation and also promotes robust and reliable samples whose neighbors also show high quality with less local inconsistencies. We demonstrate that models instruction-tuned on a curated dataset (only 5% of the original size) using T-SHIRT can outperform those trained on the entire large-scale dataset by up to 5.48 points on average across eight benchmarks. Across various LLMs and training set scales, our method consistently surpasses existing state-of-the-art data selection techniques, while also remaining both cost-effective and highly efficient. For instance, by using GPT-2 for score computation, we are able to process a dataset of 52k samples in 40 minutes on a single GPU. Our code is available at https://github.com/Dynamite321/T-SHIRT.
[830] Deep sub-ensembles meets quantile regression: uncertainty-aware imputation for time series
Ying Liu, Peng Cui, Wenbo Hu, Richang Hong
Main category: cs.LG
TL;DR: Quantile Sub-Ensembles (QSE) is a novel non-generative time series imputation method that uses ensembles of quantile-regression networks to produce accurate imputations with reliable uncertainty estimates while maintaining computational efficiency.
Details
Motivation: Real-world time series often have substantial missing values, requiring imputation. While deep learning methods are effective, they produce overconfident imputations that risk system reliability. Diffusion methods handle uncertainty well but are computationally expensive and struggle with high missing rates.Method: Quantile Sub-Ensembles estimates uncertainty using ensembles of quantile-regression-based task networks, incorporated into a non-generative time series imputation framework. This approach avoids the computational burden of generative models while providing probabilistic imputation capabilities.
Result: Extensive experiments on five real-world datasets show superior performance in both deterministic and probabilistic imputation compared to baselines across most experimental settings. The method produces accurate and reliable imputations while remaining computationally efficient.
Conclusion: Quantile Sub-Ensembles offers an effective solution for time series imputation that addresses the limitations of both overconfident deep learning methods and computationally expensive diffusion models, providing reliable uncertainty estimation within an efficient non-generative framework.
Abstract: Real-world time series data often exhibits substantial missing values, posing challenges for advanced analysis. A common approach to addressing this issue is imputation, where the primary challenge lies in determining the appropriate values to fill in. While previous deep learning methods have proven effective for time series imputation, they often produce overconfident imputations, which poses a potentially overlooked risk to the reliability of the intelligent system. Diffusion methods are proficient in estimating probability distributions but face challenges under a high missing rate and are, moreover, computationally expensive due to the nature of the generative model framework. In this paper, we propose Quantile Sub-Ensembles, a novel method that estimates uncertainty with ensembles of quantile-regression-based task networks and incorporate Quantile Sub-Ensembles into a non-generative time series imputation method. Our method not only produces accurate and reliable imputations, but also remains computationally efficient due to its non-generative framework. We conduct extensive experiments on five real-world datasets, and the results demonstrates superior performance in both deterministic and probabilistic imputation compared to baselines across most experimental settings. The code is available at https://github.com/yingliu-coder/QSE.
[831] REASONING COMPILER: LLM-Guided Optimizations for Efficient Model Serving
Annabelle Sujun Tang, Christopher Priebe, Rohan Mahapatra, Lianhui Qin, Hadi Esmaeilzadeh
Main category: cs.LG
TL;DR: LLM-guided reasoning compiler using structured MCTS achieves substantial speedups with fewer samples than existing neural compilers.
Details
Motivation: High serving costs of large-scale models hinder accessibility and innovation. Existing compilers struggle with neural workloads due to large interdependent transformation spaces, and stochastic search methods are sample-inefficient and lack context awareness.Method: Introduces Reasoning Compiler framework that formulates optimization as sequential, context-aware decision process guided by LLM and structured Monte Carlo Tree Search (MCTS). LLM acts as proposal mechanism suggesting hardware-informed transformations based on current program state and performance feedback, while MCTS balances exploration/exploitation.
Result: Achieves substantial speedups with markedly fewer samples than leading neural compilers, demonstrating superior sample efficiency.
Conclusion: LLM-guided reasoning has potential to transform compiler optimization landscape by leveraging context-aware decision spaces with improved sample efficiency.
Abstract: While model serving has unlocked unprecedented capabilities, the high cost of serving large-scale models continues to be a significant barrier to widespread accessibility and rapid innovation. Compiler optimizations have long driven substantial performance improvements, but existing compilers struggle with neural workloads due to the exponentially large and highly interdependent space of possible transformations. Although existing stochastic search techniques can be effective, they are often sample-inefficient and fail to leverage the structural context underlying compilation decisions. We set out to investigate the research question of whether reasoning with large language models (LLMs), without any retraining, can leverage the context-aware decision space of compiler optimizations to significantly improve sample efficiency. To that end, we introduce a novel compilation framework (dubbed Reasoning Compiler) that formulates optimization as a sequential, context-aware decision process guided by a large language model and structured Monte Carlo tree search (MCTS). The LLM acts as a proposal mechanism, suggesting hardware-informed transformations that reflect the current program state and accumulated performance feedback. MCTS incorporates the LLM-generated proposals to balance exploration and exploitation, facilitating structured, context-sensitive traversal of the expansive compiler optimization space. By achieving substantial speedups with markedly fewer samples than leading neural compilers, our approach demonstrates the potential of LLM-guided reasoning to transform the landscape of compiler optimization.
[832] Accelerating data-driven algorithm selection for combinatorial partitioning problems
Vaggos Chatziafratis, Ishani Karmarkar, Yingxi Li, Ellen Vitercik
Main category: cs.LG
TL;DR: The paper introduces a framework for selecting clustering algorithms using subsampling with theoretical guarantees that the best algorithm on a small subsample will also be best on the full dataset.
Details
Motivation: When dealing with massive datasets, it's expensive to evaluate clustering algorithms directly on the full dataset, especially when ground truth is only accessible through expensive oracle queries. There's a need for efficient methods to select the best clustering algorithm without running it on the entire dataset.Method: Introduces a notion of size generalization for clustering accuracy, identifies conditions under which subsampling works, and provides theoretical guarantees for three classic algorithms: single-linkage, k-means++, and a smoothed variant of Gonzalez’s k-centers heuristic.
Result: Theoretical size generalization guarantees show that under certain conditions, the algorithm with best accuracy on a small subsample will also have the best accuracy on the full dataset. Empirical validation shows that as little as 5% subsampling can identify the best algorithm on real-world datasets.
Conclusion: The proposed subsampling approach provides an efficient method for clustering algorithm selection on massive datasets with theoretical guarantees, enabling practical algorithm selection with minimal computational overhead.
Abstract: In clustering algorithm selection, we are given a massive dataset and must efficiently select which clustering algorithm to use. We study this problem in a semi-supervised setting, with an unknown ground-truth clustering that we can only access through expensive oracle queries. Ideally, the clustering algorithm’s output will be structurally close to the ground truth. We approach this problem by introducing a notion of size generalization for clustering algorithm accuracy. We identify conditions under which we can (1) subsample the massive clustering instance, (2) evaluate a set of candidate algorithms on the smaller instance, and (3) guarantee that the algorithm with the best accuracy on the small instance will have the best accuracy on the original big instance. We provide theoretical size generalization guarantees for three classic clustering algorithms: single-linkage, k-means++, and (a smoothed variant of) Gonzalez’s k-centers heuristic. We validate our theoretical analysis with empirical results, observing that on real-world clustering instances, we can use a subsample of as little as 5% of the data to identify which algorithm is best on the full dataset.
[833] Graph Persistence goes Spectral
Mattie Ji, Amauri H. Souza, Vikas Garg
Main category: cs.LG
TL;DR: SpectRe integrates spectral graph information into persistent homology diagrams, creating a more expressive topological descriptor for graphs that surpasses existing methods in expressivity while maintaining local stability.
Details
Motivation: Existing topological descriptors that combine persistent homology with vertex/edge features still fail to capture basic graph structural information, limiting their expressivity for graph representation learning.Method: Proposes SpectRe, a new topological descriptor that integrates spectral information (likely from graph Laplacians or adjacency matrices) into persistent homology diagrams, creating a hybrid representation that combines topological and spectral graph properties.
Result: SpectRe is strictly more expressive than existing descriptors on graphs, is locally stable, and experimental results on synthetic and real-world datasets demonstrate its effectiveness in enhancing graph models for learning tasks.
Conclusion: SpectRe successfully combines spectral and topological information to create a superior graph descriptor that improves upon existing methods in both expressivity and stability, offering potential for enhanced graph representation learning.
Abstract: Including intricate topological information (e.g., cycles) provably enhances the expressivity of message-passing graph neural networks (GNNs) beyond the Weisfeiler-Leman (WL) hierarchy. Consequently, Persistent Homology (PH) methods are increasingly employed for graph representation learning. In this context, recent works have proposed decorating classical PH diagrams with vertex and edge features for improved expressivity. However, these methods still fail to capture basic graph structural information. In this paper, we propose SpectRe – a new topological descriptor for graphs that integrates spectral information into PH diagrams. Notably, SpectRe is strictly more expressive than existing descriptors on graphs. We also introduce notions of global and local stability to analyze existing descriptors and establish that SpectRe is locally stable. Finally, experiments on synthetic and real-world datasets demonstrate the effectiveness of SpectRe and its potential to enhance the capabilities of graph models in relevant learning tasks. Code is available at https://github.com/Aalto-QuML/SpectRe/.
[834] LCEN: A Nonlinear, Interpretable Feature Selection and Machine Learning Algorithm
Pedro Seber, Richard D. Braatz
Main category: cs.LG
TL;DR: LCEN algorithm for nonlinear interpretable feature selection and modeling that outperforms other methods in accuracy and speed while being robust to common data issues.
Details
Motivation: Interpretable models are essential for critical applications like aviation and medicine, but many current methods are either black-box or limited in handling nonlinear relationships and data issues.Method: LASSO-Clip-EN (LCEN) algorithm for nonlinear, interpretable feature selection and machine learning modeling that combines LASSO and elastic net approaches with clipping mechanisms.
Result: LCEN constructs sparser and more accurate models than other methods, is robust to noise/multicollinearity/data scarcity, 10.3x faster than thresholded elastic net for feature selection, and can rediscover physical laws from data.
Conclusion: LCEN provides an effective interpretable alternative to black-box models, offering superior performance in both feature selection and machine learning tasks while maintaining transparency essential for critical applications.
Abstract: Interpretable models can have advantages over black-box models, and interpretability is essential for the application of machine learning in critical settings, such as aviation or medicine. This article introduces the LASSO-Clip-EN (LCEN) algorithm for nonlinear, interpretable feature selection and machine learning modeling. In a wide variety of artificial and empirical datasets, LCEN constructed sparse and frequently more accurate models than other methods, including sparse, nonlinear methods, on tested datasets. LCEN was empirically observed to be robust against many issues typically present in datasets and modeling, including noise, multicollinearity, and data scarcity. As a feature selection algorithm, LCEN matched or surpassed the thresholded elastic net but was, on average, 10.3-fold faster based on our experiments. LCEN for feature selection can also rediscover multiple physical laws from empirical data. As a machine learning algorithm, when tested on processes with no known physical laws, LCEN achieved better results than many other dense and sparse methods – including being comparable to or better than ANNs on multiple datasets.
[835] STC-ViT: Spatio Temporal Continuous Vision Transformer for Medium-range Global Weather Forecasting
Hira Saleem, Flora Salim, Cormac Purcell
Main category: cs.LG
TL;DR: STC-ViT is a space-time continuous vision transformer for weather forecasting that combines Fourier Neural Operator for spatial modeling with transformer-parameterized Neural ODE for continuous-time dynamics, achieving competitive performance with shallow architecture and fast inference.
Details
Motivation: Traditional NWP systems are computationally expensive, and existing transformer models for weather forecasting have limitations: they discretize spatio-temporal dimensions (limiting continuous dynamics modeling) and require deep architectures (increasing computational cost and parameter redundancy).Method: STC-ViT integrates a Fourier Neural Operator (FNO) for global spatial operators with a transformer-parameterized Neural ODE for continuous-time dynamics, creating a space-time continuous model. This allows modeling continuous weather processes with a shallow, single-layer transformer encoder.
Result: The method achieves competitive forecasting performance on 1.5-degree WeatherBench 2, with inference speed of 0.125 seconds per forecast trajectory. It outperforms state-of-the-art data-driven and NWP models trained on higher-resolution data while using substantially lower data and compute costs.
Conclusion: STC-ViT provides an efficient, continuous spatio-temporal modeling approach for weather forecasting that reduces computational burden while maintaining strong performance, with detailed empirical analysis showing benefits of denser time grids, higher-accuracy ODE solvers, and deeper transformer stacks.
Abstract: Operational Numerical Weather Prediction (NWP) system relies on computationally expensive physics-based models. Recently, transformer models have shown remarkable potential in weather forecasting achieving state-of-the-art results. However, traditional transformers discretize spatio-temporal dimensions, limiting their ability to model continuous dynamical weather processes. Moreover, their reliance on increased depth to capture complex dependencies results in higher computational cost and parameter redundancy. We address these issues with STC-ViT, a Spatio-Temporal Continuous Vision Transformer for weather forecasting. STC-ViT integrates a Fourier Neural Operator (FNO) for global spatial operators with a transformer parameterised Neural ODE for continuous-time dynamics, yielding a space-time continuous model for weather forecasting. Our proposed method achieves competitive forecasting performance even with a shallow, single-layer transformer encoder mitigating the reliance on deeper networks. STC-ViT generates complete forecast trajectories with an inference speed of only 0.125 seconds and achieves strong medium-range forecasting skill on 1.5-degree WeatherBench 2 as compared to state-of-the-art data-driven and NWP models trained on higher-resolution data, with substantially lower data and compute costs. We also provide detailed empirical analysis on model’s performance with respect to denser time grids, higher-accuracy ODE solvers, and deeper transformer stacks.
[836] Spectral Convolutional Conditional Neural Processes
Peiman Mohseni, Nick Duffield
Main category: cs.LG
TL;DR: Spectral ConvCNP (SConvCNP) improves neural processes by performing global convolution in frequency domain to capture long-range dependencies efficiently, overcoming limitations of local CNN kernels.
Details
Motivation: Existing neural process variants using finite-dimensional embeddings or local CNN kernels struggle to capture infinite-dimensional generative processes and long-range dependencies efficiently. Convolutional CNPs address dimensionality but still face computational costs with large kernels needed for long-range dependencies.Method: Proposes Spectral ConvCNP (SConvCNP) that performs global convolution in frequency domain, inspired by Fourier neural operators. Directly parameterizes convolution kernels in Fourier domain to leverage compact global representations of natural signals.
Result: Validated effectiveness on both synthetic and real-world datasets, demonstrating how operator learning ideas can advance neural process capabilities.
Conclusion: SConvCNP successfully addresses limitations of previous neural process variants by enabling efficient global convolution through frequency domain processing, bridging operator learning with neural processes for improved long-range dependency modeling.
Abstract: Neural processes (NPs) are probabilistic meta-learning models that map sets of observations to posterior predictive distributions, enabling inference at arbitrary domain points. Their capacity to handle variable-sized collections of unstructured observations, combined with simple maximum-likelihood training and uncertainty-aware predictions, makes them well-suited for modeling data over continuous domains. Since their introduction, several variants have been proposed. Early approaches typically represented observed data using finite-dimensional summary embeddings obtained through aggregation schemes such as mean pooling. However, this strategy fundamentally mismatches the infinite-dimensional nature of the generative processes that NPs aim to capture. Convolutional conditional neural processes (ConvCNPs) address this limitation by constructing infinite-dimensional functional embeddings processed through convolutional neural networks (CNNs) to enforce translation equivariance. Yet CNNs with local spatial kernels struggle to capture long-range dependencies without resorting to large kernels, which impose significant computational costs. To overcome this limitation, we propose the Spectral ConvCNP (SConvCNP), which performs global convolution in the frequency domain. Inspired by Fourier neural operators (FNOs) for learning solution operators of partial differential equations (PDEs), our approach directly parameterizes convolution kernels in the frequency domain, leveraging the relatively compact yet global Fourier representation of many natural signals. We validate the effectiveness of SConvCNP on both synthetic and real-world datasets, demonstrating how ideas from operator learning can advance the capabilities of NPs.
[837] CoxSE: Exploring the Potential of Self-Explaining Neural Networks with Cox Proportional Hazards Model for Survival Analysis
Abdallah Alabdallah, Omar Hamed, Mattias Ohlsson, Thorsteinn Rögnvaldsson, Sepideh Pashami
Main category: cs.LG
TL;DR: The paper proposes two new explainable survival models: CoxSE (locally explainable Cox model using self-explaining neural networks) and CoxSENAM (hybrid of SENN and neural additive models) that maintain predictive power while providing stable, consistent explanations.
Details
Motivation: Traditional Cox models are explainable but limited by linearity, while deep neural extensions sacrifice explainability for predictive power. There's a need for survival models that combine both explainability and strong predictive performance.Method: Two approaches: 1) CoxSE - uses self-explaining neural networks (SENN) to estimate locally-linear log-hazard functions, 2) CoxSENAM - hybrid model combining SENN with neural additive models (NAM) to control explanation stability and consistency.
Result: SENN-based models provide more stable and consistent explanations than NAM-based models while maintaining black-box model predictive power. NAM-based models show better robustness to non-informative features, with the hybrid CoxSENAM exhibiting the best overall robustness.
Conclusion: The proposed CoxSE and CoxSENAM models successfully bridge the gap between explainability and predictive power in survival analysis, offering stable explanations while maintaining competitive performance and robustness.
Abstract: The Cox Proportional Hazards (CPH) model has long been the preferred survival model for its explainability. However, to increase its predictive power beyond its linear log-risk, it was extended to utilize deep neural networks, sacrificing its explainability. In this work, we explore the potential of self-explaining neural networks (SENN) for survival analysis. We propose a new locally explainable Cox proportional hazards model, named CoxSE, by estimating a locally-linear log-hazard function using the SENN. We also propose a modification to the Neural additive (NAM) model, hybrid with SENN, named CoxSENAM, which enables the control of the stability and consistency of the generated explanations. Several experiments using synthetic and real datasets are presented, benchmarking CoxSE and CoxSENAM against a NAM-based model, a DeepSurv model explained with SHAP, and a linear CPH model. The results show that, unlike the NAM-based model, the SENN-based model can provide more stable and consistent explanations while maintaining the predictive power of the black-box model. The results also show that, due to their structural design, NAM-based models demonstrate better robustness to non-informative features. Among the models, the hybrid model exhibits the best robustness.
[838] Causal Feature Selection Method for Contextual Multi-Armed Bandits in Recommender System
Zhenyu Zhao, Yexi Jiang
Main category: cs.LG
TL;DR: Two model-free filter methods (HIE and HDD) for feature selection in contextual multi-armed bandits that identify features driving heterogeneous treatment effects, improving performance without complex retraining.
Details
Motivation: Traditional feature selection in CMABs focuses on outcome correlation but neglects heterogeneous treatment effects across arms, which is crucial for optimal decision-making. Suboptimal features degrade rewards, interpretability, and efficiency in large-scale online systems.Method: Proposes two model-free filter methods: 1) Heterogeneous Incremental Effect (HIE) quantifies a feature’s value based on its ability to induce changes in the optimal arm, and 2) Heterogeneous Distribution Divergence (HDD) measures a feature’s impact on reward distribution divergence across arms. Both are computationally efficient, robust to model mis-specification, and adaptable to various feature types.
Result: Validated on synthetic data with known ground truth and in a large-scale commercial recommender system. Both methods consistently identify influential HTE features and enhance CMAB performance.
Conclusion: HIE and HDD provide effective, efficient feature selection methods for CMABs that specifically target heterogeneous treatment effects, making them suitable for dynamic environments where retraining complex models is infeasible.
Abstract: Effective feature selection is essential for optimizing contextual multi-armed bandits (CMABs) in large-scale online systems, where suboptimal features can degrade rewards, interpretability, and efficiency. Traditional feature selection often prioritizes outcome correlation, neglecting the crucial role of heterogeneous treatment effects (HTE) across arms in CMAB decision-making. This paper introduces two novel, model-free filter methods, Heterogeneous Incremental Effect (HIE) and Heterogeneous Distribution Divergence (HDD), specifically designed to identify features driving HTE. HIE quantifies a feature’s value based on its ability to induce changes in the optimal arm, while HDD measures its impact on reward distribution divergence across arms. These methods are computationally efficient, robust to model mis-specification, and adaptable to various feature types, making them suitable for rapid screening in dynamic environments where retraining complex models is infeasible. We validate HIE and HDD on synthetic data with known ground truth and in a large-scale commercial recommender system, demonstrating their consistent ability to identify influential HTE features and thereby enhance CMAB performance.
[839] Dual Prototypes for Adaptive Pre-Trained Model in Class-Incremental Learning
Zhiming Xu, Suorong Yang, Baile Xu, Furao Shen, Jian Zhao
Main category: cs.LG
TL;DR: DPTA is a dual-prototype network with task-wise adaptation for class-incremental learning using pre-trained models, achieving 1-5% improvements over recent methods.
Details
Motivation: Pre-trained model based class-incremental learning suffers from catastrophic forgetting when fine-tuning on incremental task streams, requiring better adaptation strategies.Method: Uses task-specific adapter modules with center-adapt loss for better representation clustering. Employs dual prototypes (raw and augmented) for test-time adapter selection and final classification.
Result: Consistently surpasses recent methods by 1-5% on multiple benchmarks, with ~3% improvement on VTAB dataset over state-of-the-art methods.
Conclusion: DPTA effectively addresses catastrophic forgetting in PTM-based CIL through task-wise adaptation and dual-prototype mechanisms, demonstrating significant performance gains.
Abstract: Class-incremental learning (CIL) aims to learn new classes while retaining previous knowledge. Although pre-trained model (PTM) based approaches show strong performance, directly fine-tuning PTMs on incremental task streams often causes renewed catastrophic forgetting. This paper proposes a Dual-Prototype Network with Task-wise Adaptation (DPTA) for PTM-based CIL. For each incremental learning task, an adapter module is built to fine-tune the PTM, where the center-adapt loss forces the representation to be more centrally clustered and class separable. The dual prototype network improves the prediction process by enabling test-time adapter selection, where the raw prototypes deduce several possible task indexes of test samples to select suitable adapter modules for PTM, and the augmented prototypes that could separate confusable classes are utilized to determine the final result. Experiments on multiple benchmarks show that DPTA consistently surpasses recent methods by 1% - 5%. Notably, on the VTAB dataset, it achieves approximately 3% improvement over state-of-the-art methods. The code is open-sourced in https://github.com/Yorkxzm/DPTA}
[840] Adaptive Nonlinear Vector Autoregression: Robust Forecasting for Noisy Chaotic Time Series
Azimov Sherkhon, Susana Lopez-Moreno, Eric Dolores-Cuenca, Sieun Lee, Sangil Kim
Main category: cs.LG
TL;DR: Proposes data-adaptive NVAR model with trainable MLP features that outperforms standard NVAR and reservoir computing methods for chaotic system forecasting, especially under noisy conditions.
Details
Motivation: Standard NVAR and reservoir computing methods use fixed nonlinear transformations (polynomial expansions or random feature maps) which limit adaptability to noisy or complex real-world data, and suffer from poor scalability due to costly matrix inversion during optimization.Method: Combines delay-embedded linear inputs with features generated by a shallow, trainable multilayer perceptron (MLP). The MLP and linear readout are jointly trained using gradient-based optimization, enabling learning of data-driven nonlinearities while preserving simple readout structure.
Result: The adaptive model outperformed standard NVAR, leaky echo state network (ESN), and hybrid ESN in predictive accuracy across multiple chaotic systems under both noise-free and synthetically noisy conditions, showing robust forecasting under noisy conditions.
Conclusion: The proposed data-adaptive NVAR model with trainable MLP features provides improved forecasting performance for chaotic systems, especially under noisy conditions, while addressing scalability issues of traditional methods through joint gradient-based optimization.
Abstract: Nonlinear vector autoregression (NVAR) and reservoir computing (RC) have shown promise in forecasting chaotic dynamical systems, such as the Lorenz-63 model and El Nino-Southern Oscillation. However, their reliance on fixed nonlinear transformations - polynomial expansions in NVAR or random feature maps in RC - limits their adaptability to high noise or complex real-world data. Furthermore, these methods also exhibit poor scalability in high-dimensional settings due to costly matrix inversion during optimization. We propose a data-adaptive NVAR model that combines delay-embedded linear inputs with features generated by a shallow, trainable multilayer perceptron (MLP). Unlike standard NVAR and RC models, the MLP and linear readout are jointly trained using gradient-based optimization, enabling the model to learn data-driven nonlinearities, while preserving a simple readout structure and improving scalability. Initial experiments across multiple chaotic systems, tested under noise-free and synthetically noisy conditions, showed that the adaptive model outperformed in predictive accuracy the standard NVAR, a leaky echo state network (ESN) - the most common RC model - and a hybrid ESN, thereby showing robust forecasting under noisy conditions.
[841] Measuring Fingerprints of Web-filtered Text Datasets and Fingerprint Propagation Through Training
Youssef Mansour, Reinhard Heckel
Main category: cs.LG
TL;DR: LLM pretraining datasets have identifiable fingerprints that allow neural networks to classify which dataset a text came from, and these fingerprints propagate to model outputs, enabling insights into undisclosed training data characteristics.
Details
Motivation: To investigate whether popular LLM pretraining datasets contain identifiable fingerprints or biases, similar to what has been observed in computer vision datasets, and to understand how these fingerprints affect model training and generalization.Method: Conducted dataset classification experiments on major open-source pretraining datasets (C4, RefinedWeb, DolmaCC, RedPajama-V2, FineWeb, DCLM-Baseline) using neural networks to classify which dataset a single text sequence belongs to, comparing performance against human classification. Analyzed fingerprints in formatting, vocabulary, and content distributions, and tested whether these fingerprints propagate through training by classifying sequences generated by models trained on these datasets.
Result: Neural networks can classify dataset origin significantly better than humans, indicating small differences in filtering/processing pipelines create identifiable fingerprints. These fingerprints negatively impact cross-dataset generalization and propagate through training - sequences generated by models can be accurately classified using classifiers trained on original datasets.
Conclusion: LLM pretraining datasets contain identifiable fingerprints that affect model behavior and can be used to infer undisclosed training data characteristics like pretraining mixture proportions and finetuning data sources, highlighting data provenance issues in LLM development.
Abstract: We investigate fingerprints in pretraining datasets for large language models (LLMs) through dataset classification experiments. Building on prior work demonstrating the existence of fingerprints or biases in popular computer vision datasets, we analyze popular open-source pretraining datasets for LLMs derived from CommonCrawl including C4, RefinedWeb, DolmaCC, RedPajama-V2, FineWeb, and DCLM-Baseline. Despite those datasets being obtained with similar curation steps, neural networks can classify surprisingly well which dataset a single text sequence belongs to, significantly better than a human can. This indicates that small differences in filtering and processing pipelines induce fingerprints. Those fingerprints are evident in formatting, vocabulary, and content distributions, and can negatively impact cross-dataset generalization. Additionally, we show that these fingerprints propagate through training: sequences generated by models trained on those datasets can be accurately classified by a classifier trained on the original datasets. This can offer insights into data characteristics that are typically undisclosed by LLM developers, including pretraining mixture proportions and finetuning data sources.
[842] Skewed Neuronal Heterogeneity Enhances Efficiency On Various Computing Systems
Arash Golmohammadi, Jannik Luboeinski, Christian Tetzlaff
Main category: cs.LG
TL;DR: Intrinsic neuronal heterogeneity (specifically skewed time constant variations) enhances performance and robustness across diverse temporal tasks, reduces network size requirements, and lowers computational costs for both machine learning and neuromorphic applications.
Details
Motivation: Biological systems exhibit stable, task-unspecific heterogeneity that enables efficient multitasking with minimal metabolic cost, unlike artificial systems that use top-down optimization which is biologically implausible, prone to forgetting, and resource-intensive.Method: Investigated the utility of intrinsic neuronal time constant heterogeneity for solving hundreds of distinct temporal tasks, focusing on skewed heterogeneity profiles similar to those found in biology, and tested across different computational paradigms.
Result: Intrinsic heterogeneity significantly enhances performance and robustness in an implementation-independent manner, with only skewed heterogeneity profiles producing performance gains. This advantage allows comparable performance with substantially smaller networks and lower operational, metabolic, and energetic costs.
Conclusion: Intrinsic (rather than task-induced) heterogeneity provides computational advantages that can inform the design of efficient artificial systems, particularly neuromorphic devices that naturally exhibit device-to-device variability, enabling more biologically plausible and resource-efficient computing.
Abstract: Heterogeneity is a ubiquitous property of many biological systems and has profound implications for computation. While it is conceivable to optimize neuronal and synaptic heterogeneity for a specific task, such top-down optimization is biologically implausible, prone to catastrophic forgetting, and both data- and energy-intensive. In contrast, biological organisms, with remarkable capacity to perform numerous tasks with minimal metabolic cost, exhibit a heterogeneity that is inherent, stable during adulthood, and task-unspecific. Inspired by this intrinsic form of heterogeneity, we investigate the utility of variations in neuronal time constants for solving hundreds of distinct temporal tasks of varying complexity. Our results show that intrinsic heterogeneity significantly enhances performance and robustness in an implementation-independent manner, indicating its usefulness for both (rate-based) machine learning and (spike-coded) neuromorphic applications. Importantly, only skewed heterogeneity profiles-reminiscent of those found in biology-produce such performance gains. We further demonstrate that this computational advantage eliminates the need for large networks, allowing comparable performance with substantially lower operational, metabolic, and energetic costs, respectively in silico, in vivo, and on neuromorphic hardware. Finally, we discuss the implications of intrinsic (rather than task-induced) heterogeneity for the design of efficient artificial systems, particularly novel neuromorphic devices that exhibit similar device-to-device variability.
[843] A Unifying View of Linear Function Approximation in Off-Policy RL Through Matrix Splitting and Preconditioning
Zechen Wu, Amy Greenwald, Ronald Parr
Main category: cs.LG
TL;DR: The paper provides a unified mathematical framework showing that TD, PFQI, and FQI are all solving the same linear system but with different matrix splitting schemes and preconditioners, explaining why TD convergence doesn’t guarantee FQI convergence.
Details
Motivation: Traditional views treat TD, PFQI, and FQI as fundamentally different algorithms based on update counts, but this perspective fails to explain why TD convergence doesn't imply FQI convergence. The authors aim to provide a unified mathematical understanding of these methods.Method: The authors develop a new mathematical framework using linear value function approximation, showing that all three methods solve the same linear system but with different matrix splitting schemes and preconditioners. They use matrix splitting theory and introduce an encoder-decoder perspective for convergence analysis.
Result: The framework establishes tight convergence connections among TD, PFQI, and FQI, explains why TD convergence doesn’t guarantee FQI convergence, provides sharper theoretical results, characterizes convergence conditions without feature independence assumptions, and shows that smaller learning rates can help when large ones fail.
Conclusion: The paper provides a unified theoretical foundation for understanding off-policy evaluation algorithms, introduces matrix splitting to convergence analysis, reveals that target networks represent a transition from constant to data-feature adaptive preconditioners, and enables dropping restrictive feature assumptions while maintaining convergence guarantees.
Abstract: In off-policy policy evaluation (OPE) tasks within reinforcement learning, Temporal Difference Learning(TD) and Fitted Q-Iteration (FQI) have traditionally been viewed as differing in the number of updates toward the target value function: TD makes one update, FQI makes an infinite number, and Partial Fitted Q-Iteration (PFQI) performs a finite number. We show that this view is not accurate, and provide a new mathematical perspective under linear value function approximation that unifies these methods as a single iterative method solving the same linear system, but using different matrix splitting schemes and preconditioners. We show that increasing the number of updates under the same target value function, i.e., the target network technique, is a transition from using a constant preconditioner to using a data-feature adaptive preconditioner. This elucidates, for the first time, why TD convergence does not necessarily imply FQI convergence, and establishes tight convergence connections among TD, PFQI, and FQI. Our framework enables sharper theoretical results than previous work and characterization of the convergence conditions for each algorithm, without relying on assumptions about the features (e.g., linear independence). We also provide an encoder-decoder perspective to better understand the convergence conditions of TD, and prove, for the first time, that when a large learning rate doesn’t work, trying a smaller one may help. Our framework also leads to the discovery of new crucial conditions on features for convergence, and shows how common assumptions about features influence convergence, e.g., the assumption of linearly independent features can be dropped without compromising the convergence guarantees of stochastic TD in the on-policy setting. This paper is also the first to introduce matrix splitting into the convergence analysis of these algorithms.
[844] Machine learning applications in archaeological practices: a review
Mathias Bellat, Jordy D. Orellana Figueroa, Jonathan S. Reeves, Ruhollah Taghizadeh-Mehrjardi, Claudio Tennie, Thomas Scholten
Main category: cs.LG
TL;DR: Review of 135 AI/ML archaeology papers (1997-2022) shows rapid growth since 2019, with structure detection and artifact classification as top applications, but reveals methodological issues and need for better workflows.
Details
Motivation: To comprehensively examine the prevalence, success, and methodological quality of AI/ML applications across all archaeological subfields, regions, and time periods, addressing gaps in previous reviews that focused only on specific subfields.Method: Exhaustive review of 135 articles published between 1997-2022, analyzing publication trends, application areas, model types, and methodological quality. Proposed a workflow guide for archaeologists to develop coherent methodologies.
Result: Significant increase in publications from 2019 onwards; structure detection and artifact classification were most common tasks; clustering/unsupervised methods underrepresented; neural networks and ensemble learning dominate; identified methodological issues including poorly defined requirements and unclear goals.
Conclusion: ML is becoming important for analyzing large multivariate archaeological data but requires well-defined structured methodologies, better reporting, and collaborative practices to maximize potential while addressing current limitations and misunderstandings.
Abstract: Artificial intelligence and machine learning applications in archaeology have increased significantly in recent years, and these now span all subfields, geographical regions, and time periods. The prevalence and success of these applications have remained largely unexamined, as recent reviews on the use of machine learning in archaeology have only focused only on specific subfields of archaeology. Our review examined an exhaustive corpus of 135 articles published between 1997 and 2022. We observed a significant increase in the number of publications from 2019 onwards. Automatic structure detection and artefact classification were the most represented tasks in the articles reviewed, followed by taphonomy, and archaeological predictive modelling. From the review, clustering and unsupervised methods were underrepresented compared to supervised models. Artificial neural networks and ensemble learning account for two thirds of the total number of models used. However, if machine learning models are gaining in popularity they remain subject to misunderstanding. We observed, in some cases, poorly defined requirements and caveats of the machine learning methods used. Furthermore, the goals and the needs of machine learning applications for archaeological purposes are in some cases unclear or poorly expressed. To address this, we proposed a workflow guide for archaeologists to develop coherent and consistent methodologies adapted to their research questions, project scale and data. As in many other areas, machine learning is rapidly becoming an important tool in archaeological research and practice, useful for the analyses of large and multivariate data, although not without limitations. This review highlights the importance of well-defined and well-reported structured methodologies and collaborative practices to maximise the potential of applications of machine learning methods in archaeology.
[845] Symmetric Behavior Regularized Policy Optimization
Lingwei Zhu, Haseeb Shah, Zheng Chen, Yukie Nagai, Martha White
Main category: cs.LG
TL;DR: Symmetric regularization in offline RL lacks analytic optimal policy; proposed Sf-AC uses Taylor expansion of Pearson-Vajda divergences to achieve stable solutions and strong D4RL performance.
Details
Motivation: To address the open question of symmetric regularization in offline RL, which unlike asymmetric regularization doesn't have an analytic optimal policy solution, limiting its practical utility.Method: Approximate optimal policy using Taylor series of Pearson-Vajda χ^n divergences, finding analytic expression only when capped at n=5. Propose Sf-AC algorithm that Taylor expands the conditional symmetry term for numerically stable computation.
Result: Sf-AC achieves consistently strong results across D4RL MuJoCo tasks and avoids per-environment failures observed in IQL, SQL, XQL and AWAC methods.
Conclusion: Symmetric regularization via Sf-AC enables more diverse and effective regularization choices for offline RL, overcoming the analytic solution limitation of symmetric BRPO.
Abstract: Behavior Regularized Policy Optimization (BRPO) leverages asymmetric (divergence) regularization to mitigate the distribution shift in offline Reinforcement Learning. This paper is the first to study the open question of symmetric regularization. We show that symmetric regularization does not permit an analytic optimal policy $π^$, posing a challenge to practical utility of symmetric BRPO. We approximate $π^$ by the Taylor series of Pearson-Vajda $χ^n$ divergences and show that an analytic policy expression exists only when the series is capped at $n=5$. To compute the solution in a numerically stable manner, we propose to Taylor expand the conditional symmetry term of the symmetric divergence loss, leading to a novel algorithm: Symmetric $f$-Actor Critic (S$f$-AC). S$f$-AC achieves consistently strong results across various D4RL MuJoCo tasks. Additionally, S$f$-AC avoids per-environment failures observed in IQL, SQL, XQL and AWAC, opening up possibilities for more diverse and effective regularization choices for offline RL.
[846] Testing Noise Assumptions of Learning Algorithms
Surbhi Goel, Adam R. Klivans, Konstantinos Stavropoulos, Arsen Vasilyan
Main category: cs.LG
TL;DR: First efficient algorithm to test noise model assumptions on training data, showing testable learning can be tractable for Massart noise but requires super-polynomial time for random classification noise.
Details
Motivation: Address a fundamental gap in computational learning theory: whether we can efficiently test if a training set satisfies the assumptions of a given noise model, which has remained unaddressed despite decades of noise-tolerant learning research.Method: Extend the testable learning framework of Rubinfeld and Vasilyan (2023) with two conditions: (1) test acceptance requires learner to output classifier with certificate of optimality, (2) test must pass for any dataset drawn according to specified modeling assumptions. Apply to learning halfspaces over Gaussian marginals with Massart noise.
Result: Present first efficient algorithm for testable learning of halfspaces with Massart noise over Gaussian marginals. Show separation: testable learning requires super-polynomial time for random classification noise (η=1/2) while classical learning is trivial.
Conclusion: Testable learning is tractable for some noise models (Massart) but fundamentally harder than classical learning for others (random classification noise), establishing a new computational separation in learning theory.
Abstract: We pose a fundamental question in computational learning theory: can we efficiently test whether a training set satisfies the assumptions of a given noise model? This question has remained unaddressed despite decades of research on learning in the presence of noise. In this work, we show that this task is tractable and present the first efficient algorithm to test various noise assumptions on the training data. To model this question, we extend the recently proposed testable learning framework of Rubinfeld and Vasilyan (2023) and require a learner to run an associated test that satisfies the following two conditions: (1) whenever the test accepts, the learner outputs a classifier along with a certificate of optimality, and (2) the test must pass for any dataset drawn according to a specified modeling assumption on both the marginal distribution and the noise model. We then consider the problem of learning halfspaces over Gaussian marginals with Massart noise (where each label can be flipped with probability less than $1/2$ depending on the input features), and give a fully-polynomial time testable learning algorithm. We also show a separation between the classical setting of learning in the presence of structured noise and testable learning. In fact, for the simple case of random classification noise (where each label is flipped with fixed probability $η= 1/2$), we show that testable learning requires super-polynomial time while classical learning is trivial.
[847] TimeDistill: Efficient Long-Term Time Series Forecasting with MLP via Cross-Architecture Distillation
Juntong Ni, Zewen Liu, Shiyu Wang, Ming Jin, Wei Jin
Main category: cs.LG
TL;DR: TimeDistill is a knowledge distillation framework that transfers multi-scale temporal patterns from heavy models (Transformers/CNNs) to lightweight MLPs for efficient time series forecasting.
Details
Motivation: Transformer and CNN models have strong forecasting performance but suffer from high computational/storage costs that limit large-scale deployment. There's a need for lightweight yet accurate alternatives.Method: Proposes TimeDistill, a cross-architecture knowledge distillation framework that transfers complementary multi-scale and multi-period patterns from teacher models (Transformers/CNNs) to student MLPs. Includes theoretical analysis showing KD acts as specialized mixup augmentation.
Result: Improves MLP performance by up to 18.6%, surpassing teacher models on eight datasets. Achieves 7X faster inference and 130X fewer parameters while maintaining accuracy.
Conclusion: TimeDistill enables efficient time series forecasting by distilling complex patterns from heavy models into lightweight MLPs, offering practical deployment advantages without sacrificing performance.
Abstract: Transformer-based and CNN-based methods demonstrate strong performance in long-term time series forecasting. However, their high computational and storage requirements can hinder large-scale deployment. To address this limitation, we propose integrating lightweight MLP with advanced architectures using knowledge distillation (KD). Our preliminary study reveals different models can capture complementary patterns, particularly multi-scale and multi-period patterns in the temporal and frequency domains. Based on this observation, we introduce TimeDistill, a cross-architecture KD framework that transfers these patterns from teacher models (e.g., Transformers, CNNs) to MLP. Additionally, we provide a theoretical analysis, demonstrating that our KD approach can be interpreted as a specialized form of mixup data augmentation. TimeDistill improves MLP performance by up to 18.6%, surpassing teacher models on eight datasets. It also achieves up to 7X faster inference and requires 130X fewer parameters. Furthermore, we conduct extensive evaluations to highlight the versatility and effectiveness of TimeDistill.
[848] STaRFormer: Semi-Supervised Task-Informed Representation Learning via Dynamic Attention-Based Regional Masking for Sequential Data
Maximilian Forstenhäusler, Daniel Külzer, Christos Anagnostopoulos, Shameem Puthiya Parambath, Natascha Weber
Main category: cs.LG
TL;DR: STaRFormer is a Transformer-based framework for sequential modeling that handles non-stationary, irregularly sampled spatiotemporal data using dynamic attention-based regional masking and semi-supervised contrastive learning.
Details
Motivation: The paper addresses the challenge of predicting smart device user intent near vehicles using sequential spatiotemporal data. Real-world scenarios present difficulties due to environmental factors and sensor limitations that create non-stationary and irregularly sampled data.Method: STaRFormer uses a Transformer-based approach with two key innovations: 1) dynamic attention-based regional masking scheme, and 2) novel semi-supervised contrastive learning paradigm to enhance task-specific latent representations.
Result: Comprehensive experiments on 56 diverse datasets (varying in types, tasks, domains, sequence lengths, training samples, and applications) demonstrate STaRFormer’s efficacy, achieving notable improvements over state-of-the-art approaches.
Conclusion: STaRFormer serves as a universal framework for sequential modeling that effectively handles real-world challenges like non-stationary and irregularly sampled data, making it suitable for intent prediction in smart device-vehicle interaction scenarios.
Abstract: Understanding user intent is essential for situational and context-aware decision-making. Motivated by a real-world scenario, this work addresses intent predictions of smart device users in the vicinity of vehicles by modeling sequential spatiotemporal data. However, in real-world scenarios, environmental factors and sensor limitations can result in non-stationary and irregularly sampled data, posing significant challenges. To address these issues, we propose STaRFormer, a Transformer-based approach that can serve as a universal framework for sequential modeling. STaRFormer utilizes a new dynamic attention-based regional masking scheme combined with a novel semi-supervised contrastive learning paradigm to enhance task-specific latent representations. Comprehensive experiments on 56 datasets varying in types (including non-stationary and irregularly sampled), tasks, domains, sequence lengths, training samples, and applications demonstrate the efficacy of STaRFormer, achieving notable improvements over state-of-the-art approaches.
[849] Class-Conditional Distribution Balancing for Group Robust Classification
Miaoyun Zhao, Chenrong Li, Qiang Zhang
Main category: cs.LG
TL;DR: A novel robust learning method that addresses spurious correlations by reframing them as class-conditional distribution imbalances, using sample reweighting to maximize conditional entropy without needing bias annotations or predictions.
Details
Motivation: Existing methods for handling spurious correlations require expensive bias annotations or rely on large-scale pretrained foundation models, making them impractical for resource-limited domains. There's a need for a simpler approach that works without bias supervision.Method: Reframes spurious correlations as class-conditional distribution imbalances and uses sample reweighting to maximize conditional entropy of labels given spurious factors. This automatically highlights minority groups and classes to achieve class-conditional distribution balancing.
Result: The approach consistently delivers state-of-the-art performance, rivaling methods that rely on bias supervision, as demonstrated through extensive experiments and analysis.
Conclusion: The proposed method effectively dismantles spurious correlations by achieving class-conditional distribution balancing without requiring bias annotations or predictions, making it practical for resource-limited domains while maintaining competitive performance.
Abstract: Spurious correlations that lead models to correct predictions for the wrong reasons pose a critical challenge for robust real-world generalization. Existing research attributes this issue to group imbalance and addresses it by maximizing group-balanced or worst-group accuracy, which heavily relies on expensive bias annotations. A compromise approach involves predicting bias information using extensively pretrained foundation models, which requires large-scale data and becomes impractical for resource-limited rare domains. To address these challenges, we offer a novel perspective by reframing the spurious correlations as imbalances or mismatches in class-conditional distributions, and propose a simple yet effective robust learning method that eliminates the need for both bias annotations and predictions. With the goal of maximizing the conditional entropy (uncertainty) of the label given spurious factors, our method leverages a sample reweighting strategy to achieve class-conditional distribution balancing, which automatically highlights minority groups and classes, effectively dismantling spurious correlations and producing a debiased data distribution for classification. Extensive experiments and analysis demonstrate that our approach consistently delivers state-of-the-art performance, rivaling methods that rely on bias supervision.
[850] PARD: Accelerating LLM Inference with Low-Cost PARallel Draft Model Adaptation
Zihao An, Huajun Bai, Ziqiong Liu, Dong Li, Emad Barsoum
Main category: cs.LG
TL;DR: PARD is a speculative decoding method that accelerates LLM inference with target-independent parallel token prediction, achieving up to 3.67× speedup without requiring separate draft heads for each target model.
Details
Motivation: Autoregressive LLMs are bottlenecked by memory bandwidth, generating only one token per forward pass. Existing speculative decoding methods like EAGLE require training separate draft heads for each target model, which is costly and limits practical adoption.Method: PARD uses a target-independent draft model that works across model families without per-model training. It predicts multiple future tokens in parallel within a single forward pass. A COnditional Drop-token (COD) mechanism adapts autoregressive draft models to parallel prediction by leveraging prefix key-value state integrity.
Result: COD improves draft model training efficiency by 3× compared to traditional masked prediction. On vLLM framework, PARD achieves up to 3.67× speedup on LLaMA3.1-8B, reaching 264.88 tokens per second, which is 1.15× faster than EAGLE-3.
Conclusion: PARD provides an efficient speculative decoding solution that reduces adaptation costs while accelerating inference through parallel token prediction, making it practical for deployment across model families.
Abstract: The autoregressive nature of large language models (LLMs) fundamentally limits inference speed, as each forward pass generates only a single token and is often bottlenecked by memory bandwidth. Speculative decoding has emerged as a promising solution, adopting a draft-then-verify strategy to accelerate token generation. While the EAGLE series achieves strong acceleration, its requirement of training a separate draft head for each target model introduces substantial adaptation costs. In this work, we propose \textbf{PARD (PARallel Draft)}, a novel speculative decoding method featuring \textit{target-independence} and \textit{parallel token prediction}. Specifically, PARD enables a single draft model to be applied across an entire family of target models without requiring separate training for each variant, thereby minimizing adaptation costs. Meanwhile, PARD substantially accelerates inference by predicting multiple future tokens within a single forward pass of the draft phase. To further reduce the training adaptation cost of PARD, we propose a COnditional Drop-token (COD) mechanism based on the integrity of prefix key-value states, enabling autoregressive draft models to be adapted into parallel draft models at low-cost. Our experiments show that the proposed COD method improves draft model training efficiency by \textbf{3$\times$} compared with traditional masked prediction training. On the \texttt{vLLM} inference framework, PARD achieves up to \textbf{3.67$\times$} speedup on LLaMA3.1-8B, reaching \textbf{264.88} tokens per second, which is \textbf{1.15$\times$} faster than EAGLE-3. Our code is available at https://github.com/AMD-AIG-AIMA/PARD.
[851] IberFire – a detailed creation of a spatio-temporal dataset for wildfire risk assessment in Spain
Julen Erzibengoa, Meritxell Gómez-Omella, Izaro Goienetxea
Main category: cs.LG
TL;DR: IberFire is a high-resolution (1km x 1km x 1-day) wildfire dataset for Spain with 120 features across 8 categories, covering 2007-2024, designed to support advanced ML/DL wildfire risk modeling and climate analysis.
Details
Motivation: Wildfires threaten ecosystems, economies, and public safety in Mediterranean regions like Spain. Existing datasets lack fine-grained spatio-temporal resolution needed to capture complex environmental and human factors driving wildfire dynamics.Method: Created IberFire dataset by integrating open-access data and tools to compile 120 features across 8 categories (auxiliary data, fire history, geography, topography, meteorology, vegetation indices, human activity, land cover) at 1km x 1km x 1-day resolution for mainland Spain and Balearic Islands from 2007-2024.
Result: IberFire provides enhanced spatial granularity and feature diversity compared to existing European datasets, with publicly available codebase on Zenodo ensuring transparency, reproducibility, and open access for research collaboration.
Conclusion: IberFire enables advanced wildfire risk modeling via ML/DL, supports climate trend analysis, and informs fire prevention and land management strategies through its high-resolution, comprehensive, and reproducible framework.
Abstract: Wildfires pose a threat to ecosystems, economies and public safety, particularly in Mediterranean regions such as Spain. Accurate predictive models require high-resolution spatio-temporal data to capture complex dynamics of environmental and human factors. To address the scarcity of fine-grained wildfire datasets in Spain, we introduce IberFire: a spatio-temporal dataset with 1 km x 1 km x 1-day resolution, covering mainland Spain and the Balearic Islands from December 2007 to December 2024. IberFire integrates 120 features across eight categories: auxiliary data, fire history, geography, topography, meteorology, vegetation indices, human activity and land cover. All features and processing rely on open-access data and tools, with a publicly available codebase ensuring transparency and applicability. IberFire offers enhanced spatial granularity and feature diversity compared to existing European datasets, and provides a reproducible framework. It supports advanced wildfire risk modelling via Machine Learning and Deep Learning, facilitates climate trend analysis, and informs fire prevention and land management strategies. The dataset is freely available on Zenodo to promote open research and collaboration.
[852] Robust Detection of Synthetic Tabular Data under Schema Variability
G. Charbel N. Kindji, Elisa Fromont, Lina Maria Rojas-Barahona, Tanguy Urvoy
Main category: cs.LG
TL;DR: A novel transformer-based method for detecting synthetic tabular data in real-world conditions with variable schemas, achieving significant performance improvements over previous approaches.
Details
Motivation: While detection methods exist for images and text, synthetic tabular data detection has been largely overlooked despite its ubiquity. The challenge is especially difficult due to heterogeneous structures and unseen formats at test time.Method: Introduces a novel datum-wise transformer architecture with a table-adaptation component to handle variable and previously unseen table schemas in real-world conditions.
Result: Significantly outperforms the only previously published baseline, improving both AUC and accuracy by 7 points. The table-adaptation component adds another 7 accuracy points, demonstrating enhanced robustness.
Conclusion: This work provides the first strong evidence that detecting synthetic tabular data in real-world conditions is feasible, and demonstrates substantial improvements over previous approaches. Source code release is pending administrative procedures.
Abstract: The rise of powerful generative models has sparked concerns over data authenticity. While detection methods have been extensively developed for images and text, the case of tabular data, despite its ubiquity, has been largely overlooked. Yet, detecting synthetic tabular data is especially challenging due to its heterogeneous structure and unseen formats at test time. We address the underexplored task of detecting synthetic tabular data ‘‘in the wild’’, i.e. when the detector is deployed on tables with variable and previously unseen schemas. We introduce a novel datum-wise transformer architecture that significantly outperforms the only previously published baseline, improving both AUC and accuracy by 7 points. By incorporating a table-adaptation component, our model gains an additional 7 accuracy points, demonstrating enhanced robustness. This work provides the first strong evidence that detecting synthetic tabular data in real-world conditions is feasible, and demonstrates substantial improvements over previous approaches. Following acceptance of the paper, we are finalizing the administrative and licensing procedures necessary for releasing the source code. This extended version will be updated as soon as the release is complete.
[853] Exploring Equity of Climate Policies using Multi-Agent Multi-Objective Reinforcement Learning
Palok Biswas, Zuzanna Osika, Isidoro Tamassia, Adit Whorra, Jazmin Zatarain-Salazar, Jan Kwakkel, Frans A. Oliehoek, Pradeep K. Murukannaiah
Main category: cs.LG
TL;DR: Justice framework integrates Integrated Assessment Models with Multi-Objective Multi-Agent Reinforcement Learning to generate equitable climate policies that balance economic growth, temperature goals, and climate justice.
Details
Motivation: Traditional IAMs optimize based on single objectives, failing to capture trade-offs between economic growth, temperature goals, and climate justice, leading to policies that perpetuate inequalities and fuel disagreements in climate negotiations.Method: Justice integrates Integrated Assessment Models (IAMs) with Multi-Objective Multi-Agent Reinforcement Learning (MOMARL), incorporating multiple objectives and multiple agents to represent diverse policy actors and their interactions.
Result: The framework identifies equitable Pareto-optimal policies that balance climate and economic goals while highlighting equity considerations, facilitating deliberative decision-making by presenting policymakers with inherent trade-offs.
Conclusion: Justice provides a more realistic and equitable approach to climate policy assessment by capturing multi-objective trade-offs and diverse stakeholder interactions, enabling better-informed climate policy negotiations.
Abstract: Addressing climate change requires coordinated policy efforts of nations worldwide. These efforts are informed by scientific reports, which rely in part on Integrated Assessment Models (IAMs), prominent tools used to assess the economic impacts of climate policies. However, traditional IAMs optimize policies based on a single objective, limiting their ability to capture the trade-offs among economic growth, temperature goals, and climate justice. As a result, policy recommendations have been criticized for perpetuating inequalities, fueling disagreements during policy negotiations. We introduce Justice, the first framework integrating IAM with Multi-Objective Multi-Agent Reinforcement Learning (MOMARL). By incorporating multiple objectives, Justice generates policy recommendations that shed light on equity while balancing climate and economic goals. Further, using multiple agents can provide a realistic representation of the interactions among the diverse policy actors. We identify equitable Pareto-optimal policies using our framework, which facilitates deliberative decision-making by presenting policymakers with the inherent trade-offs in climate and economic policy.
[854] Conformal Prediction for Time-series Forecasting with Change Points
Sophia Sun, Rose Yu
Main category: cs.LG
TL;DR: CPTC algorithm integrates state prediction with online conformal prediction to handle uncertainty quantification in time series with change points, proving validity and showing improved performance on synthetic and real datasets.
Details
Motivation: Current conformal prediction methods struggle with time series data containing change points (sudden shifts in data-generating process), creating a gap in uncertainty quantification for non-stationary time series.Method: Proposes CPTC algorithm that integrates a model to predict underlying state with online conformal prediction to model uncertainties in non-stationary time series with change points.
Result: Proves CPTC’s validity and improved adaptivity under minimum assumptions, and demonstrates practical effectiveness on 6 synthetic and real-world datasets with improved validity and adaptivity compared to state-of-the-art baselines.
Conclusion: CPTC successfully addresses the gap in conformal prediction for time series with change points, providing valid and adaptive uncertainty quantification for non-stationary time series.
Abstract: Conformal prediction has been explored as a general and efficient way to provide uncertainty quantification for time series. However, current methods struggle to handle time series data with change points - sudden shifts in the underlying data-generating process. In this paper, we propose a novel Conformal Prediction for Time-series with Change points (CPTC) algorithm, addressing this gap by integrating a model to predict the underlying state with online conformal prediction to model uncertainties in non-stationary time series. We prove CPTC’s validity and improved adaptivity in the time series setting under minimum assumptions, and demonstrate CPTC’s practical effectiveness on 6 synthetic and real-world datasets, showing improved validity and adaptivity compared to state-of-the-art baselines.
[855] How many measurements are enough? Bayesian recovery in inverse problems with general distributions
Ben Adcock, Nick Huang
Main category: cs.LG
TL;DR: The paper establishes non-asymptotic sample complexity bounds for Bayesian inverse problems with general priors, showing dependence on prior complexity and forward/noise concentration. For DNN-based generative priors, sample complexity scales log-linearly with latent dimension, and coherence plays a fundamental role in Bayesian recovery.
Details
Motivation: To provide rigorous guarantees for Bayesian inverse problems with arbitrary distributions, unifying and extending prior work on deterministic recovery to the Bayesian setting with general priors, forward operators, and noise distributions.Method: Analyze posterior sampling with approximate prior P, establish sufficient conditions using approximate covering number to quantify prior complexity, and derive concentration bounds for forward operator and noise distributions. Specialize to generative priors where P is pushforward of latent distribution via DNN.
Result: Non-asymptotic bound showing sample complexity depends on (i) approximate covering number of prior P, and (ii) concentration bounds for forward/noise distributions. For DNN priors, sample complexity scales log-linearly with latent dimension k. Coherence of measurement matrix U with respect to prior support determines sample complexity.
Conclusion: The framework unifies and extends prior work, providing rigorous guarantees for Bayesian inverse problems. Coherence plays fundamental role in Bayesian recovery, and DNN-based priors are effective with sample complexity scaling favorably with latent dimension.
Abstract: We study the sample complexity of Bayesian recovery for solving inverse problems with general prior, forward operator and noise distributions. We consider posterior sampling according to an approximate prior $\mathcal{P}$, and establish sufficient conditions for stable and accurate recovery with high probability. Our main result is a non-asymptotic bound that shows that the sample complexity depends on (i) the intrinsic complexity of $\mathcal{P}$, quantified by its so-called approximate covering number, and (ii) concentration bounds for the forward operator and noise distributions. As a key application, we specialize to generative priors, where $\mathcal{P}$ is the pushforward of a latent distribution via a Deep Neural Network (DNN). We show that the sample complexity scales log-linearly with the latent dimension $k$, thus establishing the efficacy of DNN-based priors. Generalizing existing results on deterministic (i.e., non-Bayesian) recovery for the important problem of random sampling with an orthogonal matrix $U$, we show how the sample complexity is determined by the coherence of $U$ with respect to the support of $\mathcal{P}$. Hence, we establish that coherence plays a fundamental role in Bayesian recovery as well. Overall, our framework unifies and extends prior work, providing rigorous guarantees for the sample complexity of solving Bayesian inverse problems with arbitrary distributions.
[856] ORACLE: Explaining Feature Interactions in Neural Networks with ANOVA
Dongseok Kim, Hyoungsun Choi, Mohamed Jismy Aashik Rasool, Gisung Oh
Main category: cs.LG
TL;DR: ORACLE is a framework that explains neural networks on tabular/scientific data by fitting ANOVA-style main and pairwise interaction effects to model predictions, providing interpretable factorial structure analysis.
Details
Motivation: Need for interpretable explanations of neural networks on tabular and scientific design data, particularly for workflows requiring stable, design-of-experiments (DoE)-style interaction summaries that are comparable across models.Method: Treats trained network as black-box response, learns orthogonal factorial surrogate on discretized input grid, uses centering and μ-rebalancing to obtain L²-consistent main- and interaction-effect tables, producing grid-based interaction maps.
Result: On synthetic factorial and tabular regression benchmarks, ORACLE more accurately recovers ground-truth ANOVA interactions and hotspot structure than Monte Carlo SHAP-family methods, with better ranking, localization, and cross-backbone stability.
Conclusion: ORACLE is particularly well-suited for scientific/engineering tabular workflows requiring stable DoE-style interaction summaries, as grid-based ANOVA surrogates are most effective when features admit interpretable factorial structure.
Abstract: We introduce ORACLE, a framework that explains neural networks on tabular and scientific design data. It fits ANOVA-style main and pairwise interaction effects to a model’s prediction surface. ORACLE treats a trained network as a black-box response, learns an orthogonal factorial surrogate on a discretized input grid, and uses simple centering and $μ$-rebalancing steps to obtain main- and interaction-effect tables that remain $L^2$-consistent with the original model. The resulting grid-based interaction maps are easy to visualize, comparable across backbones, and directly connected to classical design-of-experiments analyses. On synthetic factorial and low- to medium-dimensional tabular regression benchmarks, ORACLE more accurately recovers ground-truth ANOVA interactions and hotspot structure than Monte Carlo SHAP-family interaction methods, as measured by ranking, localization, and cross-backbone stability metrics. In latent image and text settings, ORACLE instead delineates its natural scope, and our results indicate that grid-based ANOVA surrogates are most effective when features admit interpretable factorial structure, making ORACLE particularly well-suited to scientific and engineering tabular workflows that require stable, DoE-style interaction summaries.
[857] Multiclass threshold-based classification
Francesco Marchetti, Edoardo Legnaro, Sabrina Guastavino
Main category: cs.LG
TL;DR: A threshold-based framework for multiclass classification that replaces softmax probabilistic interpretation with geometric simplex analysis, enabling a posteriori threshold tuning and score-oriented losses.
Details
Motivation: Standard multiclass classification uses argmax rule with softmax outputs, but lacks the threshold tuning capability available in binary classification. The paper aims to generalize threshold-based approaches from binary to multiclass settings to improve classification performance.Method: Replaces probabilistic interpretation of softmax outputs with geometric interpretation on multidimensional simplex. Introduces multidimensional threshold that enables a posteriori optimization of classification scores through threshold tuning. Defines score-oriented losses by treating threshold as random variable.
Result: Multidimensional threshold tuning yields consistent performance improvements across various networks and datasets. Proposed multiclass score-oriented losses are competitive with standard loss functions, showing advantages similar to those observed in binary classification.
Conclusion: The threshold-based framework successfully generalizes binary threshold tuning to multiclass classification, enabling performance improvements through a posteriori optimization and introducing competitive score-oriented loss functions.
Abstract: In this paper, we introduce a threshold-based framework for multiclass classification that generalizes the standard argmax rule. This is done by replacing the probabilistic interpretation of softmax outputs with a geometric one on the multidimensional simplex, where the classification depends on a multidimensional threshold. This change of perspective enables for any trained classification network an a posteriori optimization of the classification score by means of threshold tuning, as usually carried out in the binary setting. This allows a further refinement of the prediction capability of any network. Moreover, this multidimensional threshold-based setting makes it possible to define score-oriented losses, which are based on the interpretation of the threshold as a random variable. Our experiments show that the multidimensional threshold tuning yields consistent performance improvements across various networks and datasets, and that the proposed multiclass score-oriented losses are competitive with standard loss functions, resembling the advantages observed in the binary case.
[858] ToMA: Token Merge with Attention for Diffusion Models
Wenbo Lu, Shaoyi Zheng, Yuxuan Xia, Shengjie Wang
Main category: cs.LG
TL;DR: ToMA is a GPU-efficient token reduction method for diffusion models that reformulates token merging as submodular optimization and implements it as attention-like linear transformations, achieving 24% latency reduction for SDXL generation.
Details
Motivation: Existing token reduction methods (ToMeSD, ToFu) have theoretical speedups but introduce GPU-inefficient operations (sorting, scattered writes) that negate benefits when paired with optimized attention like FlashAttention, creating a gap between theoretical and practical efficiency.Method: 1) Reformulates token merge as submodular optimization to select diverse tokens; 2) Implements merge/unmerge as attention-like linear transformations using GPU-friendly matrix operations; 3) Exploits latent locality and sequential redundancy (pattern reuse) to minimize overhead.
Result: Reduces SDXL generation latency by 24% and Flux generation latency by 23% with minimal quality degradation (DINO Δ < 0.07), outperforming prior token reduction methods.
Conclusion: ToMA bridges the gap between theoretical and practical efficiency for transformers in diffusion models by providing GPU-aligned token reduction that delivers actual speed improvements while maintaining image quality.
Abstract: Diffusion models excel in high-fidelity image generation but face scalability limits due to transformers’ quadratic attention complexity. Plug-and-play token reduction methods like ToMeSD and ToFu reduce FLOPs by merging redundant tokens in generated images but rely on GPU-inefficient operations (e.g., sorting, scattered writes), introducing overheads that negate theoretical speedups when paired with optimized attention implementations (e.g., FlashAttention). To bridge this gap, we propose Token Merge with Attention (ToMA), an off-the-shelf method that redesigns token reduction for GPU-aligned efficiency, with three key contributions: 1) a reformulation of token merge as a submodular optimization problem to select diverse tokens; 2) merge/unmerge as an attention-like linear transformation via GPU-friendly matrix operations; and 3) exploiting latent locality and sequential redundancy (pattern reuse) to minimize overhead. ToMA reduces SDXL/Flux generation latency by 24%/23%, respectively (with DINO $Δ< 0.07$), outperforming prior methods. This work bridges the gap between theoretical and practical efficiency for transformers in diffusion. Code available at https://github.com/WenboLuu/ToMA.
[859] Fourier-Invertible Neural Encoder (FINE) for Homogeneous Flows
Anqiao Ouyang, Hongyi Ke, Qi Wang
Main category: cs.LG
TL;DR: FINE is a compact, interpretable neural encoder for translation-equivariant data that uses reversible filters, monotonic activations, and Fourier truncation to achieve information-preserving compression while respecting translational symmetry.
Details
Motivation: To create a principled framework for interpretable, low-parameter, symmetry-preserving dimensional reduction that bridges Fourier representations with modern neural architectures for scientific and physics-informed learning.Method: Integrates reversible filters and monotonic activation functions with a Fourier truncation bottleneck, creating an information-preserving compression architecture that respects translational symmetry through spectral truncation and group-equivariant representations.
Result: Achieves 4.9-9.1 times lower reconstruction error than convolutional autoencoders while using only 13-21% of their parameters across three test cases: 1D nonlinear wave interaction, 1D Kuramoto-Sivashinsky turbulence, and 2D turbulence datasets.
Conclusion: FINE provides an effective framework for representing complex physical systems with minimal latent space dimensions, offering interpretable, low-parameter, symmetry-preserving dimensional reduction that bridges Fourier analysis with neural architectures for scientific applications.
Abstract: We present the Fourier-Invertible Neural Encoder (FINE), a compact and interpretable architecture for dimension reduction in translation-equivariant datasets. FINE integrates reversible filters and monotonic activation functions with a Fourier truncation bottleneck, achieving information-preserving compression that respects translational symmetry. This design offers a new perspective on symmetry-aware learning, linking spectral truncation to group-equivariant representations. The proposed FINE architecture is tested on one-dimensional nonlinear wave interaction, one-dimensional Kuramoto-Sivashinsky turbulence dataset, and a two-dimensional turbulence dataset. FINE achieves an overall 4.9-9.1 times lower reconstruction error than convolutional autoencoders while using only 13-21% of their parameters. The results highlight FINE’s effectiveness in representing complex physical systems with minimal dimension in the latent space. The proposed framework provides a principled framework for interpretable, low-parameter, and symmetry-preserving dimensional reduction, bridging the gap between Fourier representations and modern neural architectures for scientific and physics-informed learning.
[860] Homeostatic Ubiquity of Hebbian Dynamics in Regularized Learning Rules
David Koplow, Tomaso Poggio, Liu Ziyin
Main category: cs.LG
TL;DR: L2 weight decay regularization causes learning signals to align with Hebbian direction at stationarity, and anti-Hebbian plasticity can arise from gradient/input noise in regression tasks.
Details
Motivation: To understand the theoretical basis of Hebbian and anti-Hebbian plasticity observed in biological brains, which remains limited despite widespread observation.Method: Analyzed learning methods with L2 weight decay regularization, showing learning signals align with Hebbian direction at stationarity. Also provided theoretical explanation for anti-Hebbian plasticity arising from gradient or input noise in regression tasks.
Result: Found that Hebbian-like behavior emerges not just with SGD but with almost any learning rule including random ones. Demonstrated how anti-Hebbian plasticity can naturally arise from noise in regression tasks, potentially explaining anti-Hebbian effects in the brain.
Conclusion: Proposed mechanisms for Hebbian and anti-Hebbian plasticity could coexist with established forms in the brain, highlighting the need for experimental methods to distinguish between different types of Hebbian observations in neurophysiology.
Abstract: Hebbian and anti-Hebbian plasticity are widely observed in the biological brain, yet their theoretical understanding remains limited. In this work, we find that when a learning method is regularized with L2 weight decay, its learning signal will gradually align with the direction of the Hebbian learning signal as it approaches stationarity. This Hebbian-like behavior is not unique to SGD: almost any learning rule, including random ones, can exhibit the same signature long before learning has ceased. We also provide a theoretical explanation for anti-Hebbian plasticity in regression tasks, demonstrating how it can arise naturally from gradient or input noise, and offering a potential reason for the observed anti-Hebbian effects in the brain. Certainly, our proposed mechanisms do not rule out any conventionally established forms of Hebbian plasticity and could coexist with them extensively in the brain. A key insight for neurophysiology is the need to develop ways to experimentally distinguish these two types of Hebbian observations.
[861] Towards Robust Influence Functions with Flat Validation Minima
Xichen Ye, Yifan Wu, Weizhong Zhang, Cheng Jin, Yifan Chen
Main category: cs.LG
TL;DR: The paper identifies that existing Influence Function methods fail in deep neural networks with noisy data due to deficiencies in loss change estimation from sharp validation risk, not parameter change estimation. It connects influence estimation error to validation risk sharpness and introduces a new Influence Function form for flat validation minima.
Details
Motivation: Existing Influence Function methods often provide unreliable influence estimates in deep neural networks, especially with noisy training data. Prior research focused on parameter change estimation accuracy, but the real problem lies in loss change estimation deficiencies caused by sharp validation risk.Method: The paper establishes a theoretical connection between influence estimation error, validation set risk, and its sharpness. It then introduces a novel estimation form of Influence Function specifically designed for flat validation minima.
Result: Experimental results across various tasks validate the superiority of the proposed approach over existing Influence Function methods.
Conclusion: Flat validation minima are crucial for accurate influence estimation, and the newly introduced Influence Function form designed for such minima provides more reliable influence estimates in deep neural networks with noisy data.
Abstract: The Influence Function (IF) is a widely used technique for assessing the impact of individual training samples on model predictions. However, existing IF methods often fail to provide reliable influence estimates in deep neural networks, particularly when applied to noisy training data. This issue does not stem from inaccuracies in parameter change estimation, which has been the primary focus of prior research, but rather from deficiencies in loss change estimation, specifically due to the sharpness of validation risk. In this work, we establish a theoretical connection between influence estimation error, validation set risk, and its sharpness, underscoring the importance of flat validation minima for accurate influence estimation. Furthermore, we introduce a novel estimation form of Influence Function specifically designed for flat validation minima. Experimental results across various tasks validate the superiority of our approach.
[862] Diffusion Models are Kelly Gamblers
Akhil Premkumar
Main category: cs.LG
TL;DR: Connects diffusion models to Kelly criterion for betting, showing diffusion models capture joint correlations in data and conditional models store mutual information between variables.
Details
Motivation: To establish a theoretical connection between diffusion models and the Kelly criterion from betting theory, and to understand what information diffusion models store about data correlations and conditional relationships.Method: Analytical framework connecting diffusion models to Kelly criterion, analyzing information storage in unconditional vs. conditional diffusion models, examining mutual information between data and conditioning variables.
Result: Unconditional diffusion models capture joint correlations between data components. Conditional models store additional mutual information between data X and conditioning Y, which is only a small fraction of total network information for low-dimensional data, yet doesn’t hinder conditional generation.
Conclusion: Diffusion models operate similarly to Kelly criterion by concentrating probability mass, and conditional generation works effectively despite storing relatively small mutual information because the network leverages the joint correlation structure already captured.
Abstract: We draw a connection between diffusion models and the Kelly criterion for maximizing returns in betting games. A signal that is correlated with the outcome of such a game can be used to focus the bets on a narrow range of high probability predictions. Diffusion models share the same paradigm in that they gradually concentrate the probability mass to fit the training data. We show that the information stored in an unconditional diffusion model captures, in part, the joint correlation between the components of the data variable $X$. Conditional diffusion models store additional information to bind the signal $X$ with the conditioning information $Y$, equal to the mutual information between them. The latter is only a small fraction of the total information in the neural network if the data is low-dimensional. We examine why this does not hinder conditional generation.
[863] In Search of Adam’s Secret Sauce
Antonio Orvieto, Robert M. Gower
Main category: cs.LG
TL;DR: Adam with equal momentum parameters (beta1=beta2) maintains near-optimal performance while enabling theoretical insights and statistical interpretation as a mean-field Gaussian variational inference algorithm.
Details
Motivation: To understand Adam's remarkable efficacy in training transformer-based language models and explore simplified variants to gain deeper optimization insights.Method: Extensive empirical study training over 1500 language models across different data configurations and scales, comparing Adam to several simplified variants including signed gradient and signed momentum methods.
Result: Signed momentum methods are faster than SGD but consistently underperform Adam. However, constraining Adam momentum parameters to be equal (beta1=beta2) preserves near-optimal performance while enabling theoretical insights.
Conclusion: Adam with beta1=beta2 is a compelling option that maintains performance while providing theoretical clarity: it implements a natural online algorithm for estimating gradient mean and variance from a mean-field Gaussian variational inference perspective.
Abstract: Understanding the remarkable efficacy of Adam when training transformer-based language models has become a central research topic within the optimization community. To gain deeper insights, several simplifications of Adam have been proposed, such as the signed gradient and signed momentum methods. In this work, we conduct an extensive empirical study - training over 1500 language models across different data configurations and scales - comparing Adam to several known simplified variants. We find that signed momentum methods are faster than SGD, but consistently underperform relative to Adam, even after careful tuning of momentum, clipping setting and learning rates. However, our analysis reveals a compelling option that preserves near-optimal performance while allowing for new insightful reformulations: constraining the Adam momentum parameters to be equal, beta1 = beta2. Beyond robust performance, this choice affords new theoretical insights, highlights the “secret sauce” on top of signed momentum, and grants a precise statistical interpretation: we show that Adam in this setting implements a natural online algorithm for estimating the mean and variance of gradients-one that arises from a mean-field Gaussian variational inference perspective.
[864] Vision Language Models are Biased
An Vo, Khai-Nguyen Nguyen, Mohammad Reza Taesiri, Vy Tuong Dang, Anh Totti Nguyen, Daeyoung Kim
Main category: cs.LG
TL;DR: VLMs show strong biases from prior knowledge that hurt performance on objective visual tasks like counting and identification, with accuracy dropping to ~17% across diverse domains.
Details
Motivation: While LLMs memorize useful knowledge from the internet, this same knowledge can cause biases that lead to wrong answers. The paper investigates how knowledge about popular subjects negatively affects VLMs' accuracy on objective visual tasks.Method: Tested state-of-the-art VLMs on standard visual tasks of counting and identification across 7 diverse domains (animals, logos, chess, board games, optical illusions, patterned grids). Used a human-supervised automated framework to evaluate biases, including experiments with background removal and analysis of reasoning patterns with varying thinking tokens.
Result: VLMs scored only 17.05% accuracy on counting tasks across domains. Removing image backgrounds nearly doubled accuracy (21.09 percentage point improvement). Counting accuracy initially improved with thinking tokens (reaching ~40%) but declined with excessive reasoning. The Adidas logo example showed VLMs couldn’t recognize when a 4th stripe was added to a 3-stripe logo.
Conclusion: VLMs exhibit interesting failure modes where prior knowledge biases their visual perception, demonstrating that contextual cues trigger biased responses. The work provides a framework for testing VLM biases and shows the need to address how memorized knowledge interferes with objective visual reasoning.
Abstract: Large language models (LLMs) memorize a vast amount of prior knowledge from the Internet that helps them on downstream tasks but also may notoriously sway their outputs towards wrong or biased answers. In this work, we test how the knowledge about popular subjects hurt the accuracy of vision language models (VLMs) on standard, objective visual tasks of counting and identification. We find that state-of-the-art VLMs are strongly biased (e.g., unable to recognize the 4th stripe has been added to a 3-stripe Adidas logo) scoring an average of 17.05% accuracy in counting (e.g., counting stripes in an Adidas-like logo) across 7 diverse domains from animals, logos, chess, board games, optical illusions, to patterned grids. Removing image backgrounds nearly doubles accuracy (21.09 percentage points), revealing that contextual visual cues trigger these biased responses. Further analysis of VLMs’ reasoning patterns shows that counting accuracy initially rises with thinking tokens, reaching ~40%, before declining with excessive reasoning. Our work presents an interesting failure mode in VLMs and a human-supervised automated framework for testing VLM biases. Code and data are available at: vlmsarebiased.github.io.
[865] Exploring System 1 and 2 communication for latent reasoning in LLMs
Julian Coda-Forno, Zhuokai Zhao, Qiang Zhang, Dipesh Tamboli, Weiwei Li, Xiangjun Fan, Lizhu Zhang, Eric Schulz, Hsiao-Ping Tseng
Main category: cs.LG
TL;DR: Dual-architecture latent reasoning (Base + Coprocessor) shows limited gains over unified single-model baseline; joint finetuning helps modestly but scaling latent tokens doesn’t improve robustness, suggesting current dual designs mostly add compute without qualitative reasoning improvements.
Details
Motivation: To determine whether LLM reasoning should be implemented in separate modules or within a single model's forward pass, testing if dual-architecture latent reasoning can improve over previous approaches.Method: Tested two hypotheses: (H1) increase channel capacity, and (H2) learn communication via joint finetuning. Compared dual-architecture (Base + Coprocessor exchanging latent messages) against unified soft-embedding baseline (single model with same forward pass and shared representations) under matched latent-token budgets on GPT-2 and Qwen-3 models.
Result: H2 (joint finetuning) was consistently strongest, while H1 yielded modest gains. The unified baseline nearly matched H2 and surpassed H1, suggesting dual designs mostly add compute rather than qualitatively improving reasoning. Scaling latent-token budget beyond small values failed to improve robustness across GSM8K, ProsQA, and Countdown stress tests. Latent analyses showed overlapping subspaces with limited specialization.
Conclusion: Dual-model latent reasoning remains promising in principle but likely requires objectives and training schedules that explicitly shape latent spaces for algorithmic planning, rather than just increasing capacity or adding compute.
Abstract: Should LLM reasoning live in a separate module, or within a single model’s forward pass and representational space? We study dual-architecture latent reasoning, where a fluent Base exchanges latent messages with a Coprocessor, and test two hypotheses aimed at improving latent communication over Liu et al. (2024): (H1) increase channel capacity; (H2) learn communication via joint finetuning. Under matched latent-token budgets on GPT-2 and Qwen-3, H2 is consistently strongest while H1 yields modest gains. A unified soft-embedding baseline, a single model with the same forward pass and shared representations, using the same latent-token budget, nearly matches H2 and surpasses H1, suggesting current dual designs mostly add compute rather than qualitatively improving reasoning. Across GSM8K, ProsQA, and a Countdown stress test with increasing branching factor, scaling the latent-token budget beyond small values fails to improve robustness. Latent analyses show overlapping subspaces with limited specialization, consistent with weak reasoning gains. We conclude dual-model latent reasoning remains promising in principle, but likely requires objectives and training schedules that explicitly shape latent spaces for algorithmic planning.
[866] Adaptive Plane Reformatting for 4D Flow MRI using Deep Reinforcement Learning
Javier Bisbal, Julio Sotelo, Maria I Valdés, Pablo Irarrazaval, Marcelo E Andia, Julio García, José Rodriguez-Palomarez, Francesca Raimondi, Cristián Tejos, Sergio Uribe
Main category: cs.LG
TL;DR: AdaPR is a deep reinforcement learning framework that enables accurate, orientation-independent plane reformatting for 4D flow MRI without requiring detailed landmarks or consistent spatial alignment across datasets.
Details
Motivation: Manual plane reformatting for 4D flow MRI is time-consuming and variable between observers. Current DRL methods assume consistent spatial alignment across datasets, limiting generalization across different scanners and institutions.Method: AdaPR uses a local coordinate system with the Asynchronous Advantage Actor-Critic (A3C) algorithm to navigate volumes with arbitrary positions and orientations, enabling adaptive plane reformatting without requiring detailed anatomical landmarks.
Result: AdaPR achieved mean angular error of 6.32° ± 4.15° and distance error of 3.40 ± 2.75 mm, outperforming global-coordinate DRL and non-DRL methods. Flow measurements showed excellent correlation with manual observers (R² = 0.972 and 0.968), comparable to inter-observer agreement.
Conclusion: AdaPR provides robust, orientation-independent plane reformatting for 4D flow MRI with flow quantification comparable to expert observers. Its adaptability across datasets and scanners makes it promising for broader medical imaging applications.
Abstract: Background and Objective: Plane reformatting for four-dimensional phase contrast MRI (4D flow MRI) is time-consuming and prone to inter-observer variability, which limits fast cardiovascular flow assessment. Deep reinforcement learning (DRL) trains agents to iteratively adjust plane position and orientation, enabling accurate plane reformatting without the need for detailed landmarks, making it suitable for images with limited contrast and resolution such as 4D flow MRI. However, current DRL methods assume that test volumes share the same spatial alignment as the training data, limiting generalization across scanners and institutions. To address this limitation, we introduce AdaPR (Adaptive Plane Reformatting), a DRL framework that uses a local coordinate system to navigate volumes with arbitrary positions and orientations. Methods: We implemented AdaPR using the Asynchronous Advantage Actor-Critic (A3C) algorithm and validated it on 88 4D flow MRI datasets acquired from multiple vendors, including patients with congenital heart disease. Results: AdaPR achieved a mean angular error of 6.32 +/- 4.15 degrees and a distance error of 3.40 +/- 2.75 mm, outperforming global-coordinate DRL methods and alternative non-DRL methods. AdaPR maintained consistent accuracy under different volume orientations and positions. Flow measurements from AdaPR planes showed no significant differences compared to two manual observers, with excellent correlation (R^2 = 0.972 and R^2 = 0.968), comparable to inter-observer agreement (R^2 = 0.969). Conclusion: AdaPR provides robust, orientation-independent plane reformatting for 4D flow MRI, achieving flow quantification comparable to expert observers. Its adaptability across datasets and scanners makes it a promising candidate for medical imaging applications beyond 4D flow MRI.
[867] Connecting Neural Models Latent Geometries with Relative Geodesic Representations
Hanlin Yu, Berfin Inal, Georgios Arvanitidis, Soren Hauberg, Francesco Locatello, Marco Fumero
Main category: cs.LG
TL;DR: The paper proposes a method to capture transformations between different neural model representations by exploiting their differential-geometric structure and assuming they parametrize the same underlying manifold.
Details
Motivation: Different neural models learning the same task on the same data can produce different representations due to various factors (training stochasticity, architectures, inductive biases). However, when these models share an underlying latent structure, their representations may preserve relative distances up to distortions. The authors aim to capture the precise transformations between such representational spaces.Method: The method assumes distinct neural models parametrize approximately the same underlying manifold. It introduces a representation based on the pullback metric that captures the intrinsic structure of the latent space while scaling efficiently to large models. This differential-geometric approach exploits the structure of latent spaces to capture transformations between representational spaces trained on similar data distributions.
Result: The method is experimentally validated on model stitching and retrieval tasks, covering autoencoders and vision foundation discriminative models across diverse architectures, datasets, and pretraining schemes.
Conclusion: By leveraging the differential-geometric structure of latent spaces and the pullback metric representation, it’s possible to capture precise transformations between different neural model representations that share an underlying manifold structure, enabling effective model stitching and retrieval across diverse model configurations.
Abstract: Neural models learn representations of high-dimensional data on low-dimensional manifolds. Multiple factors, including stochasticities in the training process, model architectures, and additional inductive biases, may induce different representations, even when learning the same task on the same data. However, it has recently been shown that when a latent structure is shared between distinct latent spaces, relative distances between representations can be preserved, up to distortions. Building on this idea, we demonstrate that exploiting the differential-geometric structure of latent spaces of neural models, it is possible to capture precisely the transformations between representational spaces trained on similar data distributions. Specifically, we assume that distinct neural models parametrize approximately the same underlying manifold, and introduce a representation based on the pullback metric that captures the intrinsic structure of the latent space, while scaling efficiently to large models. We validate experimentally our method on model stitching and retrieval tasks, covering autoencoders and vision foundation discriminative models, across diverse architectures, datasets, and pretraining schemes.
[868] Multiscale guidance of protein structure prediction with heterogeneous cryo-EM data
Rishwanth Raghu, Axel Levy, Gordon Wetzstein, Ellen D. Zhong
Main category: cs.LG
TL;DR: CryoBoltz combines cryo-EM density maps with protein structure prediction models to capture conformational diversity in dynamic biomolecular complexes.
Details
Motivation: Protein structure prediction models fail to capture conformational diversity of dynamic complexes, while cryo-EM struggles with transforming raw data into atomic models. There's a need to bridge these modalities.Method: CryoBoltz guides sampling of pretrained structure prediction models using global and local structural constraints from cryo-EM density maps, driving predictions toward experimental data-consistent conformational states.
Result: The method successfully builds atomic models into heterogeneous cryo-EM maps across various dynamic systems including transporters and antibodies.
Conclusion: CryoBoltz provides a flexible inference-time approach that combines experimental cryo-EM data with learned biophysical priors to model conformational diversity in biomolecular complexes.
Abstract: Protein structure prediction models are now capable of generating accurate 3D structural hypotheses from sequence alone. However, they routinely fail to capture the conformational diversity of dynamic biomolecular complexes, often requiring heuristic MSA subsampling approaches for generating alternative states. In parallel, cryo-electron microscopy (cryo-EM) has emerged as a powerful tool for imaging near-native structural heterogeneity, but is challenged by arduous pipelines to transform raw experimental data into atomic models. Here, we bridge the gap between these modalities, combining cryo-EM density maps with the rich sequence and biophysical priors learned by protein structure prediction models. Our method, CryoBoltz, guides the sampling trajectory of a pretrained biomolecular structure prediction model using both global and local structural constraints derived from density maps, driving predictions towards conformational states consistent with the experimental data. We demonstrate that this flexible yet powerful inference-time approach allows us to build atomic models into heterogeneous cryo-EM maps across a variety of dynamic biomolecular systems including transporters and antibodies. Code is available at https://github.com/ml-struct-bio/cryoboltz .
[869] State Entropy Regularization for Robust Reinforcement Learning
Yonatan Ashlag, Uri Koren, Mirco Mutti, Esther Derman, Pierre-Luc Bacon, Shie Mannor
Main category: cs.LG
TL;DR: State entropy regularization improves robustness to structured perturbations in RL, unlike standard methods that focus on small uncorrelated changes.
Details
Motivation: State entropy regularization has shown empirical benefits in exploration and sample complexity, but lacks theoretical understanding. The paper aims to provide theoretical guarantees for its robustness properties, particularly against structured and spatially correlated perturbations common in transfer learning.Method: The paper provides a comprehensive theoretical characterization of state entropy regularization’s robustness properties, including formal guarantees under reward and transition uncertainty. It contrasts state entropy with policy entropy regularization to highlight their different benefits.
Result: State entropy regularization improves robustness to structured and spatially correlated perturbations, which are common in transfer learning. The analysis shows formal guarantees under reward and transition uncertainty, and identifies settings where the method performs poorly. The robustness advantages are more sensitive to the number of rollouts used for policy evaluation compared to policy entropy.
Conclusion: State entropy regularization provides theoretical robustness guarantees against structured perturbations, addressing limitations of standard robust RL methods. The method’s benefits differ from policy entropy regularization, with practical implications for rollout sensitivity in policy evaluation.
Abstract: State entropy regularization has empirically shown better exploration and sample complexity in reinforcement learning (RL). However, its theoretical guarantees have not been studied. In this paper, we show that state entropy regularization improves robustness to structured and spatially correlated perturbations. These types of variation are common in transfer learning but often overlooked by standard robust RL methods, which typically focus on small, uncorrelated changes. We provide a comprehensive characterization of these robustness properties, including formal guarantees under reward and transition uncertainty, as well as settings where the method performs poorly. Much of our analysis contrasts state entropy with the widely used policy entropy regularization, highlighting their different benefits. Finally, from a practical standpoint, we illustrate that compared with policy entropy, the robustness advantages of state entropy are more sensitive to the number of rollouts used for policy evaluation.
[870] Implicit Models: Expressive Power Scales with Test-Time Compute
Jialin Liu, Lisang Ding, Stanley Osher, Wotao Yin
Main category: cs.LG
TL;DR: Implicit models use iterative fixed-point computation with constant memory, achieving performance scaling with test-time compute through progressive complexity expression.
Details
Motivation: To understand why compact implicit models can match or exceed larger explicit networks by allocating more test-time compute, despite the underlying mechanism being poorly understood.Method: Nonparametric analysis of expressive power, providing strict mathematical characterization showing that simple implicit operators can progressively express more complex mappings through iteration, with expressive power scaling with test-time compute.
Result: Proved that implicit models’ expressive power scales with test-time compute to match richer function classes. Validated across four domains: image reconstruction, scientific computing, operations research, and LLM reasoning, showing increased mapping complexity and improved/stabilized solution quality with more iterations.
Conclusion: Implicit models can achieve performance comparable to larger explicit networks by leveraging test-time compute to progressively express more complex mappings, with their expressive power mathematically proven to scale with computational allocation.
Abstract: Implicit models, an emerging model class, compute outputs by iterating a single parameter block to a fixed point. This architecture realizes an infinite-depth, weight-tied network that trains with constant memory, significantly reducing memory needs for the same level of performance compared to explicit models. While it is empirically known that these compact models can often match or even exceed the accuracy of larger explicit networks by allocating more test-time compute, the underlying mechanism remains poorly understood. We study this gap through a nonparametric analysis of expressive power. We provide a strict mathematical characterization, showing that a simple and regular implicit operator can, through iteration, progressively express more complex mappings. We prove that for a broad class of implicit models, this process lets the model’s expressive power scale with test-time compute, ultimately matching a much richer function class. The theory is validated across four domains: image reconstruction, scientific computing, operations research, and LLM reasoning, demonstrating that as test-time iterations increase, the complexity of the learned mapping rises, while the solution quality simultaneously improves and stabilizes.
[871] AWP: Activation-Aware Weight Pruning and Quantization with Projected Gradient Descent
Jing Liu, Toshiaki Koike-Akino, Ye Wang, Hassan Mansour, Matthew Brand
Main category: cs.LG
TL;DR: A unified method called AWP (Activation-aware Weight pruning and quantization via Projected gradient descent) is proposed for LLM compression, outperforming state-of-the-art methods with theoretical convergence guarantees.
Details
Motivation: To address the enormous size of Large Language Models (LLMs) for deployment on edge devices, model compression methods like quantization and pruning are needed. The paper focuses on layer-wise post-training quantization and pruning.Method: The authors propose AWP, a unified method for activation-aware weight pruning and quantization via projected gradient descent. They draw connections between activation-aware weight pruning and sparse approximation problems, and are motivated by the success of Iterative Hard Thresholding (IHT).
Result: Experiments demonstrate that AWP outperforms state-of-the-art LLM pruning and quantization methods. Theoretical convergence guarantees for the proposed pruning method are also provided.
Conclusion: AWP provides an effective unified approach for LLM compression through activation-aware weight pruning and quantization, with superior performance over existing methods and theoretical convergence guarantees.
Abstract: To address the enormous size of Large Language Models (LLMs), model compression methods, such as quantization and pruning, are often deployed, especially on edge devices. In this work, we focus on layer-wise post-training quantization and pruning. Drawing connections between activation-aware weight pruning and sparse approximation problems, and motivated by the success of Iterative Hard Thresholding (IHT), we propose a unified method for Activation-aware Weight pruning and quantization via Projected gradient descent (AWP). Our experiments demonstrate that AWP outperforms state-of-the-art LLM pruning and quantization methods. Theoretical convergence guarantees of the proposed method for pruning are also provided.
[872] CORAL: Disentangling Latent Representations in Long-Tailed Diffusion
Esther Rodriguez, Monica Welfert, Samuel McDowell, Nathan Stromberg, Julian Antolin Camarena, Lalitha Sankar
Main category: cs.LG
TL;DR: CORAL improves diffusion model performance on long-tailed datasets by using contrastive regularization to separate latent class representations, addressing overlap between head and tail class features.
Details
Motivation: Standard diffusion models struggle with long-tailed datasets, producing low-diversity and lower-quality samples for tail classes. The underlying cause of this degradation is poorly understood, despite being well-documented.Method: Proposes CORAL (COntrastive Regularization for Aligning Latents), a contrastive latent alignment framework that uses supervised contrastive losses to encourage well-separated latent class representations in the U-Net bottleneck layer.
Result: CORAL significantly improves both diversity and visual quality of samples generated for tail classes relative to state-of-the-art methods.
Conclusion: The key issue in long-tailed diffusion models is latent representation overlap between head and tail classes, which CORAL effectively addresses through contrastive regularization, leading to better generation quality for underrepresented classes.
Abstract: Diffusion models have achieved impressive performance in generating high-quality and diverse synthetic data. However, their success typically assumes a class-balanced training distribution. In real-world settings, multi-class data often follow a long-tailed distribution, where standard diffusion models struggle – producing low-diversity and lower-quality samples for tail classes. While this degradation is well-documented, its underlying cause remains poorly understood. In this work, we investigate the behavior of diffusion models trained on long-tailed datasets and identify a key issue: the latent representations (from the bottleneck layer of the U-Net) for tail class subspaces exhibit significant overlap with those of head classes, leading to feature borrowing and poor generation quality. Importantly, we show that this is not merely due to limited data per class, but that the relative class imbalance significantly contributes to this phenomenon. To address this, we propose COntrastive Regularization for Aligning Latents (CORAL), a contrastive latent alignment framework that leverages supervised contrastive losses to encourage well-separated latent class representations. Experiments demonstrate that CORAL significantly improves both the diversity and visual quality of samples generated for tail classes relative to state-of-the-art methods.
[873] NeuroRVQ: Multi-Scale EEG Tokenization for Generative Large Brainwave Models
Konstantinos Barmpas, Na Lee, Alexandros Koliousis, Yannis Panagakis, Dimitrios A. Adamos, Nikolaos Laskaris, Stefanos Zafeiriou
Main category: cs.LG
TL;DR: NeuroRVQ introduces a scalable Large Brainwave Model with a novel codebook-based tokenizer that preserves high-frequency EEG dynamics through multi-scale feature extraction, hierarchical residual vector quantization, and phase/amplitude-aware loss, achieving better reconstruction and downstream task performance.
Details
Motivation: Existing EEG foundation models have limited performance due to tokenization modules that fail to preserve high-frequency dynamics, hindering accurate EEG signal reconstruction and representation learning.Method: Developed NeuroRVQ with a codebook-based tokenizer featuring: (1) multi-scale feature extraction to capture full frequency spectrum, (2) hierarchical residual vector quantization codebooks for high-resolution encoding, and (3) phase- and amplitude-aware loss function for efficient training.
Result: NeuroRVQ achieves lower reconstruction error and outperforms existing Large Brainwave Models on various downstream tasks, establishing a strong prior for codebook-based general-purpose brainwave models.
Conclusion: NeuroRVQ provides an effective tokenizer for EEG foundation models that enables efficient compression while supporting accurate reconstruction across all frequency bands, advancing neural decoding, generative modeling, and multimodal biosignal integration.
Abstract: Electroencephalography (EEG) captures neural activity across multiple temporal and spectral scales, yielding signals that are rich but complex for representation learning. Recently, EEG foundation models trained to predict masked signal-tokens have shown promise for learning generalizable representations. However, their performance is hindered by their signal tokenization modules. Existing neural tokenizers fail to preserve high-frequency dynamics, limiting their ability to reconstruct EEG signals with high fidelity. We introduce NeuroRVQ, a scalable Large Brainwave Model (LBM) centered on a codebook-based tokenizer. Our tokenizer integrates: (i) multi-scale feature extraction modules that capture the full frequency neural spectrum; (ii) hierarchical residual vector quantization (RVQ) codebooks for high-resolution encoding; and, (iii) an EEG signal phase- and amplitude-aware loss function for efficient training. This design enables efficient EEG compression while supporting accurate reconstruction across all frequency bands, leading to robust generative masked modeling. Our empirical results demonstrate that NeuroRVQ achieves lower reconstruction error and outperforms existing LBMs on a variety of downstream tasks. More broadly, NeuroRVQ tokenizer establishes a strong prior for codebook-based general-purpose brainwave models, enabling advances in neural decoding, generative modeling and multimodal biosignal integration.
[874] Tensor-Parallelism with Partially Synchronized Activations
Itay Lamprecht, Asaf Karnieli, Yair Hanani, Niv Giladi, Daniel Soudry
Main category: cs.LG
TL;DR: CAAT-Net reduces tensor-parallel communication by up to 50% in LLM training/inference without significant accuracy loss, using communication-aware architecture design.
Details
Motivation: Current tensor-parallel training of LLMs requires substantial communication overhead for activation synchronization, which creates bandwidth bottlenecks.Method: Communication-Aware Architecture for Tensor-parallelism (CAAT-Net) - minor adjustments to current practices to train LLMs without fully synchronizing activations.
Result: 50% reduction in tensor-parallel communication with no significant pretraining accuracy drop across benchmarks; validation loss sometimes improves; demonstrated on 130M, 1.1B, and 7B parameter models.
Conclusion: CAAT-Net enables more efficient LLM training and inference by reducing communication overhead while maintaining or improving model performance, showing robustness and scalability across model sizes.
Abstract: Training and inference of Large Language Models (LLMs) with tensor-parallelism requires substantial communication to synchronize activations. Our findings suggest that with a few minor adjustments to current practices, LLMs can be trained without fully synchronizing activations, reducing bandwidth demands. We name this “Communication-Aware Architecture for Tensor-parallelism” (CAAT-Net). We train a 7B parameter CAAT-Net model and show that tensor-parallel communication can be reduced by up to 50% with no significant drop in pretraining accuracy across nearly all evaluated benchmarks. We also experiment with smaller 130M and 1.1B models to show the robustness and scalability of our method. We find that, in some scenarios, validation loss can even improve when reducing communication. Finally, we demonstrate how CAAT-Net accelerates both training and inference workloads across various settings and model sizes.
[875] Efficient Low Rank Attention for Long-Context Inference in Large Language Models
Tenghui Li, Guoxu Zhou, Xuyang Zhao, Yuning Qiu, Qibin Zhao
Main category: cs.LG
TL;DR: LRQK is a two-stage framework that decomposes query and key matrices into low-rank factors to reduce KV cache memory usage while maintaining exact attention outputs through a mixed GPU-CPU cache system.
Details
Motivation: As input text grows, KV cache in LLMs imposes prohibitive GPU memory costs, limiting long-context inference on resource-constrained devices. Existing approaches like KV quantization and pruning reduce memory but suffer from precision loss or suboptimal KV pair retention.Method: Two-stage framework: 1) Jointly decomposes full-precision query and key matrices into compact rank-r factors during prefill stage; 2) Uses low-dimensional projections to compute proxy attention scores in O(lr) time at each decode step. Employs mixed GPU-CPU cache with hit-and-miss mechanism that transfers only missing full-precision KV pairs, selecting top-k tokens and recent tokens.
Result: Extensive experiments on RULER and LongBench benchmarks with LLaMA-3-8B and Qwen2.5-7B show LRQK matches or surpasses leading sparse-attention methods in long context settings while delivering significant memory savings with minimal accuracy loss.
Conclusion: LRQK effectively addresses KV cache memory limitations for long-context LLM inference, preserving exact attention outputs while reducing memory usage and CPU-GPU data movement, making it suitable for resource-constrained devices.
Abstract: As the length of input text grows, the key-value (KV) cache in LLMs imposes prohibitive GPU memory costs and limits long-context inference on resource constrained devices. Existing approaches, such as KV quantization and pruning, reduce memory usage but suffer from numerical precision loss or suboptimal retention of key-value pairs. We introduce Low Rank Query and Key attention (LRQK), a two-stage framework that jointly decomposes the full-precision query and key matrices into compact rank-(r) factors during the prefill stage, and then uses these low-dimensional projections to compute proxy attention scores in (\mathcal{O}(lr)) time at each decode step. By selecting only the top-(k) tokens and a small fixed set of recent tokens, LRQK employs a mixed GPU-CPU cache with a hit-and-miss mechanism that transfers only missing full-precision KV pairs, thereby preserving exact attention outputs while reducing CPU-GPU data movement. Extensive experiments on the RULER and LongBench benchmarks with LLaMA-3-8B and Qwen2.5-7B demonstrate that LRQK matches or surpasses leading sparse-attention methods in long context settings, while delivering significant memory savings with minimal loss in accuracy. Our code is available at https://github.com/tenghuilee/LRQK.
[876] Bridging Prediction and Intervention Problems in Social Systems
Lydia T. Liu, Inioluwa Deborah Raji, Angela Zhou, Luke Guerdan, Jessica Hullman, Daniel Malinsky, Bryan Wilder, Simone Zhang, Hammaad Adam, Amanda Coston, Ben Laufer, Ezinne Nwankwo, Michael Zanger-Tishler, Eli Ben-Michael, Solon Barocas, Avi Feller, Marissa Gerchick, Talia Gillis, Shion Guha, Daniel Ho, Lily Hu, Kosuke Imai, Sayash Kapoor, Joshua Loftus, Razieh Nabi, Arvind Narayanan, Ben Recht, Juan Carlos Perdomo, Matthew Salganik, Mark Sendak, Alexander Tolbert, Berk Ustun, Suresh Venkatasubramanian, Angelina Wang, Ashia Wilson
Main category: cs.LG
TL;DR: The paper argues for shifting from a prediction-focused paradigm to an interventionist paradigm for automated decision systems, emphasizing that ADS operationalize policy interventions in deployment and shape population outcomes.
Details
Motivation: Current automated decision systems are designed as prediction problems but actually function as policy interventions in deployment, shaping population outcomes through decision-making changes. The prediction-focused paradigm fails to account for how ADS interact with social systems, organizational infrastructure, and stakeholder dynamics.Method: The paper proposes a new default problem setup for ADS that moves beyond prediction to consider predictions as decision support, final decisions, and outcomes. It highlights how modern statistical frameworks and tools can be unified to study ADS design, implementation, and evaluation from an interventionist perspective.
Result: The analysis characterizes the limitations of focusing on isolated prediction tasks and lays the foundation for a more intervention-oriented approach to developing and deploying ADS. It identifies necessary research directions to operationalize this paradigm shift.
Conclusion: A paradigm shift from prediction to intervention is essential for properly understanding and deploying automated decision systems within social systems, requiring new frameworks that account for how ADS shape outcomes through policy changes and stakeholder interactions.
Abstract: Many automated decision systems (ADS) are designed to solve prediction problems – where the goal is to learn patterns from a sample of the population and apply them to individuals from the same population. In reality, these prediction systems operationalize holistic policy interventions in deployment. Once deployed, ADS can shape impacted population outcomes through an effective policy change in how decision-makers operate, while also being defined by past and present interactions between stakeholders and the limitations of existing organizational, as well as societal, infrastructure and context. In this work, we consider the ways in which we must shift from a prediction-focused paradigm to an interventionist paradigm when considering the impact of ADS within social systems. We argue this requires a new default problem setup for ADS beyond prediction, to instead consider predictions as decision support, final decisions, and outcomes. We highlight how this perspective unifies modern statistical frameworks and other tools to study the design, implementation, and evaluation of ADS systems, and point to the research directions necessary to operationalize this paradigm shift. Using these tools, we characterize the limitations of focusing on isolated prediction tasks, and lay the foundation for a more intervention-oriented approach to developing and deploying ADS.
[877] Quantile Reward Policy Optimization: Alignment with Pointwise Regression and Exact Partition Functions
Simon Matrenok, Skander Moalla, Caglar Gulcehre
Main category: cs.LG
TL;DR: QRPO enables offline learning from absolute rewards (not just preferences) using quantile rewards, matching PPO performance with DPO simplicity.
Details
Motivation: Current offline methods like DPO and REBEL require preference pairs, while online methods like PPO can use absolute rewards but are complex. Need a simple offline method that works with absolute rewards.Method: QRPO uses quantile rewards to enable regression to KL-regularized RL solution. Quantile rewards yield tractable partition function, eliminating need for relative signals. Scales with compute for quantile estimation.
Result: Consistently achieves top performance on chat/coding evaluations (reward model scores, AlpacaEval 2, LeetCode) vs DPO, REBEL, SimPO across diverse datasets and 8B models. Reduces length bias.
Conclusion: QRPO bridges gap between absolute reward learning and offline simplicity, offering new pre-computation scaling dimension while reducing length bias.
Abstract: Aligning large language models with pointwise absolute rewards has so far required online, on-policy algorithms such as PPO and GRPO. In contrast, simpler methods that can leverage offline or off-policy data, such as DPO and REBEL, are limited to learning from preference pairs or relative signals. To bridge this gap, we introduce Quantile Reward Policy Optimization (QRPO), which learns from pointwise absolute rewards while preserving the simplicity and offline applicability of DPO-like methods. QRPO uses quantile rewards to enable regression to the closed-form solution of the KL-regularized RL objective. This reward yields an analytically tractable partition function, removing the need for relative signals to cancel this term. Moreover, QRPO scales with increased compute to estimate quantile rewards, opening a new dimension for pre-computation scaling. Empirically, QRPO consistently achieves top performance on chat and coding evaluations–reward model scores, AlpacaEval 2, and LeetCode–compared to DPO, REBEL, and SimPO across diverse datasets and 8B-scale models. Finally, we find that training with robust rewards instead of converting them to preferences induces less length bias.
[878] Flow Equivariant Recurrent Neural Networks
T. Anderson Keller
Main category: cs.LG
TL;DR: Extends equivariant network theory to handle continuous time-parameterized symmetries (flows) in sequence models, showing how to make RNNs flow-equivariant and demonstrating improved performance.
Details
Motivation: Real-world data arrives as continuous streams with smooth transformations over time, representing continuous symmetries. While equivariant networks have benefits for static transformations, they haven't been applied to sequence models and time-parameterized transformations.Method: Extends equivariant network theory to handle ‘flows’ (one-parameter Lie subgroups capturing natural transformations over time). Shows standard RNNs are not flow-equivariant and demonstrates how to introduce flow equivariance into sequence models.
Result: Flow-equivariant models significantly outperform non-equivariant counterparts in training speed, length generalization, and velocity generalization on both next-step prediction and sequence classification tasks.
Conclusion: This work represents a first step toward building sequence models that respect the time-parameterized symmetries governing real-world data, extending equivariance from static to dynamic transformations.
Abstract: Data arrives at our senses as a continuous stream, smoothly transforming from one instant to the next. These smooth transformations can be viewed as continuous symmetries of the environment that we inhabit, defining equivalence relations between stimuli over time. In machine learning, neural network architectures that respect symmetries of their data are called equivariant and have provable benefits in terms of generalization ability and sample efficiency. To date, however, equivariance has been considered only for static transformations and feed-forward networks, limiting its applicability to sequence models, such as recurrent neural networks (RNNs), and corresponding time-parameterized sequence transformations. In this work, we extend equivariant network theory to this regime of ‘flows’ – one-parameter Lie subgroups capturing natural transformations over time, such as visual motion. We begin by showing that standard RNNs are generally not flow equivariant: their hidden states fail to transform in a geometrically structured manner for moving stimuli. We then show how flow equivariance can be introduced, and demonstrate that these models significantly outperform their non-equivariant counterparts in terms of training speed, length generalization, and velocity generalization, on both next step prediction and sequence classification. We present this work as a first step towards building sequence models that respect the time-parameterized symmetries which govern the world around us.
[879] Predictive Scaling Laws for Efficient GRPO Training of Large Reasoning Models
Datta Nimmaturi, Vaishnavi Bhargava, Rajat Ghosh, Johnu George, Debojyoti Dutta
Main category: cs.LG
TL;DR: The paper proposes a predictive framework to optimize compute resources for fine-tuning LLMs using GRPO, identifying empirical scaling laws and training phases to enable earlier stopping without performance loss.
Details
Motivation: Fine-tuning large language models for reasoning tasks using reinforcement learning methods like GRPO is computationally expensive, creating a need for more efficient resource optimization strategies.Method: Developed a predictive framework that models training dynamics through experiments on Llama and Qwen models (3B and 8B sizes), deriving empirical scaling laws based on model size, initial performance, and training progress.
Result: Identified three consistent training phases (slow start, rapid improvement, plateau) and found that training beyond certain epochs offers minimal gains, enabling earlier stopping to reduce compute by up to 50% without performance sacrifice.
Conclusion: The predictive framework generalizes across model types and provides practical guidance for efficient GRPO-based fine-tuning, significantly reducing computational costs while maintaining performance.
Abstract: Fine-tuning large language models (LLMs) for reasoning tasks using reinforcement learning methods like Group Relative Policy Optimization (GRPO) is computationally expensive. To address this, we propose a predictive framework that models training dynamics and helps optimize resource usage. Through experiments on Llama and Qwen models (3B 8B), we derive an empirical scaling law based on model size, initial performance, and training progress. This law predicts reward trajectories and identifies three consistent training phases: slow start, rapid improvement, and plateau. We find that training beyond certain number of an epoch offers little gain, suggesting earlier stopping can significantly reduce compute without sacrificing performance. Our approach generalizes across model types, providing a practical guide for efficient GRPO-based fine-tuning.
[880] Optimal Scheduling Algorithms for LLM Inference: Theory and Practice
Agrim Bari, Parikshit Hegde, Gustavo de Veciana
Main category: cs.LG
TL;DR: The paper proposes SLAI, an SLO-aware scheduler for LLM inference systems that optimizes throughput while meeting latency constraints for both prefill (TTFT) and decode (TBT) phases.
Details
Motivation: LLM inference systems have a unique two-phase computation structure (prefill and decode) requiring new routing and scheduling strategies to handle growing industry demand efficiently.Method: Developed theoretical framework for LLM inference scheduling, identified optimal tiling and dynamic resource allocation principles, proposed RAD scheduler for throughput optimality, and designed SLAI scheduler with real-time measurements to prioritize decode requests near TBT deadlines and reorder prefill requests based on prompt lengths.
Result: On Openchat ShareGPT4 dataset with Mistral-7B model on NVIDIA RTX ADA 6000 GPU, SLAI reduces median TTFT by 53% and increases maximum serving capacity by 26% while meeting tail TBT latency constraints.
Conclusion: SLAI effectively addresses practical SLO requirements for LLM inference systems by optimizing both throughput and latency constraints through intelligent scheduling of prefill and decode phases.
Abstract: With the growing use of Large Language Model (LLM)-based tools like ChatGPT, Perplexity, and Gemini across industries, there is a rising need for efficient LLM inference systems. These systems handle requests with a unique two-phase computation structure: a prefill-phase that processes the full input prompt and a decode-phase that autoregressively generates tokens one at a time. This structure calls for new strategies for routing and scheduling requests. In this paper, we take a comprehensive approach to this challenge by developing a theoretical framework that models routing and scheduling in LLM inference systems. We identify two key design principles-optimal tiling and dynamic resource allocation-that are essential for achieving high throughput. Guided by these principles, we propose the Resource-Aware Dynamic (RAD) scheduler and prove that it achieves throughput optimality under mild conditions. To address practical Service Level Objectives (SLOs) such as serving requests with different Time Between Token (TBT) constraints, we design the SLO-Aware LLM Inference (SLAI) scheduler. SLAI uses real-time measurements to prioritize decode requests that are close to missing their TBT deadlines and reorders prefill requests based on known prompt lengths to further reduce the Time To First Token (TTFT) delays. We evaluate SLAI on the Openchat ShareGPT4 dataset using the Mistral-7B model on an NVIDIA RTX ADA 6000 GPU. Compared to Sarathi-Serve, SLAI reduces the median TTFT by 53% and increases the maximum serving capacity by 26% such that median TTFT is below 0.5 seconds, while meeting tail TBT latency constraints.
[881] Signals, Concepts, and Laws: Toward Universal, Explainable Time-Series Forecasting
Hongwei Ma, Junbin Gao, Minh-Ngoc Tran
Main category: cs.LG
TL;DR: DORIC is a domain-universal transformer for time-series forecasting that uses self-supervised concepts and enforces ODE-based physical constraints for interpretable and physically credible predictions.
Details
Motivation: Addressing the challenge of accurate, explainable, and physically credible forecasting for multivariate time-series with varying statistical properties across different domains.Method: Proposes DORIC: a Domain-Universal, ODE-Regularized, Interpretable-Concept Transformer that generates predictions through five self-supervised, domain-agnostic concepts while enforcing differentiable residuals grounded in first-principles constraints.
Result: Not specified in the abstract (paper likely presents experimental results demonstrating improved forecasting accuracy, interpretability, and physical credibility across multiple domains).
Conclusion: DORIC offers a novel approach to time-series forecasting that combines domain universality with interpretability and physical credibility through ODE-regularized concept learning.
Abstract: Accurate, explainable and physically credible forecasting remains a persistent challenge for multivariate time-series whose statistical properties vary across domains. We propose DORIC, a Domain-Universal, ODE-Regularized, Interpretable-Concept Transformer for Time-Series Forecasting that generates predictions through five self-supervised, domain-agnostic concepts while enforcing differentiable residuals grounded in first-principles constraints.
[882] Addressing divergent representations from causal interventions on neural networks
Satchel Grant, Simon Jerome Han, Alexa R. Tartaglini, Christopher Potts
Main category: cs.LG
TL;DR: Causal interventions in mechanistic interpretability often create out-of-distribution representations, which can be harmless (behavioral null-space) or pernicious (activating hidden pathways). The paper proposes using Counterfactual Latent loss to keep interventions closer to natural distributions while preserving interpretability.
Details
Motivation: To investigate whether causal interventions in mechanistic interpretability create out-of-distribution representations that may compromise the faithfulness of explanations to the model's natural state.Method: 1) Theoretical and empirical demonstration that common causal interventions shift representations away from natural distributions; 2) Theoretical analysis of harmless vs. pernicious divergences; 3) Application and modification of Counterfactual Latent loss to mitigate pernicious divergences.
Result: Causal interventions often create divergent representations. Harmless divergences occur in behavioral null-space, while pernicious ones activate hidden pathways causing dormant behavioral changes. CL loss helps keep interventions closer to natural distributions.
Conclusion: The paper highlights the importance of addressing distributional shifts in interpretability methods and proposes a path toward more reliable interpretability through modified intervention techniques that preserve natural distributions.
Abstract: A common approach to mechanistic interpretability is to causally manipulate model representations via targeted interventions in order to understand what those representations encode. Here we ask whether such interventions create out-of-distribution (divergent) representations, and whether this raises concerns about how faithful their resulting explanations are to the target model in its natural state. First, we demonstrate theoretically and empirically that common causal intervention techniques often do shift internal representations away from the natural distribution of the target model. Then, we provide a theoretical analysis of two cases of such divergences: “harmless” divergences that occur in the behavioral null-space of the layer(s) of interest, and “pernicious” divergences that activate hidden network pathways and cause dormant behavioral changes. Finally, in an effort to mitigate the pernicious cases, we apply and modify the Counterfactual Latent (CL) loss from Grant (2025) allowing representations from causal interventions to remain closer to the natural distribution, reducing the likelihood of harmful divergences while preserving the interpretive power of the interventions. Together, these results highlight a path towards more reliable interpretability methods.
[883] Revitalizing Canonical Pre-Alignment for Irregular Multivariate Time Series Forecasting
Ziyu Zhou, Yiming Huang, Yanyun Wang, Yuankai Wu, James Kwok, Yuxuan Liang
Main category: cs.LG
TL;DR: KAFNet is a compact architecture for irregular multivariate time series forecasting that retains canonical pre-alignment but addresses its computational overhead through learnable compression and frequency-domain processing.
Details
Motivation: Irregular multivariate time series (IMTS) with uneven sampling and inter-variate asynchrony are challenging to model efficiently. While canonical pre-alignment (CPA) helps unify series length, its dense zero-padding causes prohibitive compute overhead, especially with many variates. Recent graph-based models avoid CPA but struggle to capture global inter-variate correlations.Method: KAFNet retains CPA but handles pre-aligned series efficiently through: (1) Pre-Convolution module for sequence smoothing and sparsity mitigation, (2) Temporal Kernel Aggregation module for learnable compression and modeling intra-series irregularity, and (3) Frequency Linear Attention blocks for low-cost inter-series correlation modeling in the frequency domain.
Result: Experiments on multiple IMTS datasets show KAFNet achieves state-of-the-art forecasting performance with 7.2× parameter reduction and 8.4× training-inference acceleration compared to existing methods.
Conclusion: CPA should be retained for IMTS forecasting when properly handled by the model, as demonstrated by KAFNet’s superior performance and efficiency over graph-based baselines that avoid CPA.
Abstract: Irregular multivariate time series (IMTS), characterized by uneven sampling and inter-variate asynchrony, fuel many forecasting applications yet remain challenging to model efficiently. Canonical Pre-Alignment (CPA) has been widely adopted in IMTS modeling by padding zeros at every global timestamp, thereby alleviating inter-variate asynchrony and unifying the series length, but its dense zero-padding inflates the pre-aligned series length, especially when numerous variates are present, causing prohibitive compute overhead. Recent graph-based models with patching strategies sidestep CPA, but their local message passing struggles to capture global inter-variate correlations. Therefore, we posit that CPA should be retained, with the pre-aligned series properly handled by the model, enabling it to outperform state-of-the-art graph-based baselines that sidestep CPA. Technically, we propose KAFNet, a compact architecture grounded in CPA for IMTS forecasting that couples (1) Pre-Convolution module for sequence smoothing and sparsity mitigation, (2) Temporal Kernel Aggregation module for learnable compression and modeling of intra-series irregularity, and (3) Frequency Linear Attention blocks for the low-cost inter-series correlations modeling in the frequency domain. Experiments on multiple IMTS datasets show that KAFNet achieves state-of-the-art forecasting performance, with a 7.2$\times$ parameter reduction and a 8.4$\times$ training-inference acceleration.
[884] Evaluating Spatio-Temporal Forecasting Trade-offs Between Graph Neural Networks and Foundation Models
Ragini Gupta, Naman Raina, Bo Chen, Li Chen, Claudiu Danilov, Josh Eckhardt, Keyshla Bernard, Klara Nahrstedt
Main category: cs.LG
TL;DR: Systematic benchmarking of forecasting models under varying spatial density and sampling frequency shows STGNNs excel with sparse deployments, TSFMs need dense coverage, and Moirai outperforms all by learning cross-sensor dependencies.
Details
Motivation: Existing IoT data optimization techniques overlook how sampling frequency and spatial coverage variations affect downstream forecasting model performance, creating a gap in understanding the interplay between data collection strategies and model architectures.Method: Systematic study comparing classical models (VAR), neural networks (GRU, Transformer), spatio-temporal GNNs (STGNNs), and time series foundation models (TSFMs: Chronos, Moirai, TimesFM) under varying spatial sensor density and sampling intervals using real-world temperature data from wireless sensor networks.
Result: STGNNs perform best with sparse deployments and moderate sampling rates by leveraging spatial correlations via graph structure. TSFMs work well at high frequencies but degrade with reduced spatial coverage. Moirai (multivariate TSFM) outperforms all models by natively learning cross-sensor dependencies.
Conclusion: The findings provide actionable insights for building efficient forecasting pipelines in spatio-temporal systems, highlighting that model selection should consider deployment density and sampling rates, with Moirai showing superior performance through learned cross-sensor dependencies.
Abstract: Modern IoT deployments for environmental sensing produce high volume spatiotemporal data to support downstream tasks such as forecasting, typically powered by machine learning models. While existing filtering and strategic deployment techniques optimize collected data volume at the edge, they overlook how variations in sampling frequencies and spatial coverage affect downstream model performance. In many forecasting models, incorporating data from additional sensors denoise predictions by providing broader spatial contexts. This interplay between sampling frequency, spatial coverage and different forecasting model architectures remain underexplored. This work presents a systematic study of forecasting models - classical models (VAR), neural networks (GRU, Transformer), spatio-temporal graph neural networks (STGNNs), and time series foundation models (TSFMs: Chronos Moirai, TimesFM) under varying spatial sensor nodes density and sampling intervals using real-world temperature data in a wireless sensor network. Our results show that STGNNs are effective when sensor deployments are sparse and sampling rate is moderate, leveraging spatial correlations via encoded graph structure to compensate for limited coverage. In contrast, TSFMs perform competitively at high frequencies but degrade when spatial coverage from neighboring sensors is reduced. Crucially, the multivariate TSFM Moirai outperforms all models by natively learning cross-sensor dependencies. These findings offer actionable insights for building efficient forecasting pipelines in spatio-temporal systems. All code for model configurations, training, dataset, and logs are open-sourced for reproducibility: https://github.com/UIUC-MONET-Projects/Benchmarking-Spatiotemporal-Forecast-Models
[885] Do Vision-Language Models Leak What They Learn? Adaptive Token-Weighted Model Inversion Attacks
Ngoc-Bao Nguyen, Sy-Tuyen Ho, Koh Jun Hao, Ngai-Man Cheung
Main category: cs.LG
TL;DR: First systematic study of model inversion attacks on vision-language models, showing VLMs are vulnerable to training data leakage with 61.21% attack accuracy.
Details
Motivation: Model inversion attacks pose significant privacy risks by reconstructing private training data from trained models. While prior work focused on unimodal networks, the vulnerability of vision-language models (VLMs) remains unexplored despite their increasing deployment in sensitive domains like healthcare and finance.Method: Introduces token-based and sequence-based model inversion strategies tailored to VLMs’ token-generative nature. Proposes SMI-AW (Sequence-based Model Inversion with Adaptive Token Weighting) that dynamically reweights each token’s loss gradient based on visual grounding to focus optimization on visually informative tokens for better image reconstruction.
Result: VLMs are susceptible to training data leakage. Human evaluation shows 61.21% attack accuracy on reconstructed images. Publicly released VLMs are vulnerable to such attacks. Comprehensive experiments across state-of-the-art VLMs and multiple datasets confirm these privacy risks.
Conclusion: Vision-language models are vulnerable to model inversion attacks that can reconstruct private training data. The study highlights urgent need for privacy safeguards as VLMs become increasingly deployed in sensitive domains.
Abstract: Model inversion (MI) attacks pose significant privacy risks by reconstructing private training data from trained neural networks. While prior studies have primarily examined unimodal deep networks, the vulnerability of vision-language models (VLMs) remains largely unexplored. In this work, we present the first systematic study of MI attacks on VLMs to understand their susceptibility to leaking private visual training data. Our work makes two main contributions. First, tailored to the token-generative nature of VLMs, we introduce a suite of token-based and sequence-based model inversion strategies, providing a comprehensive analysis of VLMs’ vulnerability under different attack formulations. Second, based on the observation that tokens vary in their visual grounding, and hence their gradients differ in informativeness for image reconstruction, we propose Sequence-based Model Inversion with Adaptive Token Weighting (SMI-AW) as a novel MI for VLMs. SMI-AW dynamically reweights each token’s loss gradient according to its visual grounding, enabling the optimization to focus on visually informative tokens and more effectively guide the reconstruction of private images. Through extensive experiments and human evaluations on a range of state-of-the-art VLMs across multiple datasets, we show that VLMs are susceptible to training data leakage. Human evaluation of the reconstructed images yields an attack accuracy of 61.21%, underscoring the severity of these privacy risks. Notably, we demonstrate that publicly released VLMs are vulnerable to such attacks. Our study highlights the urgent need for privacy safeguards as VLMs become increasingly deployed in sensitive domains such as healthcare and finance. Additional experiments are provided in Supp.
[886] Semantic Energy: Detecting LLM Hallucination Beyond Entropy
Huan Ma, Jiadong Pan, Jing Liu, Yan Chen, Joey Tianyi Zhou, Guangyu Wang, Qinghua Hu, Hua Wu, Changqing Zhang, Haifeng Wang
Main category: cs.LG
TL;DR: Semantic Energy is a new uncertainty estimation framework that improves hallucination detection in LLMs by using logits from the penultimate layer and Boltzmann-inspired energy distribution, outperforming semantic entropy methods.
Details
Motivation: LLMs are increasingly deployed in real-world applications but suffer from hallucinations that produce fluent but incorrect responses, leading to erroneous decisions. Existing uncertainty estimation methods like semantic entropy rely on post-softmax probabilities and fail to capture the model's inherent uncertainty, making them ineffective in certain scenarios.Method: The authors introduce Semantic Energy, a novel uncertainty estimation framework that leverages LLMs’ inherent confidence by operating directly on logits of the penultimate layer. The method combines semantic clustering with a Boltzmann-inspired energy distribution to better capture uncertainty where semantic entropy fails.
Result: Experiments across multiple benchmarks show that Semantic Energy significantly improves hallucination detection and uncertainty estimation compared to existing methods, offering more reliable signals for downstream applications.
Conclusion: Semantic Energy provides a more effective approach to uncertainty estimation for LLMs, addressing limitations of semantic entropy and offering improved hallucination detection capabilities for real-world deployment.
Abstract: Large Language Models (LLMs) are being increasingly deployed in real-world applications, but they remain susceptible to hallucinations, which produce fluent yet incorrect responses and lead to erroneous decision-making. Uncertainty estimation is a feasible approach to detect such hallucinations. For example, semantic entropy estimates uncertainty by considering the semantic diversity across multiple sampled responses, thus identifying hallucinations. However, semantic entropy relies on post-softmax probabilities and fails to capture the model’s inherent uncertainty, causing it to be ineffective in certain scenarios. To address this issue, we introduce Semantic Energy, a novel uncertainty estimation framework that leverages the inherent confidence of LLMs by operating directly on logits of penultimate layer. By combining semantic clustering with a Boltzmann-inspired energy distribution, our method better captures uncertainty in cases where semantic entropy fails. Experiments across multiple benchmarks show that Semantic Energy significantly improves hallucination detection and uncertainty estimation, offering more reliable signals for downstream applications such as hallucination detection.
[887] Gaming and Cooperation in Federated Learning: What Can Happen and How to Monitor It
Dongseok Kim, Hyoungsun Choi, Mohamed Jismy Aashik Rasool, Gisung Oh
Main category: cs.LG
TL;DR: The paper develops a strategic framework for federated learning that separates welfare-improving behavior from metric gaming, introduces indices to quantify manipulability and cooperation costs, and provides design principles to reduce gaming while maintaining stable cooperation.
Details
Motivation: Current federated learning formulations treat FL as static optimization problems, ignoring how strategic participants behave under partial observability. The authors recognize that FL deployments are governed strategic systems that require analytical frameworks to distinguish between genuine cooperation and metric gaming.Method: Developed an analytical framework that separates welfare-improving behavior from metric gaming. Introduced indices to quantify manipulability, price of gaming, and price of cooperation. Studied how rules, information disclosure, evaluation metrics, and aggregator-switching policies reshape incentives. Derived threshold conditions for deterring harmful gaming while preserving cooperation. Built a design toolkit including governance checklist and audit-budget allocation algorithm with provable performance guarantee.
Result: Simulations across diverse stylized environments and a federated learning case study consistently matched the qualitative and quantitative patterns predicted by the framework. The results provide threshold conditions for deterring harmful gaming while preserving benign cooperation, and for triggering auto-switch rules when early-warning indicators become critical.
Conclusion: The framework provides design principles and operational guidelines for reducing metric gaming while sustaining stable, high-welfare cooperation in FL platforms. The governance checklist and audit-budget allocation algorithm offer practical tools for implementing these principles in real FL deployments.
Abstract: The success of federated learning (FL) ultimately depends on how strategic participants behave under partial observability, yet most formulations still treat FL as a static optimization problem. We instead view FL deployments as governed strategic systems and develop an analytical framework that separates welfare-improving behavior from metric gaming. Within this framework, we introduce indices that quantify manipulability, the price of gaming, and the price of cooperation, and we use them to study how rules, information disclosure, evaluation metrics, and aggregator-switching policies reshape incentives and cooperation patterns. We derive threshold conditions for deterring harmful gaming while preserving benign cooperation, and for triggering auto-switch rules when early-warning indicators become critical. Building on these results, we construct a design toolkit including a governance checklist and a simple audit-budget allocation algorithm with a provable performance guarantee. Simulations across diverse stylized environments and a federated learning case study consistently match the qualitative and quantitative patterns predicted by our framework. Taken together, our results provide design principles and operational guidelines for reducing metric gaming while sustaining stable, high-welfare cooperation in FL platforms.
[888] PRISM: Diversifying Dataset Distillation by Decoupling Architectural Priors
Brian B. Moser, Shalini Sarode, Federico Raue, Stanislav Frolov, Krzysztof Adamkiewicz, Arundhati Shanbhag, Joachim Folz, Tobias C. Nauen, Andreas Dengel
Main category: cs.LG
TL;DR: PRISM introduces multi-teacher dataset distillation that disentangles architectural priors, using different teacher models for logit-matching and regularization to improve synthetic data diversity and generalization.
Details
Motivation: Existing dataset distillation methods inherit inductive bias from a single teacher model, leading to overly smooth, homogeneous synthetic samples with reduced intra-class diversity and limited generalization as dataset size increases.Method: PRISM decouples logit-matching and regularization objectives, supervising them with different teacher architectures: a primary model for logits and a stochastic subset for batch-normalization alignment. It also introduces teacher selection strategies and a scalable cross-class batch formation scheme for parallel synthesis.
Result: On ImageNet-1K, PRISM consistently outperforms single-teacher methods (SRe2L) and recent multi-teacher variants (G-VBSM) at low- and mid-IPC regimes. Generated data shows significantly richer intra-class diversity with notable drop in cosine similarity between features.
Conclusion: PRISM effectively addresses the architectural bias problem in dataset distillation by disentangling priors from diverse source models, producing more diverse and generalizable synthetic data while introducing scalable synthesis techniques.
Abstract: Dataset distillation (DD) promises compact yet faithful synthetic data, but existing approaches often inherit the inductive bias of a single teacher model. As dataset size increases, this bias drives generation toward overly smooth, homogeneous samples, reducing intra-class diversity and limiting generalization. We present PRISM (PRIors from diverse Source Models), a framework that disentangles architectural priors during synthesis. PRISM decouples the logit-matching and regularization objectives, supervising them with different teacher architectures: a primary model for logits and a stochastic subset for batch-normalization (BN) alignment. On ImageNet-1K, PRISM consistently and reproducibly outperforms single-teacher methods (e.g., SRe2L) and recent multi-teacher variants (e.g., G-VBSM) at low- and mid-IPC regimes. The generated data also show significantly richer intra-class diversity, as reflected by a notable drop in cosine similarity between features. We further analyze teacher selection strategies (pre- vs. intra-distillation) and introduce a scalable cross-class batch formation scheme for fast parallel synthesis. Code will be released after the review period.
[889] Exploring Variational Graph Autoencoders for Distribution Grid Data Generation
Syed Zain Abbas, Ehimare Okoyomon
Main category: cs.LG
TL;DR: VGAEs for synthetic distribution grid generation show promise but have limitations - simple decoders fail, GCN-based ones work on simpler datasets but struggle with complex ones.
Details
Motivation: Address the lack of public power system data for machine learning research in energy networks by generating synthetic distribution grids.Method: Use variational graph autoencoders (VGAEs) with four decoder variants, evaluated on two open-source datasets (ENGAGE and DINGO) using structural and spectral metrics.
Result: Simple decoders fail to capture realistic topologies. GCN-based approaches achieve strong fidelity on simpler ENGAGE dataset but struggle on more complex DINGO dataset, producing disconnected components and repeated motifs.
Conclusion: VGAEs show promise for grid synthesis but have limitations, highlighting need for more expressive generative models and robust evaluation. Models and analysis released as open source to support benchmarking.
Abstract: To address the lack of public power system data for machine learning research in energy networks, we investigate the use of variational graph autoencoders (VGAEs) for synthetic distribution grid generation. Using two open-source datasets, ENGAGE and DINGO, we evaluate four decoder variants and compare generated networks against the original grids using structural and spectral metrics. Results indicate that simple decoders fail to capture realistic topologies, while GCN-based approaches achieve strong fidelity on ENGAGE but struggle on the more complex DINGO dataset, producing artifacts such as disconnected components and repeated motifs. These findings highlight both the promise and limitations of VGAEs for grid synthesis, underscoring the need for more expressive generative models and robust evaluation. We release our models and analysis as open source to support benchmarking and accelerate progress in ML-driven power system research.
[890] FedHK-MVFC: Federated Heat Kernel Multi-View Clustering
Kristina P. Sinaga
Main category: cs.LG
TL;DR: A novel multi-view clustering framework combining quantum field theory with federated learning for healthcare analytics, using heat kernel coefficients to transform Euclidean distances into geometry-aware similarity measures with privacy-preserving protocols.
Details
Motivation: To address the need for privacy-focused medical data analysis in distributed AI settings, enabling collaborative healthcare analytics across hospitals while maintaining data privacy and HIPAA compliance.Method: Proposes heat kernel distance (HKD) transformation with convergence guarantees. Develops two algorithms: HK-MVFC for central analysis and FedHK-MVFC for federated learning with differential privacy and secure aggregation.
Result: Tests on synthetic cardiovascular datasets (10,000 patient records across two hospitals) show increased clustering accuracy, reduced communication overhead, and retained efficiency compared to centralized methods.
Conclusion: The framework establishes a new standard for geometry-aware federated learning in healthcare, translating advanced mathematics into practical, privacy-preserving solutions for sensitive medical data analysis.
Abstract: In the realm of distributed artificial intelligence (AI) and privacy-focused medical applications, this paper proposes a multi-view clustering framework that links quantum field theory with federated healthcare analytics. The method uses heat kernel coefficients from spectral analysis to convert Euclidean distances into geometry-aware similarity measures that capture the structure of diverse medical data. The framework is presented through the heat kernel distance (HKD) transformation, which has convergence guarantees. Two algorithms have been developed: The first, Heat Kernel-Enhanced Multi-View Fuzzy Clustering (HK-MVFC), is used for central analysis. The second, Federated Heat Kernel Multi-View Fuzzy Clustering (FedHK-MVFC), is used for secure, privacy-preserving learning across hospitals. FedHK-MVFC uses differential privacy and secure aggregation to enable HIPAA-compliant collaboration. Tests on synthetic cardiovascular patient datasets demonstrate increased clustering accuracy, reduced communication, and retained efficiency compared to centralized methods. After being validated on 10,000 synthetic patient records across two hospitals, the methods proved useful for collaborative phenotyping involving electrocardiogram (ECG) data, cardiac imaging data, and behavioral data. The proposed methods’ theoretical contributions include update rules with proven convergence, adaptive view weighting, and privacy-preserving protocols. These contributions establish a new standard for geometry-aware federated learning in healthcare, translating advanced mathematics into practical solutions for analyzing sensitive medical data while ensuring rigor and clinical relevance.
[891] Remote Sensing-Oriented World Model
Yuxi Lu, Biao Wu, Zhidong Li, Kunqi Li, Chenya Huang, Huacan Wang, Qizhen Lan, Ronghao Chen, Ling Chen, Bin Liang
Main category: cs.LG
TL;DR: First framework for world modeling in remote sensing using direction-conditioned spatial extrapolation, with new benchmark RSWISE and model RemoteBAGEL that outperforms SOTA baselines.
Details
Motivation: Existing world models are limited to synthetic/constrained environments, while remote sensing urgently needs spatial reasoning for disaster response and urban planning. Need to bridge gap between world modeling capabilities and real-world remote sensing applications.Method: Formulate remote sensing world modeling as direction-conditioned spatial extrapolation (generate adjacent image tiles given central observation + directional instruction). Create RSWISE benchmark with 1,600 tasks across 4 scenarios. Develop RemoteBAGEL, a unified multimodal model fine-tuned on remote sensing data for spatial extrapolation.
Result: RemoteBAGEL consistently outperforms state-of-the-art baselines on the RSWISE benchmark, demonstrating effective spatial reasoning capabilities in remote sensing contexts.
Conclusion: Successfully bridges world modeling with remote sensing, providing first framework for spatial reasoning in this domain with rigorous evaluation methodology using semantic judgment via GPT-4o.
Abstract: World models have shown potential in artificial intelligence by predicting and reasoning about world states beyond direct observations. However, existing approaches are predominantly evaluated in synthetic environments or constrained scene settings, limiting their validation in real-world contexts with broad spatial coverage and complex semantics. Meanwhile, remote sensing applications urgently require spatial reasoning capabilities for disaster response and urban planning. This paper bridges these gaps by introducing the first framework for world modeling in remote sensing. We formulate remote sensing world modeling as direction-conditioned spatial extrapolation, where models generate semantically consistent adjacent image tiles given a central observation and directional instruction. To enable rigorous evaluation, we develop RSWISE (Remote Sensing World-Image Spatial Evaluation), a benchmark containing 1,600 evaluation tasks across four scenarios: general, flood, urban, and rural. RSWISE combines visual fidelity assessment with instruction compliance evaluation using GPT-4o as a semantic judge, ensuring models genuinely perform spatial reasoning rather than simple replication. Afterwards, we present RemoteBAGEL, a unified multimodal model fine-tuned on remote sensing data for spatial extrapolation tasks. Extensive experiments demonstrate that RemoteBAGEL consistently outperforms state-of-the-art baselines on RSWISE.
[892] SCI: A Metacognitive Control for Signal Dynamics
Vishal Joshua Meesala
Main category: cs.LG
TL;DR: SCI is a closed-loop metacognitive control layer that wraps existing stochastic models to make prediction iterative, monitoring confidence and adaptively deciding when to stop, continue sampling, or abstain to regulate interpretive error and expose safety signals.
Details
Motivation: Current deep learning systems operate as open-loop function approximators without regulating computation per case, making them brittle in safety-critical settings where easy and ambiguous inputs receive identical processing and uncertainty is only read from raw probabilities.Method: SCI wraps existing stochastic models (Monte Carlo dropout classifiers) with a lightweight closed-loop metacognitive control layer that monitors interpretive state SP(t) (normalized entropy-based confidence) and adaptively decides whether to stop, continue sampling, or abstain, turning prediction into an iterative process.
Result: SCI allocates more inference steps to misclassified inputs than correct ones (3-4x on MNIST and bearings, 1.4x on MIT-BIH). The resulting ΔSP acts as a usable safety signal for detecting misclassifications with AUROC 0.63 on MNIST, 0.70 on MIT-BIH, and 0.86 on bearings.
Conclusion: SCI provides a practical approach to making deep learning systems more robust in safety-critical applications by regulating interpretive error and exposing safety signals that track when underlying models are likely to fail, without necessarily improving accuracy.
Abstract: Modern deep learning systems are typically deployed as open-loop function approximators: they map inputs to outputs in a single pass, without regulating how much computation or explanatory effort is spent on a given case. In safety-critical settings, this is brittle: easy and ambiguous inputs receive identical processing, and uncertainty is only read off retrospectively from raw probabilities. We introduce the Surgical Cognitive Interpreter (SCI), a lightweight closed-loop metacognitive control layer that wraps an existing stochastic model and turns prediction into an iterative process. SCI monitors a scalar interpretive state SP(t), here instantiated as a normalized entropy-based confidence signal, and adaptively decides whether to stop, continue sampling, or abstain. The goal is not to improve accuracy per se, but to regulate interpretive error ΔSP and expose a safety signal that tracks when the underlying model is likely to fail. We instantiate SCI around Monte Carlo dropout classifiers in three domains: vision (MNIST digits), medical time series (MIT-BIH arrhythmia), and industrial condition monitoring (rolling-element bearings). In all cases, the controller allocates more inference steps to misclassified inputs than to correct ones (up to about 3-4x on MNIST and bearings, and 1.4x on MIT-BIH). The resulting ΔSP acts as a usable safety signal for detecting misclassifications (AUROC 0.63 on MNIST, 0.70 on MIT-BIH, 0.86 on bearings). Code and reproducibility: https://github.com/vishal-1344/sci
[893] Decoupled-Value Attention for Prior-Data Fitted Networks: GP Inference for Physical Equations
Kaustubh Sharma, Simardeep Singh, Parikshit Pareek
Main category: cs.LG
TL;DR: PFNs with Decoupled-Value Attention (DVA) achieve fast GP-like inference by computing similarities from inputs only and propagating labels through values, enabling efficient high-dimensional regression with 80x speedup over exact GP.
Details
Motivation: Prior-data fitted networks (PFNs) offer fast alternatives to Gaussian process inference but struggle with high-dimensional regression using standard Transformer attention. The authors aim to improve PFN effectiveness for high-dimensional tasks while maintaining computational efficiency.Method: Introduces Decoupled-Value Attention (DVA) that mirrors GP properties: computes similarities from inputs only (like kernel) and propagates labels solely through values (like predictive mean). DVA remains kernel-free while capturing GP-like behavior. Also shows PFNs are backbone architecture invariant.
Result: DVA reduces validation loss by >50% in 5D and 10D cases; attention rule is more decisive than backbone architecture (CNN/RNN/LSTM perform at par with Transformers); achieves 64D power flow approximations with MAE ~E-03 while being 80x faster than exact GP inference.
Conclusion: PFNs with DVA provide efficient, scalable alternatives to GP inference for high-dimensional regression, with attention mechanism being the crucial factor rather than backbone architecture, enabling fast and accurate surrogate modeling.
Abstract: Prior-data fitted networks (PFNs) are a promising alternative to time-consuming Gaussian process (GP) inference for creating fast surrogates of physical systems. PFN reduces the computational burden of GP-training by replacing Bayesian inference in GP with a single forward pass of a learned prediction model. However, with standard Transformer attention, PFNs show limited effectiveness on high-dimensional regression tasks. We introduce Decoupled-Value Attention (DVA)– motivated by the GP property that the function space is fully characterized by the kernel over inputs and the predictive mean is a weighted sum of training targets. DVA computes similarities from inputs only and propagates labels solely through values. Thus, the proposed DVA mirrors the GP update while remaining kernel-free. We demonstrate that PFNs are backbone architecture invariant and the crucial factor for scaling PFNs is the attention rule rather than the architecture itself. Specifically, our results demonstrate that (a) localized attention consistently reduces out-of-sample validation loss in PFNs across different dimensional settings, with validation loss reduced by more than 50% in five- and ten-dimensional cases, and (b) the role of attention is more decisive than the choice of backbone architecture, showing that CNN, RNN and LSTM-based PFNs can perform at par with their Transformer-based counterparts. The proposed PFNs provide 64-dimensional power flow equation approximations with a mean absolute error of the order of E-03, while being over 80x faster than exact GP inference.
[894] TRiCo: Triadic Game-Theoretic Co-Training for Robust Semi-Supervised Learning
Hongyang He, Xinyuan Song, Yangfan He, Zeyu Zhang, Yanshu Li, Haochen You, Lifan Sun, Wenqiao Zhang
Main category: cs.LG
TL;DR: TRiCo is a triadic game-theoretic co-training framework for semi-supervised learning that introduces structured interactions among two students, a meta-learned teacher, and an adversarial generator, formalized as a Stackelberg game.
Details
Motivation: To address key limitations in existing SSL frameworks: static view interactions, unreliable pseudo-labels, and lack of hard sample modeling. The paper aims to provide a more principled and generalizable solution for semi-supervised learning.Method: TRiCo formulates SSL as a triadic interaction: (1) two student classifiers trained on frozen complementary representations, (2) a meta-learned teacher that adaptively regulates pseudo-label selection and loss balancing via validation feedback, and (3) a non-parametric generator that perturbs embeddings to uncover decision boundary weaknesses. Pseudo-labels are selected based on mutual information rather than confidence. The framework is formalized as a Stackelberg game where the teacher leads strategy optimization.
Result: Extensive experiments on CIFAR-10, SVHN, STL-10, and ImageNet demonstrate that TRiCo consistently achieves state-of-the-art performance in low-label regimes while remaining architecture-agnostic and compatible with frozen vision backbones.
Conclusion: TRiCo provides a principled and generalizable solution for semi-supervised learning by rethinking SSL structure through triadic game-theoretic co-training, addressing key limitations of existing approaches and achieving superior performance across multiple benchmarks.
Abstract: We introduce TRiCo, a novel triadic game-theoretic co-training framework that rethinks the structure of semi-supervised learning by incorporating a teacher, two students, and an adversarial generator into a unified training paradigm. Unlike existing co-training or teacher-student approaches, TRiCo formulates SSL as a structured interaction among three roles: (i) two student classifiers trained on frozen, complementary representations, (ii) a meta-learned teacher that adaptively regulates pseudo-label selection and loss balancing via validation-based feedback, and (iii) a non-parametric generator that perturbs embeddings to uncover decision boundary weaknesses. Pseudo-labels are selected based on mutual information rather than confidence, providing a more robust measure of epistemic uncertainty. This triadic interaction is formalized as a Stackelberg game, where the teacher leads strategy optimization and students follow under adversarial perturbations. By addressing key limitations in existing SSL frameworks, such as static view interactions, unreliable pseudo-labels, and lack of hard sample modeling, TRiCo provides a principled and generalizable solution. Extensive experiments on CIFAR-10, SVHN, STL-10, and ImageNet demonstrate that TRiCo consistently achieves state-of-the-art performance in low-label regimes, while remaining architecture-agnostic and compatible with frozen vision backbones.Code:https://github.com/HoHongYeung/NeurIPS25-TRiCo.
[895] AEGIS: Authentic Edge Growth In Sparsity for Link Prediction in Edge-Sparse Bipartite Knowledge Graphs
Hugh Xuechen Liu, Kıvanç Tatar
Main category: cs.LG
TL;DR: AEGIS framework for edge augmentation in sparse bipartite knowledge graphs improves link prediction by resampling existing edges rather than creating fabricated ones, with semantic KNN augmentation showing best results when node descriptions are available.
Details
Motivation: Bipartite knowledge graphs in niche domains are typically data-poor and edge-sparse, which hinders link prediction performance. Existing augmentation methods often create fabricated endpoints that may not preserve graph authenticity.Method: AEGIS (Authentic Edge Growth In Sparsity) is an edge-only augmentation framework that resamples existing training edges either uniformly simple or with inverse-degree bias (degree-aware), preserving the original node set and avoiding fabricated endpoints. The framework is tested on naturally sparse graphs (game design pattern network) and induced-sparsity benchmarks (Amazon, MovieLens) using high-rate bond percolation.
Result: On Amazon and MovieLens, copy-based AEGIS variants match baseline performance while semantic KNN augmentation is the only method that restores AUC and calibration. On the text-rich GDP graph, semantic KNN achieves the largest AUC improvement and Brier score reduction, and simple augmentation also lowers the Brier score relative to sparse control.
Conclusion: Authenticity-constrained resampling is a data-efficient strategy for sparse bipartite link prediction, with semantic augmentation providing additional boost when informative node descriptions are available. Random and synthetic edges remain detrimental to performance.
Abstract: Bipartite knowledge graphs in niche domains are typically data-poor and edge-sparse, which hinders link prediction. We introduce AEGIS (Authentic Edge Growth In Sparsity), an edge-only augmentation framework that resamples existing training edges -either uniformly simple or with inverse-degree bias degree-aware -thereby preserving the original node set and sidestepping fabricated endpoints. To probe authenticity across regimes, we consider naturally sparse graphs (game design pattern’s game-pattern network) and induce sparsity in denser benchmarks (Amazon, MovieLens) via high-rate bond percolation. We evaluate augmentations on two complementary metrics: AUC-ROC (higher is better) and the Brier score (lower is better), using two-tailed paired t-tests against sparse baselines. On Amazon and MovieLens, copy-based AEGIS variants match the baseline while the semantic KNN augmentation is the only method that restores AUC and calibration; random and synthetic edges remain detrimental. On the text-rich GDP graph, semantic KNN achieves the largest AUC improvement and Brier score reduction, and simple also lowers the Brier score relative to the sparse control. These findings position authenticity-constrained resampling as a data-efficient strategy for sparse bipartite link prediction, with semantic augmentation providing an additional boost when informative node descriptions are available.
[896] Countering adversarial evasion in regression analysis
David Benfield, Phan Tu Vuong, Alain Zemkoho
Main category: cs.LG
TL;DR: This paper proposes a pessimistic bilevel optimization framework for adversarial regression scenarios, extending previous game-theoretic approaches that were limited to classification problems.
Details
Motivation: Adversarial evasion attacks challenge the assumption of consistent data distributions between training and deployment. While game-theoretic models and recent pessimistic bilevel optimization approaches have been effective for classification problems (like spam filtering and malware detection), these methods haven't been adapted to regression scenarios.Method: The authors propose a pessimistic bilevel optimization program specifically designed for regression scenarios. This approach removes assumptions about the convexity and uniqueness of the adversary’s optimal strategy, capturing the antagonistic nature of adversarial attacks in regression settings.
Result: The paper introduces a novel formulation that extends adversarial robustness techniques from classification to regression problems, providing a framework for training resilient regression models against adaptive adversaries.
Conclusion: This work bridges an important gap in adversarial machine learning by adapting pessimistic bilevel optimization to regression scenarios, enabling more robust regression models against evasion attacks without restrictive assumptions about adversary behavior.
Abstract: Adversarial machine learning challenges the assumption that the underlying distribution remains consistent throughout the training and implementation of a prediction model. In particular, adversarial evasion considers scenarios where adversaries adapt their data to influence particular outcomes from established prediction models, such scenarios arise in applications such as spam email filtering, malware detection and fake-image generation, where security methods must be actively updated to keep up with the ever-improving generation of malicious data. Game theoretic models have been shown to be effective at modelling these scenarios and hence training resilient predictors against such adversaries. Recent advancements in the use of pessimistic bilevel optimsiation which remove assumptions about the convexity and uniqueness of the adversary’s optimal strategy have proved to be particularly effective at mitigating threats to classifiers due to its ability to capture the antagonistic nature of the adversary. However, this formulation has not yet been adapted to regression scenarios. This article serves to propose a pessimistic bilevel optimisation program for regression scenarios which makes no assumptions on the convexity or uniqueness of the adversary’s solutions.
[897] ProtoTS: Learning Hierarchical Prototypes for Explainable Time Series Forecasting
Ziheng Peng, Shijie Ren, Xinyue Gu, Linxiao Yang, Xiting Wang, Liang Sun
Main category: cs.LG
TL;DR: ProtoTS is an interpretable time series forecasting framework that uses hierarchical prototypes to capture temporal patterns, achieving both high accuracy and transparent decision-making through expert-steerable interpretations.
Details
Motivation: Deep learning models for time series forecasting lack transparency in their decision-making process, which is crucial for trust in high-stakes scenarios. Existing interpretable models only provide local and partial explanations, failing to show how heterogeneous input variables jointly shape overall temporal patterns in forecasts.Method: ProtoTS computes instance-prototype similarity using denoised representations that preserve heterogeneous information. It organizes prototypes hierarchically to capture both global temporal patterns (coarse prototypes) and finer-grained local variations (detailed prototypes), enabling expert steering and multi-level interpretability.
Result: Experiments on multiple realistic benchmarks, including a newly released LOF dataset, show that ProtoTS exceeds existing methods in forecast accuracy while delivering expert-steerable interpretations for better model understanding and decision support.
Conclusion: ProtoTS successfully addresses the need for both accurate and interpretable time series forecasting by modeling prototypical temporal patterns hierarchically, providing transparent decision-making that can be steered by domain experts.
Abstract: While deep learning has achieved impressive performance in time series forecasting, it becomes increasingly crucial to understand its decision-making process for building trust in high-stakes scenarios. Existing interpretable models often provide only local and partial explanations, lacking the capability to reveal how heterogeneous and interacting input variables jointly shape the overall temporal patterns in the forecast curve. We propose ProtoTS, a novel interpretable forecasting framework that achieves both high accuracy and transparent decision-making through modeling prototypical temporal patterns. ProtoTS computes instance-prototype similarity based on a denoised representation that preserves abundant heterogeneous information. The prototypes are organized hierarchically to capture global temporal patterns with coarse prototypes while capturing finer-grained local variations with detailed prototypes, enabling expert steering and multi-level interpretability. Experiments on multiple realistic benchmarks, including a newly released LOF dataset, show that ProtoTS not only exceeds existing methods in forecast accuracy but also delivers expert-steerable interpretations for better model understanding and decision support.
[898] Adaptive Canonicalization with Application to Invariant Anisotropic Geometric Networks
Ya-Wei Eileen Lin, Ron Levie
Main category: cs.LG
TL;DR: Adaptive canonicalization framework for equivariant ML that makes canonicalization input- and network-dependent via prior maximization, yielding continuous symmetry-respecting models with universal approximation properties.
Details
Motivation: Standard canonicalization in equivariant ML introduces discontinuities that affect training stability, limit generalization, and complicate universal approximation theorems.Method: Adaptive canonicalization based on prior maximization, where the standard form of input is chosen to maximize the network’s predictive confidence. This makes canonicalization both input- and network-dependent.
Result: The framework yields continuous and symmetry-respecting models with universal approximation properties. Applied to spectral graph neural networks (eigenbasis ambiguities) and point clouds (rotational symmetries), it outperforms data augmentation, standard canonicalization, and equivariant architectures on molecular/protein classification and point cloud tasks.
Conclusion: Adaptive canonicalization provides a superior solution to equivariant ML problems by addressing the limitations of traditional canonicalization while maintaining symmetry properties and theoretical guarantees.
Abstract: Canonicalization is a widely used strategy in equivariant machine learning, enforcing symmetry in neural networks by mapping each input to a standard form. Yet, it often introduces discontinuities that can affect stability during training, limit generalization, and complicate universal approximation theorems. In this paper, we address this by introducing adaptive canonicalization, a general framework in which the canonicalization depends both on the input and the network. Specifically, we present the adaptive canonicalization based on prior maximization, where the standard form of the input is chosen to maximize the predictive confidence of the network. We prove that this construction yields continuous and symmetry-respecting models that admit universal approximation properties. We propose two applications of our setting: (i) resolving eigenbasis ambiguities in spectral graph neural networks, and (ii) handling rotational symmetries in point clouds. We empirically validate our methods on molecular and protein classification, as well as point cloud classification tasks. Our adaptive canonicalization outperforms the three other common solutions to equivariant machine learning: data augmentation, standard canonicalization, and equivariant architectures.
[899] Periodic Asynchrony: An Effective Method for Accelerating Reinforcement Learning
Jian Lu
Main category: cs.LG
TL;DR: The paper proposes a periodically asynchronous RL framework that separates inference and training deployment, achieving 3x performance improvement on NPU platforms while maintaining algorithm accuracy equivalent to synchronous methods.
Details
Motivation: Current RL frameworks deploy inference and training on same devices, creating computational coupling that prevents concurrent execution and limits training efficiency. The synchronous approach restricts independent scaling of components.Method: Separates inference and training deployment with improved data loader, transforming synchronous architecture into periodically asynchronous framework. Uses unified tri-model architecture in training phase and shared-prompt attention mask to reduce repetitive computation.
Result: Achieves at least threefold overall performance improvement in RL training on NPU platforms while maintaining algorithm accuracy completely equivalent to synchronous methods.
Conclusion: The periodically asynchronous framework enables demand-driven, independent, and elastic scaling of inference and training components, showing potential for widespread application in RL training.
Abstract: Since the introduction of the GRPO algorithm, reinforcement learning (RL) has attracted increasing attention, with growing efforts to reproduce and apply it. However, training efficiency remains a critical challenge. In mainstream RL frameworks, inference and training are typically deployed on the same devices. While this approach reduces costs through resource consolidation, its synchronous execution imposes a computational coupling that prevents concurrent inference and training. In this study, we are returning to the strategy of separating inference and training deployment, and by introducing improvements in the data loader, we transform the conventional synchronous architecture into a periodically asynchronous framework, which allows for demand-driven, independent, and elastic scaling of each component, while the accuracy of the algorithm remains completely equivalent to the synchronization method, with both belonging to the on-policy strategy. It is worth emphasizing that we apply a unified tri-model architecture in the training phase, and we also proposed a shared-prompt attention mask to reduce repetitive computation. In practice, these works have achieved at least a threefold overall performance improvement in RL training on NPU platforms, indicating its potential for widespread application.
[900] EXP-CAM: Explanation Generation and Circuit Discovery Using Classifier Activation Matching
Pirzada Suhail, Aditya Anand, Amit Sethi
Main category: cs.LG
TL;DR: EXP-CAM generates minimal, faithful explanations for image classifiers using Classifier Activation Matching and binary masks, enabling both input-level explanations and mechanistic circuit discovery.
Details
Motivation: Machine learning models learn many decision rules that are hard to identify and interpret in high-dimensional spaces. Current explanations lack conciseness and human-readability while preserving model decisions.Method: Train lightweight auto-encoder to produce binary masks highlighting decision-critical image regions. Training integrates: activation alignment across layers, output label consistency, sparsity/compactness priors, and robustness constraints for faithfulness.
Result: Generates minimal and faithful explanations that preserve model decisions while being concise and human-readable. Also enables mechanistic interpretation through circuit readout procedure identifying active channels and constructing channel-level graphs.
Conclusion: EXP-CAM bridges minimal input-level explanations with mechanistic understanding of internal computations, providing practical tools for both explanation generation and circuit discovery in image classifiers.
Abstract: Machine learning models, by virtue of training, learn a large repertoire of decision rules for any given input, and any one of these may suffice to justify a prediction. However, in high-dimensional input spaces, such rules are difficult to identify and interpret. In this paper, we introduce EXP-CAM: an explanation generation and circuit discovery approach using Classifier Activation Matching. EXP-CAM can generate minimal and faithful explanations for the decisions of pre-trained image classifiers that not only preserve the model’s decision but are also concise and human-readable. We aim to identify minimal explanations that not only preserve the model’s decision but are also concise and human-readable. To achieve this, we train a lightweight auto-encoder to produce binary masks that learns to highlight the decision-wise critical regions of an image while discarding irrelevant background. The training objective integrates activation alignment across multiple layers, consistency at the output label, priors that encourage sparsity, and compactness, along with a robustness constraint that enforces faithfulness. The minimal explanations so generated also lead us to mechanistically interpreting the model internals. In this regard we also introduce a circuit readout procedure wherein using the explanation’s forward pass and gradients, we identify active channels and construct a channel-level graph, scoring inter-layer edges by ingress weight magnitude times source activation and feature-to-class links by classifier weight magnitude times feature activation. Together, these contributions provide a practical bridge between minimal input-level explanations and a mechanistic understanding of the internal computations driving model decisions.
[901] Less is More: Towards Simple Graph Contrastive Learning
Yanan Zhao, Feng Ji, Jingyang Dai, Jiaze Ma, Wee Peng Tay
Main category: cs.LG
TL;DR: Simple GCL model without data augmentation or negative sampling achieves SOTA on heterophilic graphs by using GCN for structural features and MLP for node feature noise.
Details
Motivation: Current Graph Contrastive Learning methods struggle with heterophilic graphs (where connected nodes belong to different classes) and rely on complex augmentation schemes, intricate encoders, or negative sampling. The authors question whether such complexity is necessary and seek a simpler approach.Method: Propose a simple GCL model using two complementary views: (1) GCN encoder to capture structural features from graph topology, and (2) MLP encoder to isolate node feature noise. The approach requires neither data augmentation nor negative sampling.
Result: Achieves state-of-the-art results on heterophilic benchmarks with minimal computational and memory overhead. Also shows advantages in homophilic graphs regarding complexity, scalability, and robustness. Validated through extensive experiments including robustness evaluations against adversarial attacks.
Conclusion: A simple GCL approach that leverages complementary views from original node features and graph structure can effectively handle heterophilic graphs without complex augmentations or negative sampling, offering theoretical justification and practical advantages.
Abstract: Graph Contrastive Learning (GCL) has shown strong promise for unsupervised graph representation learning, yet its effectiveness on heterophilic graphs, where connected nodes often belong to different classes, remains limited. Most existing methods rely on complex augmentation schemes, intricate encoders, or negative sampling, which raises the question of whether such complexity is truly necessary in this challenging setting. In this work, we revisit the foundations of supervised and unsupervised learning on graphs and uncover a simple yet effective principle for GCL: mitigating node feature noise by aggregating it with structural features derived from the graph topology. This observation suggests that the original node features and the graph structure naturally provide two complementary views for contrastive learning. Building on this insight, we propose an embarrassingly simple GCL model that uses a GCN encoder to capture structural features and an MLP encoder to isolate node feature noise. Our design requires neither data augmentation nor negative sampling, yet achieves state-of-the-art results on heterophilic benchmarks with minimal computational and memory overhead, while also offering advantages in homophilic graphs in terms of complexity, scalability, and robustness. We provide theoretical justification for our approach and validate its effectiveness through extensive experiments, including robustness evaluations against both black-box and white-box adversarial attacks.
[902] Auditing Algorithmic Bias in Transformer-Based Trading
Armin Gerami, Ramani Duraiswami
Main category: cs.LG
TL;DR: Transformer models in finance show biases: they ignore data volatility and prefer lower-frequency price movements, raising concerns about their reliability for financial decision-making.
Details
Motivation: Despite the growing popularity of transformer models in financial applications, their potential risks and biases remain insufficiently explored. The authors aim to audit how these models rely on volatile data for decision-making and quantify how price movement frequency affects prediction confidence.Method: The researchers employ a transformer model for financial prediction and introduce a novel metric based on Partial Information Decomposition (PID) to measure the influence of each asset on the model’s decision-making process.
Result: The analysis reveals two significant findings: 1) The transformer model completely disregards data volatility in its decision-making, and 2) The model exhibits a bias toward data with lower-frequency price movements.
Conclusion: Transformer models in financial applications demonstrate concerning biases - they ignore volatility and favor lower-frequency data, which could lead to unreliable predictions and highlight the need for better auditing of AI models in finance.
Abstract: Transformer models have become increasingly popular in financial applications, yet their potential risk making and biases remain under-explored. The purpose of this work is to audit the reliance of the model on volatile data for decision-making, and quantify how the frequency of price movements affects the model’s prediction confidence. We employ a transformer model for prediction, and introduce a metric based on Partial Information Decomposition (PID) to measure the influence of each asset on the model’s decision making. Our analysis reveals two key observations: first, the model disregards data volatility entirely, and second, it is biased toward data with lower-frequency price movements.
[903] Get RICH or Die Scaling: Profitably Trading Inference Compute for Robustness
Tavish McDonald, Bo Lei, Stanislav Fort, Bhavya Kailkhura, Brian Bartoldson
Main category: cs.LG
TL;DR: Inference compute (test-time reasoning) improves robustness to adversarial OOD attacks when models can compositionally generalize from ID components, creating a rich-get-richer dynamic where robust models benefit more from test compute.
Details
Motivation: Previous work showed LLM reasoning improves robustness to jailbreaks, but this benefit fades with gradient-based or multimodal attacks. The paper addresses this gap by investigating when and why inference compute helps against adversarial OOD data.Method: Proposes the Robustness from Inference Compute Hypothesis (RICH): inference compute defenses work better when training data reflects attacked data’s components, enabling compositional generalization. Empirically tests across vision language models and attack types, examining how scaling test compute interacts with base model robustness.
Result: Robustness gains from test-time compute occur when compositional generalization enables specification following on OOD data. For example, InternVL 3.5 gains little robustness from test compute scaling alone, but significant robustness when first robustifying its vision encoder, showing base model robustness correlates with inference compute benefits.
Conclusion: Inference compute offers robustness benefits even against gradient-based and multimodal attacks when models can compositionally generalize. The RICH dynamic creates a rich-get-richer effect where robust models benefit more from test compute. Layering train-time and test-time defenses provides synergistic benefits.
Abstract: Models are susceptible to adversarially out-of-distribution (OOD) data despite large training-compute investments into their robustification. Zaremba et al. (2025) make progress on this problem at test time, showing LLM reasoning improves satisfaction of model specifications designed to thwart attacks, resulting in a correlation between reasoning effort and robustness to jailbreaks. However, this benefit of test compute fades when attackers are given access to gradients or multimodal inputs. We address this gap, clarifying that inference-compute offers benefits even in such cases. Our approach argues that compositional generalization, through which OOD data is understandable via its in-distribution (ID) components, enables adherence to defensive specifications on adversarially OOD inputs. Namely, we posit the Robustness from Inference Compute Hypothesis (RICH): inference-compute defenses profit as the model’s training data better reflects the attacked data’s components. We empirically support this hypothesis across vision language model and attack types, finding robustness gains from test-time compute if specification following on OOD data is unlocked by compositional generalization. For example, InternVL 3.5 gpt-oss 20B gains little robustness when its test compute is scaled, but such scaling adds significant robustness if we first robustify its vision encoder. This correlation of inference-compute’s robustness benefit with base model robustness is the rich-get-richer dynamic of the RICH: attacked data components are more ID for robustified models, aiding compositional generalization to OOD data. Thus, we advise layering train-time and test-time defenses to obtain their synergistic benefit.
[904] Arbitrary Entropy Policy Optimization Breaks The Exploration Bottleneck of Reinforcement Learning
Chen Wang, Zhaochun Li, Jionghao Bai, Yuzhi Zhang, Shisheng Cui, Zhou Zhao, Yue Wang
Main category: cs.LG
TL;DR: AEPO is a new RL method that solves entropy collapse in LLM reasoning by reformulating entropy regularization as a policy-gradient problem, enabling precise entropy control without optimization bias.
Details
Motivation: Current RL methods like GRPO suffer from entropy collapse, causing premature policy convergence and vanishing exploration, which limits LLM reasoning capabilities. Existing entropy-regularized methods create a trade-off between reward and entropy, introducing optimization bias.Method: AEPO reformulates entropy regularization as a policy-gradient optimization problem. Instead of directly manipulating entropy, it implicitly regulates entropy by applying a REINFORCE regularization term on temperature-adjusted samples, ensuring entropy is controlled but never dominates optimization.
Result: AEPO outperforms RL baselines on both pass@1 and pass@k metrics, and even surpasses the base model on pass@1024. It achieves more effective optimization dynamics and provides empirical evidence linking entropy, exploration, and performance.
Conclusion: AEPO enables arbitrary and principled entropy regulation in RL for LLMs, solving the entropy collapse problem and demonstrating that proper entropy control can expand the reasoning frontier of language models.
Abstract: Reinforcement Learning (RL) is essential for enhancing the reasoning capabilities of large language models (LLMs), yet the widely adopted Group Relative Policy Optimization (GRPO) suffers from entropy collapse, causing exploration to vanish and policies to converge prematurely. As a result, RL is widely believed to be incapable of expanding the reasoning frontier of LLMs. Existing entropy-regularized methods introduce an inevitable trade-off between reward and entropy, leading to exploration accompanied by non-negligible optimization bias. In this work, we prove that temperature-guided REINFORCE can modulate policy entropy, and propose Arbitrary Entropy Policy Optimization (AEPO), which reformulates entropy regularization as a policy-gradient optimization problem. Rather than manipulating entropy directly, AEPO implicitly regulates it by applying a REINFORCE regularization term on temperature-adjusted samples, ensuring that entropy is controlled but never dominates optimization, thereby enabling arbitrary and principled entropy regulation. Experiments show that AEPO outperforms RL baselines on both pass@1 and pass@$k$, and even surpasses the base model on pass@1024. By modulating entropy precisely, AEPO achieves more effective optimization dynamics and provides direct empirical evidence that entropy, exploration, and performance are intrinsically linked.
[905] Multi-View Graph Learning with Graph-Tuple
Shiyu Chen, Ningyuan Huang, Soledad Villar
Main category: cs.LG
TL;DR: Multi-view graph-tuple framework partitions graphs into disjoint subgraphs to capture multiple interaction scales, using heterogeneous message-passing for better expressivity than single-graph GNNs.
Details
Motivation: Traditional GNNs scale with graph edges, making them inefficient on dense graphs like point clouds. Existing sparsification methods force arbitrary single-scale choices and discard multi-scale information.Method: Introduces multi-view graph-tuple framework that partitions graphs into disjoint subgraphs capturing different interaction scales. Uses heterogeneous message-passing architecture inspired by non-commuting operators theory.
Result: Formally proves the framework is more expressive with lower oracle risk than single-graph models. Demonstrates better performance on molecular property prediction from Coulomb matrices and cosmological parameter inference from point clouds.
Conclusion: Multi-view graph-tuple approach effectively captures multi-scale interactions, outperforming single-graph baselines and offering versatile solution for dense graph applications.
Abstract: Graph Neural Networks (GNNs) typically scale with the number of graph edges, making them well suited for sparse graphs but less efficient on dense graphs, such as point clouds or molecular interactions. A common remedy is to sparsify the graph via similarity thresholding or distance pruning, but this forces an arbitrary choice of a single interaction scale and discards crucial information from other scales. To overcome this limitation, we introduce a multi-view graph-tuple framework. Instead of a single graph, our graph-tuple framework partitions the graph into disjoint subgraphs, capturing primary local interactions and weaker, long-range connections. We then learn multi-view representations from the graph-tuple via a heterogeneous message-passing architecture inspired by the theory of non-commuting operators, which we formally prove is strictly more expressive and guarantees a lower oracle risk compared to single-graph message-passing models. We instantiate our framework on two scientific domains: molecular property prediction from feature-scarce Coulomb matrices and cosmological parameter inference from geometric point clouds. On both applications, our multi-view graph-tuple models demonstrate better performance than single-graph baselines, highlighting the power and versatility of our multi-view approach.
[906] Enhancing Time Series Forecasting through Selective Representation Spaces: A Patch Perspective
Xingjian Wu, Xiangfei Qiu, Hanyin Cheng, Zhengyu Li, Jilin Hu, Chenjuan Guo, Bin Yang
Main category: cs.LG
TL;DR: SRSNet introduces Selective Representation Space with learnable patching and dynamic reassembly to flexibly select informative patches for better time series forecasting.
Details
Motivation: Conventional patching creates fixed representation spaces with adjacent patches, leading to insufficiently expressive representations. The paper aims to create a selective representation space that can flexibly include the most informative patches.Method: Proposes Selective Representation Space (SRS) module with learnable Selective Patching and Dynamic Reassembly techniques to adaptively select and shuffle patches from contextual time series. SRSNet combines SRS with an MLP head.
Result: SRSNet achieves state-of-the-art performance on real-world datasets from multiple domains. The SRS module also enhances existing patch-based models as a plug-and-play component.
Conclusion: The selective representation space approach effectively improves time series forecasting by flexibly selecting informative patches, offering both a standalone solution (SRSNet) and an enhancement for existing models.
Abstract: Time Series Forecasting has made significant progress with the help of Patching technique, which partitions time series into multiple patches to effectively retain contextual semantic information into a representation space beneficial for modeling long-term dependencies. However, conventional patching partitions a time series into adjacent patches, which causes a fixed representation space, thus resulting in insufficiently expressful representations. In this paper, we pioneer the exploration of constructing a selective representation space to flexibly include the most informative patches for forecasting. Specifically, we propose the Selective Representation Space (SRS) module, which utilizes the learnable Selective Patching and Dynamic Reassembly techniques to adaptively select and shuffle the patches from the contextual time series, aiming at fully exploiting the information of contextual time series to enhance the forecasting performance of patch-based models. To demonstrate the effectiveness of SRS module, we propose a simple yet effective SRSNet consisting of SRS and an MLP head, which achieves state-of-the-art performance on real-world datasets from multiple domains. Furthermore, as a novel plug-and-play module, SRS can also enhance the performance of existing patch-based models. The resources are available at https://github.com/decisionintelligence/SRSNet.
[907] From Noise to Laws: Regularized Time-Series Forecasting via Denoised Dynamic Graphs
Hongwei Ma, Junbin Gao, Minh-ngoc Tran
Main category: cs.LG
TL;DR: PRISM: A diffusion-based model for long-horizon multivariate time-series forecasting that combines score-based diffusion, dynamic graph encoding, and physics regularization to achieve state-of-the-art performance.
Details
Motivation: Long-horizon multivariate time-series forecasting faces three key challenges: (i) denoising heterogeneous signals, (ii) tracking time-varying cross-series dependencies, and (iii) maintaining stability and physical plausibility over long rollout horizons.Method: PRISM couples a score-based diffusion preconditioner with a dynamic, correlation-thresholded graph encoder and a forecast head regularized by generic physics penalties. The model includes theoretical guarantees with contraction of induced horizon dynamics under mild conditions and Lipschitz bounds for graph blocks.
Result: On six standard benchmarks, PRISM achieves consistent state-of-the-art (SOTA) performance with strong MSE and MAE gains.
Conclusion: PRISM effectively addresses the challenges of long-horizon multivariate forecasting through its diffusion-based architecture with dynamic graph encoding and physics regularization, demonstrating both theoretical robustness and empirical superiority.
Abstract: Long-horizon multivariate time-series forecasting is challenging because realistic predictions must (i) denoise heterogeneous signals, (ii) track time-varying cross-series dependencies, and (iii) remain stable and physically plausible over long rollout horizons. We present PRISM, which couples a score-based diffusion preconditioner with a dynamic, correlation-thresholded graph encoder and a forecast head regularized by generic physics penalties. We prove contraction of the induced horizon dynamics under mild conditions and derive Lipschitz bounds for graph blocks, explaining the model’s robustness. On six standard benchmarks , PRISM achieves consistent SOTA with strong MSE and MAE gains.
[908] VeFA: Vector-Based Feature Space Adaptation for Robust Model Fine-Tuning
Peng Wang, Minghao Gu, Qiang Huang
Main category: cs.LG
TL;DR: VeFA is a feature-space fine-tuning method that prevents catastrophic forgetting by avoiding intruder dimensions and preserving pre-trained knowledge through element-wise feature adaptation.
Details
Motivation: Current parameter-efficient fine-tuning methods operate in weight space and can cause catastrophic forgetting by introducing intruder dimensions that overwrite pre-trained knowledge, especially when downstream data is limited or differs from pre-training distribution.Method: Vector-based Feature Adaptation (VeFA) operates directly in feature space with element-wise adaptation on individual features, ensuring fine-tuned weights stay within the column space of pre-trained weight matrix. Inspired by effect equivalence modeling, it compensates for downstream lurking variables via lightweight feature-level transformations.
Result: VeFA achieves comparable fine-tuning performance to LoRA while consistently exhibiting stronger robustness across image classification, NLU, and NLG benchmarks, preserving pre-trained representations and improving generalization under distribution shift.
Conclusion: Feature-space adaptation via VeFA effectively mitigates catastrophic forgetting by avoiding intruder dimensions, preserving pre-trained knowledge, and enhancing model robustness while maintaining competitive fine-tuning performance.
Abstract: Catastrophic forgetting is a well-documented challenge in model fine-tuning, particularly when the downstream domain has limited labeled data or differs substantially from the pre-training distribution. Existing parameter-efficient fine-tuning methods largely operate in the weight space by modifying or augmenting the parameters of the pre-trained model, which can lead to models that are overly specialized to the observed downstream data. Recent studies suggest that one mechanism underlying such forgetting is the introduction of intruder dimensions into the representation space during fine-tuning. To mitigate the risk of overwriting pre-trained knowledge and to enhance robustness, we propose Vector-based Feature Adaptation (VeFA), a new fine-tuning method that operates directly in the feature space, which naturally avoids generating intruder dimensions. VeFA performs element-wise adaptation on individual features, thereby ensuring that the effective fine-tuned weights always remain within the column space of the pre-trained weight matrix. This feature-space adaptation perspective is inspired by the idea of effect equivalence modeling (EEM) of downstream lurking variables that induce distribution shifts, which posits that the influence of unobserved factors can be represented as an equivalent aggregate effect on observed features. By compensating for the effects of downstream lurking variables via a lightweight feature-level transformation, VeFA preserves the pre-trained representations and improves model generalization under distribution shift. We evaluate VeFA against LoRA on image classification, NLU, and NLG benchmarks, considering both standard fine-tuning performance and robustness; across these tasks, VeFA achieves comparable fine-tuning performance while consistently exhibiting stronger robustness.
[909] How Muon’s Spectral Design Benefits Generalization: A Study on Imbalanced Data
Bhavya Vasudeva, Puneesh Deora, Yize Zhao, Vatsal Sharan, Christos Thrampoulidis
Main category: cs.LG
TL;DR: Spectral optimizers like Muon and Shampoo outperform vanilla GD on imbalanced data by learning all data principal components equally, unlike GD which prioritizes dominant components first.
Details
Motivation: The growing adoption of spectrum-aware matrix-valued optimizers in deep learning motivates understanding their generalization properties and when they outperform competitive algorithms like Adam and vanilla gradient descent.Method: Study Spectral Gradient Descent (SpecGD) as canonical form of spectral optimizers. Analyze on Gaussian mixture data with linear/bilinear models. Compare SpecGD vs vanilla GD, including adaptive step-size variants. Extend to deep linear models. Empirically validate on imbalanced datasets comparing Muon, Shampoo, and their Euclidean counterparts.
Result: SpecGD learns all principal components equally while GD prioritizes dominant components first. This leads to growing balanced accuracy gap favoring SpecGD early in training, persisting even with adaptive step-sizes. Depth amplifies these effects. Empirical results validate superior generalization of spectral optimizers on imbalanced data.
Conclusion: Spectral optimizers achieve better generalization on imbalanced data by promoting balanced learning of all data components, unlike Euclidean methods that focus on dominant features. This explains practical advantages of Muon and Shampoo in deep learning applications.
Abstract: The growing adoption of spectrum-aware matrix-valued optimizers such as Muon and Shampoo in deep learning motivates a systematic study of their generalization properties and, in particular, when they might outperform competitive algorithms. We approach this question by introducing appropriate simplifying abstractions as follows: First, we use imbalanced data as a testbed. Second, we study the canonical form of such optimizers, which is Spectral Gradient Descent (SpecGD) – each update step is $UV^T$ where $UΣV^T$ is the truncated SVD of the gradient. Third, within this framework we identify a canonical setting for which we precisely quantify when SpecGD outperforms vanilla Euclidean GD. For a Gaussian mixture data model and both linear and bilinear models, we show that unlike GD, which prioritizes learning dominant principal components of the data first, SpecGD learns all principal components of the data at equal rates. We demonstrate how this translates to a growing gap in balanced accuracy favoring SpecGD early in training and further show that the gap remains consistent even when the GD counterpart uses adaptive step-sizes via normalization. By extending the analysis to deep linear models, we show that depth amplifies these effects. We empirically verify our theoretical findings on a variety of imbalanced datasets. Our experiments compare practical variants of spectral methods, like Muon and Shampoo, against their Euclidean counterparts and Adam. The results validate our findings that these spectral optimizers achieve superior generalization by promoting a more balanced learning of the data’s underlying components.
[910] Methodology for Comparing Machine Learning Algorithms for Survival Analysis
Lucas Buk Cardoso, Simone Aldrey Angelo, Yasmin Pacheco Gil Bonilha, Fernando Maia, Adeylson Guimarães Ribeiro, Maria Paula Curado, Gisele Aparecida Fernandes, Vanderlei Cunha Parro, Flávio Almeida de Magalhães Cipparrone, Alexandre Dias Porto Chiavegatto Filho, Victor Wünsch Filho, Tatiana Natasha Toporcov
Main category: cs.LG
TL;DR: Comparative analysis of 6 ML survival models on 45K colorectal cancer patients shows XGB-AFT performs best, highlighting ML’s potential for survival prediction and clinical decision support.
Details
Motivation: To evaluate and compare machine learning models for survival analysis (MLSA) in colorectal cancer prediction, addressing the need for improved survival prediction methods that can handle censored data and support clinical decision-making.Method: Used data from 45,000 colorectal cancer patients; evaluated 6 MLSA models (RSF, GBSA, SSVM, XGB-Cox, XGB-AFT, LGBM) with hyperparameter optimization; assessed performance using C-Index, C-Index IPCW, time-dependent AUC, and IBS; compared survival curves with classification algorithms; used SHAP and permutation importance for predictor interpretation.
Result: XGB-AFT achieved best performance (C-Index = 0.7618; IPCW = 0.7532), followed by GBSA and RSF. MLSA models demonstrated superior predictive capability compared to traditional classification approaches.
Conclusion: Machine learning survival analysis models show strong potential for improving survival prediction in colorectal cancer, with XGB-AFT emerging as the top performer, offering valuable tools for clinical decision support and personalized medicine.
Abstract: This study presents a comparative methodological analysis of six machine learning models for survival analysis (MLSA). Using data from nearly 45,000 colorectal cancer patients in the Hospital-Based Cancer Registries of São Paulo, we evaluated Random Survival Forest (RSF), Gradient Boosting for Survival Analysis (GBSA), Survival SVM (SSVM), XGBoost-Cox (XGB-Cox), XGBoost-AFT (XGB-AFT), and LightGBM (LGBM), capable of predicting survival considering censored data. Hyperparameter optimization was performed with different samplers, and model performance was assessed using the Concordance Index (C-Index), C-Index IPCW, time-dependent AUC, and Integrated Brier Score (IBS). Survival curves produced by the models were compared with predictions from classification algorithms, and predictor interpretation was conducted using SHAP and permutation importance. XGB-AFT achieved the best performance (C-Index = 0.7618; IPCW = 0.7532), followed by GBSA and RSF. The results highlight the potential and applicability of MLSA to improve survival prediction and support decision making.
[911] Limits of Generalization in RLVR: Two Case Studies in Mathematical Reasoning
Md Tanvirul Alam, Nidhi Rastogi
Main category: cs.LG
TL;DR: RLVR improves evaluation metrics on combinatorial problems but often reinforces superficial heuristics rather than teaching genuine mathematical reasoning, revealing limits in generalization.
Details
Motivation: To investigate whether Reinforcement Learning with Verifiable Rewards (RLVR) actually fosters genuine mathematical reasoning in LLMs, or if it merely teaches superficial heuristics that exploit shortcuts.Method: Tested RLVR on two combinatorial problems (Activity Scheduling and Longest Increasing Subsequence) with fully verifiable solutions, using carefully curated datasets with unique optima across multiple reward designs.
Result: RLVR improves evaluation metrics but often does so by reinforcing superficial heuristics rather than acquiring new reasoning strategies, showing limited generalization.
Conclusion: Highlights the limits of RLVR generalization and emphasizes the need for benchmarks that can disentangle genuine mathematical reasoning from shortcut exploitation to provide faithful measures of progress.
Abstract: Mathematical reasoning is a central challenge for large language models (LLMs), requiring not only correct answers but also faithful reasoning processes. Reinforcement Learning with Verifiable Rewards (RLVR) has emerged as a promising approach for enhancing such capabilities; however, its ability to foster genuine reasoning remains unclear. We investigate RLVR on two combinatorial problems with fully verifiable solutions: \emph{Activity Scheduling} and the \emph{Longest Increasing Subsequence}, using carefully curated datasets with unique optima. Across multiple reward designs, we find that RLVR improves evaluation metrics but often by reinforcing superficial heuristics rather than acquiring new reasoning strategies. These findings highlight the limits of RLVR generalization, emphasizing the importance of benchmarks that disentangle genuine mathematical reasoning from shortcut exploitation and provide faithful measures of progress. Code available at https://github.com/xashru/rlvr-seq-generalization.
[912] SAD-Flower: Flow Matching for Safe, Admissible, and Dynamically Consistent Planning
Tzu-Yuan Huang, Armin Lederer, Dai-Jie Wu, Xiaobing Dai, Sihua Zhang, Stefan Sosnowski, Shao-Hua Sun, Sandra Hirche
Main category: cs.LG
TL;DR: SAD-Flower is a novel flow matching framework that generates safe, admissible, and dynamically consistent trajectories by augmenting flows with virtual control inputs, providing formal guarantees without retraining.
Details
Motivation: Existing flow matching planners lack formal guarantees for state/action constraints and dynamic consistency, which are crucial for safety and executability of planned trajectories.Method: Augments flow with virtual control inputs and uses nonlinear control theory techniques to derive principled guidance, ensuring constraint satisfaction and dynamic consistency without retraining.
Result: Outperforms various generative-model-based baselines in ensuring constraint satisfaction across several tasks, with test-time satisfaction of unseen constraints.
Conclusion: SAD-Flower provides a principled framework for generating safe, admissible, and dynamically consistent trajectories with formal guarantees, addressing critical limitations of existing flow matching approaches.
Abstract: Flow matching (FM) has shown promising results in data-driven planning. However, it inherently lacks formal guarantees for ensuring state and action constraints, whose satisfaction is a fundamental and crucial requirement for the safety and admissibility of planned trajectories on various systems. Moreover, existing FM planners do not ensure the dynamical consistency, which potentially renders trajectories inexecutable. We address these shortcomings by proposing SAD-Flower, a novel framework for generating Safe, Admissible, and Dynamically consistent trajectories. Our approach relies on an augmentation of the flow with a virtual control input. Thereby, principled guidance can be derived using techniques from nonlinear control theory, providing formal guarantees for state constraints, action constraints, and dynamic consistency. Crucially, SAD-Flower operates without retraining, enabling test-time satisfaction of unseen constraints. Through extensive experiments across several tasks, we demonstrate that SAD-Flower outperforms various generative-model-based baselines in ensuring constraint satisfaction.
[913] Multivariate Variational Autoencoder
Mehmet Can Yavuz
Main category: cs.LG
TL;DR: MVAE introduces a full-covariance VAE with tractable multivariate Gaussian posterior that enables correlated latent factors while maintaining closed-form KL divergence and simple reparameterization.
Details
Motivation: Standard VAEs with diagonal Gaussian posteriors cannot capture correlated uncertainty and often produce entangled or redundant latent dimensions, limiting their expressiveness and calibration.Method: MVAE extends VAE with sample-specific diagonal scales and a global coupling matrix to create multivariate Gaussian posterior N(μ_φ(x), C diag(σ_φ²(x)) C⊤), preserving tractable KL divergence and reparameterization.
Result: MVAE consistently matches or outperforms diagonal-covariance VAEs across MNIST variants, Fashion-MNIST, and CIFAR-10/100, with notable gains in calibration and clustering metrics at both low and high latent dimensions.
Conclusion: MVAE provides a tractable full-covariance VAE framework that improves latent structure, calibration, and downstream performance while maintaining computational efficiency, with released code for reproducible research.
Abstract: Learning latent representations that are simultaneously expressive, geometrically well-structured, and reliably calibrated remains a central challenge for Variational Autoencoders (VAEs). Standard VAEs typically assume a diagonal Gaussian posterior, which simplifies optimization but rules out correlated uncertainty and often yields entangled or redundant latent dimensions. We introduce the Multivariate Variational Autoencoder (MVAE), a tractable full-covariance extension of the VAE that augments the encoder with sample-specific diagonal scales and a global coupling matrix. This induces a multivariate Gaussian posterior of the form $N(μ_φ(x), C \operatorname{diag}(σ_φ^2(x)) C^\top)$, enabling correlated latent factors while preserving a closed-form KL divergence and a simple reparameterization path. Beyond likelihood, we propose a multi-criterion evaluation protocol that jointly assesses reconstruction quality (MSE, ELBO), downstream discrimination (linear probes), probabilistic calibration (NLL, Brier, ECE), and unsupervised structure (NMI, ARI). Across Larochelle-style MNIST variants, Fashion-MNIST, and CIFAR-10/100, MVAE consistently matches or outperforms diagonal-covariance VAEs of comparable capacity, with particularly notable gains in calibration and clustering metrics at both low and high latent dimensions. Qualitative analyses further show smoother, more semantically coherent latent traversals and sharper reconstructions. All code, dataset splits, and evaluation utilities are released to facilitate reproducible comparison and future extensions of multivariate posterior models.
[914] Anomaly Detection in High-Dimensional Bank Account Balances via Robust Methods
Federico Maddanu, Tommaso Proietti, Riccardo Crupi
Main category: cs.LG
TL;DR: Proposes robust statistical methods for detecting anomalies in high-dimensional bank account balance data that are computationally efficient with high breakdown points.
Details
Motivation: Detecting anomalies in bank account balances is crucial for fraud detection and operational monitoring, but traditional robust statistics methods become inefficient and computationally expensive in high-dimensional settings.Method: Proposes several robust statistical approaches designed for computational efficiency in medium to high-dimensional datasets, with high breakdown points and low computational time.
Result: Empirical evaluation on a dataset of approximately 2.6 million daily records of anonymous users’ bank account balances demonstrates the methods’ effectiveness.
Conclusion: The proposed robust approaches offer practical solutions for financial institutions to efficiently detect anomalies in high-dimensional banking data while maintaining statistical robustness.
Abstract: Detecting point anomalies in bank account balances is essential for financial institutions, as it enables the identification of potential fraud, operational issues, or other irregularities. Robust statistics is useful for flagging outliers and for providing estimates of the data distribution parameters that are not affected by contaminated observations. However, such a strategy is often less efficient and computationally expensive under high dimensional setting. In this paper, we propose and evaluate empirically several robust approaches that may be computationally efficient in medium and high dimensional datasets, with high breakdown points and low computational time. Our application deals with around 2.6 million daily records of anonymous users’ bank account balances.
[915] Statistically Accurate and Robust Generative Prediction of Rock Discontinuities with A Tabular Foundation Model
Han Meng, Gang Mei, Hong Tian, Nengxiong Xu, Jianbing Peng
Main category: cs.LG
TL;DR: A tabular foundation model approach for generative prediction of rock discontinuities that outperforms conventional methods in accuracy and robustness, especially under data-sparse conditions.
Details
Motivation: Rock discontinuities are crucial for rock mass stability but their internal distributions are largely unobservable. Current generative prediction approaches either fail to capture complex distribution patterns or lack robustness when surface-exposed observations are sparse.Method: Proposed a simple yet robust approach using a tabular foundation model that leverages powerful sample learning capabilities specifically designed for small data, enabling effective capture of complex distribution patterns within limited measured discontinuities.
Result: Comparative experiments on ten datasets with diverse scales and distribution patterns demonstrate superior accuracy and robustness over conventional statistical models and deep generative approaches.
Conclusion: The work advances quantitative characterization of rock mass structures and supports safer, more reliable data-driven geotechnical design through statistically accurate generative prediction of rock discontinuities.
Abstract: Rock discontinuities critically govern the mechanical behavior and stability of rock masses. Their internal distributions remain largely unobservable and are typically inferred from surface-exposed discontinuities using generative prediction approaches. However, surface-exposed observations are inherently sparse, and existing generative prediction approaches either fail to capture the underlying complex distribution patterns or lack robustness under data-sparse conditions. Here, we proposed a simple yet robust approach for statistically accurate generative prediction of rock discontinuities by utilizing a tabular foundation model. By leveraging the powerful sample learning capability of the foundation model specifically designed for small data, our approach can effectively capture the underlying complex distribution patterns within limited measured discontinuities. Comparative experiments on ten datasets with diverse scales and distribution patterns of discontinuities demonstrate superior accuracy and robustness over conventional statistical models and deep generative approaches. This work advances quantitative characterization of rock mass structures, supporting safer and more reliable data-driven geotechnical design.
[916] AdamNX: An Adam improvement algorithm based on a novel exponential decay mechanism for the second-order moment estimate
Meng Zhu, Quan Xiao, Weidong Min
Main category: cs.LG
TL;DR: AdamNX: A new Adam variant with adaptive second-order moment decay that transitions to momentum SGD in stable training, improving generalization by converging to flatter minima.
Details
Motivation: Adam optimization tends to converge to non-flat minima compared to SGD, which may harm generalization. As large language models scale up, this issue becomes more critical despite Adam remaining the mainstream optimizer.Method: Proposes AdamNX with a novel adaptive second-order moment estimation exponential decay rate that gradually weakens learning step correction strength during training, effectively degrading to momentum SGD in stable periods.
Result: Experimental results show the proposed second-order moment decay rate outperforms current methods, and AdamNX consistently surpasses Adam and its variants in performance.
Conclusion: AdamNX improves training stability in stable periods and potentially enhances generalization by converging to flatter minima, offering a better optimization solution for large-scale models.
Abstract: Since the 21st century, artificial intelligence has been leading a new round of industrial revolution. Under the training framework, the optimization algorithm aims to stably converge high-dimensional optimization to local and even global minima. Entering the era of large language models, although the scale of model parameters and data has increased, Adam remains the mainstream optimization algorithm. However, compared with stochastic gradient descent (SGD) based optimization algorithms, Adam is more likely to converge to non-flat minima. To address this issue, the AdamNX algorithm is proposed. Its core innovation lies in the proposition of a novel type of second-order moment estimation exponential decay rate, which gradually weakens the learning step correction strength as training progresses, and degrades to momentum SGD in the stable training period, thereby improving the stability of training in the stable period and possibly enhancing generalization ability. Experimental results show that our second-order moment estimation exponential decay rate is better than the current second-order moment estimation exponential decay rate, and AdamNX can stably outperform Adam and its variants in terms of performance. Our code is open-sourced at https://github.com/mengzhu0308/AdamNX.
[917] Extended Physics Informed Neural Network for Hyperbolic Two-Phase Flow in Porous Media
Saif Ur Rehman, Wajid Yousuf
Main category: cs.LG
TL;DR: XPINN framework outperforms standard PINNs for solving nonlinear Buckley-Leverett equation by dynamically decomposing domain into pre-shock/post-shock subdomains with Rankine-Hugoniot coupling, achieving accurate sharp front resolution with simpler Adam optimizer.
Details
Motivation: Standard PINNs struggle with steep gradients, discontinuities, and complex nonlinear wave interactions in hyperbolic PDEs like the Buckley-Leverett equation for two-phase flow in porous media.Method: Extended PINN (XPINN) framework with dynamic space-time domain decomposition into evolving pre-shock and post-shock subdomains, using localized subnetworks with Rankine-Hugoniot jump condition coupling for flux continuity.
Result: XPINN consistently outperforms standard PINNs and variants (PINN with artificial viscosity, Welge construction, Oleinik entropy condition), accurately resolving sharp fronts and correct physical behavior with simpler Adam optimizer.
Conclusion: XPINN is a robust, scalable approach for challenging hyperbolic PDEs without needing artificial diffusion or entropy corrections, demonstrating superior performance for nonlinear Buckley-Leverett equation.
Abstract: The accurate solution of nonlinear hyperbolic partial differential equations (PDEs) remains challenging due to steep gradients, discontinuities, and multiscale structures that make conventional solvers computationally demanding. Physics-Informed Neural Networks (PINNs) embed the governing equations into the learning process, enabling mesh-free solution of PDEs, yet they often struggle to capture steep gradients, discontinuities, and complex nonlinear wave interactions. To address these limitations, we employ the Extended Physics-Informed Neural Network (XPINN) framework to solve the nonlinear Buckley-Leverett equation with a nonconvex flux, modeling immiscible two-phase flow in porous media. The computational domain is dynamically decomposed in space and time into evolving pre-shock and post-shock subdomains, allowing localized subnetworks to efficiently learn distinct flow behaviors, with coupling enforced via the Rankine-Hugoniot jump condition to ensure physically consistent flux continuity. We compare XPINN with standard PINNs and its variants, including PINN with artificial viscosity, PINN with Welge construction, and PINN with the Oleinik entropy condition, and across all cases, XPINN consistently outperforms the other methods, accurately resolving sharp fronts and capturing the correct physical behavior. Importantly, XPINN achieves this using the simpler Adam optimizer, whereas some PINN variants require more complex or higher-order strategies such as L-BFGS to reach comparable accuracy, demonstrating that XPINN is a robust and scalable approach for challenging hyperbolic PDEs without artificial diffusion or entropy corrections. The code is available at github.com/saifkhanengr/XPINN-for-Buckley-Leverett.
[918] Learning Robust Social Strategies with Large Language Models
Dereck Piche, Mohammed Muqeeth, Milad Aghajohari, Juan Duque, Michael Noukhovitch, Aaron Courville
Main category: cs.LG
TL;DR: LLMs trained with standard RL in multi-agent settings develop opportunistic, exploitative behavior despite cooperative priors. The paper adapts Advantage Alignment algorithm with group-relative baseline to train LLMs for cooperation and introduces Trust-and-Split social dilemma environment.
Details
Motivation: As agentic AI becomes widespread, multi-agent interactions with conflicting goals pose challenges in social dilemmas where individual incentives undermine collective welfare. Standard RL in multi-agent settings often converges to defecting policies, even in LLMs with cooperative priors.Method: Adapted Advantage Alignment algorithm for opponent-learning awareness, introduced group-relative baseline to simplify advantage computation in iterated games, and created Trust-and-Split social dilemma environment requiring natural language communication for high collective welfare.
Result: Across various social dilemmas, policies learned with Advantage Alignment achieve higher collective payoffs while remaining robust against exploitation by greedy agents. RL-trained LLM agents develop opportunistic behavior that can exploit even advanced closed-source models.
Conclusion: The proposed Advantage Alignment approach effectively trains LLMs for multi-agent cooperation and non-exploitability, addressing RL’s tendency to converge to poor equilibria. The work provides tools for future multi-agent RL training for LLMs.
Abstract: As agentic AI becomes more widespread, agents with distinct and possibly conflicting goals will interact in complex ways. These multi-agent interactions pose a fundamental challenge, particularly in social dilemmas, where agents’ individual incentives can undermine collective welfare. While reinforcement learning (RL) has been effective for aligning large language models (LLMs) in the single-agent regime, prior small-network results suggest that standard RL in multi-agent settings often converges to defecting, self-interested policies. We show the same effect in LLMs: despite cooperative priors, RL-trained LLM agents develop opportunistic behavior that can exploit even advanced closed-source models. To address this tendency of RL to converge to poor equilibria, we adapt a recent opponent-learning awareness algorithm, Advantage Alignment, to fine-tune LLMs toward multi-agent cooperation and non-exploitability. We then introduce a group-relative baseline that simplifies advantage computation in iterated games, enabling multi-agent training at LLM scale. We also contribute a novel social dilemma environment, Trust-and-Split, which requires natural language communication to achieve high collective welfare. Across a wide range of social dilemmas, policies learned with Advantage Alignment achieve higher collective payoffs while remaining robust against exploitation by greedy agents. We release all of our code to support future work on multi-agent RL training for LLMs.
[919] Probabilistic Hash Embeddings for Online Learning of Categorical Features
Aodong Li, Abishek Sankararaman, Balakrishnan Narayanaswamy
Main category: cs.LG
TL;DR: Probabilistic Hash Embeddings (PHE) for online learning with evolving categorical vocabularies, using Bayesian inference to prevent forgetting and maintain memory efficiency.
Details
Motivation: Traditional deterministic hash embeddings for categorical features suffer from forgetting and sensitivity to arrival order in online learning settings where vocabularies evolve and grow unboundedly over time.Method: Proposes Probabilistic Hash Embeddings (PHE) that treat hash embeddings as stochastic variables, using Bayesian online learning with scalable inference algorithms to incrementally learn from streaming data while maintaining bounded parameter sets.
Result: PHE demonstrates superior performance in classification, sequence modeling, and recommendation systems while consuming only 2-4% of memory compared to one-hot embedding tables, with invariance to item arrival order and no forgetting of old items.
Conclusion: PHE provides an effective solution for online learning with evolving categorical vocabularies, addressing key limitations of deterministic hash embeddings through probabilistic modeling and Bayesian inference.
Abstract: We study streaming data with categorical features where the vocabulary of categorical feature values is changing and can even grow unboundedly over time. Feature hashing is commonly used as a pre-processing step to map these categorical values into a feature space of fixed size before learning their embeddings. While these methods have been developed and evaluated for offline or batch settings, in this paper we consider online settings. We show that deterministic embeddings are sensitive to the arrival order of categories and suffer from forgetting in online learning, leading to performance deterioration. To mitigate this issue, we propose a probabilistic hash embedding (PHE) model that treats hash embeddings as stochastic and applies Bayesian online learning to learn incrementally from data. Based on the structure of PHE, we derive a scalable inference algorithm to learn model parameters and infer/update the posteriors of hash embeddings and other latent variables. Our algorithm (i) can handle an evolving vocabulary of categorical items, (ii) is adaptive to new items without forgetting old items, (iii) is implementable with a bounded set of parameters that does not grow with the number of distinct observed values on the stream, and (iv) is invariant to the item arrival order. Experiments in classification, sequence modeling, and recommendation systems in online learning setups demonstrate the superior performance of PHE while maintaining high memory efficiency (consumes as low as 2~4 memory of a one-hot embedding table). Supplementary materials are at https://github.com/aodongli/probabilistic-hash-embeddings
[920] Semantic Superiority vs. Forensic Efficiency: A Comparative Analysis of Deep Learning and Psycholinguistics for Business Email Compromise Detection
Yaw Osei Adjei, Frederick Ayivor
Main category: cs.LG
TL;DR: This paper compares two BEC detection approaches: a psycholinguistic CatBoost model analyzing linguistic cues and a semantic DistilBERT model using deep learning, showing both achieve high accuracy with different latency/resource tradeoffs suitable for different deployment scenarios.
Details
Motivation: BEC causes massive financial losses ($2.9B annually) with extreme economic asymmetry where fraud losses far exceed review costs (5,480:1 ratio), creating urgent need for effective detection systems.Method: Two detection paradigms: 1) Forensic Psycholinguistic Stream using CatBoost to analyze linguistic cues like urgency and authority, and 2) Semantic Stream using DistilBERT for contextual understanding. Evaluated on hybrid dataset (N=7,990) with human-legitimate and AI-synthesized adversarial fraud on Tesla T4 infrastructure.
Result: DistilBERT achieved near-perfect detection on synthetic threats (AUC >0.99, F1=0.998) with 7.4ms latency. CatBoost achieved competitive detection (AUC=0.991, F1=0.949) at 8.4x lower latency (0.8ms) with negligible resource consumption.
Conclusion: DistilBERT offers maximum accuracy for GPU-equipped organizations, while CatBoost provides cost-effective alternative for edge deployments. Both approaches demonstrate theoretical ROI exceeding 99.9% when optimized via cost-sensitive learning.
Abstract: Business Email Compromise (BEC) is a sophisticated social engineering threat that manipulates organizational hierarchies, leading to significant financial damage. According to the 2024 FBI Internet Crime Report, BEC accounts for over $2.9 billion in annual losses, presenting a massive economic asymmetry: the financial cost of a False Negative (fraud loss) exceeds the operational cost of a False Positive (manual review) by a ratio of approximately 5,480:1. This paper contrasts two detection paradigms: a Forensic Psycholinguistic Stream (CatBoost), which analyzes linguistic cues like urgency and authority with high interpretability, and a Semantic Stream (DistilBERT), which utilizes deep learning for contextual understanding. We evaluated both streams on a hybrid dataset (N=7,990) containing human-legitimate and AI-synthesized adversarial fraud. Benchmarked on Tesla T4 infrastructure, DistilBERT achieved near-perfect detection on synthetic threats (AUC >0.99, F1 =0.998) with acceptable real-time latency (7.4 ms). CatBoost achieved competitive detection (AUC =0.991, F1 =0.949) at 8.4x lower latency (0.8 ms) with negligible resource consumption. We conclude that while DistilBERT offers maximum accuracy for GPU-equipped organizations, CatBoost provides a viable, cost-effective alternative for edge deployments. Both approaches demonstrate a theoretical ROI exceeding 99.9% when optimized via cost-sensitive learning.
[921] DSD: A Distributed Speculative Decoding Solution for Edge-Cloud Agile Large Model Serving
Fengze Yu, Leshu Li, Brad McDanel, Sai Qian Zhang
Main category: cs.LG
TL;DR: DSD is a distributed speculative decoding framework that extends speculative decoding to multi-device deployments across edge-cloud environments, achieving up to 1.1x speedup and 9.7% higher throughput over existing baselines.
Details
Motivation: LLM inference suffers from high decoding latency and limited scalability across heterogeneous edge-cloud environments. Existing speculative decoding techniques are confined to single-node execution and don't leverage distributed computing resources effectively.Method: Proposes DSD framework with coordinated draft-target execution across multiple devices. Introduces DSD-Sim simulator to capture network, batching, and scheduling dynamics. Designs Adaptive Window Control policy that dynamically adjusts speculation window size based on simulation insights.
Result: DSD achieves up to 1.1x speedup and 9.7% higher throughput over existing speculative decoding baselines. Enables agile and scalable LLM serving across edge and cloud environments.
Conclusion: Distributed speculative decoding through DSD framework effectively addresses LLM inference latency and scalability challenges in heterogeneous edge-cloud deployments, outperforming single-node speculative decoding approaches.
Abstract: Large language model (LLM) inference often suffers from high decoding latency and limited scalability across heterogeneous edge-cloud environments. Existing speculative decoding (SD) techniques accelerate token generation but remain confined to single-node execution. We propose DSD, a distributed speculative decoding framework that extends SD to multi-device deployments through coordinated draft-target execution. Given the lack of prior work on simulating this paradigm, we first introduce DSD-Sim, a discrete-event simulator that captures network, batching, and scheduling dynamics. Building on insights from DSD-Sim, we further design an Adaptive Window Control (AWC) policy that dynamically adjusts speculation window size to optimize throughput. Experiments across diverse workloads show that DSD achieves up to 1.1x speedup and 9.7% higher throughput over existing SD baselines, enabling agile and scalable LLM serving across edge and cloud.
cs.MA
[922] Toward a Safe Internet of Agents
Juan A. Wibowo, George C. Polyzos
Main category: cs.MA
TL;DR: This survey provides an architectural framework for engineering safe and reliable autonomous agent systems powered by LLMs, identifying vulnerabilities at three complexity levels to establish secure design principles for the emerging Internet of Agents.
Details
Motivation: The paper addresses the novel and systemic safety and security risks introduced by the paradigm shift toward an "Internet of Agents" (IoA) powered by Large Language Models. Unlike threat-centric approaches, it aims to provide a principled architectural framework for secure agentic system design.Method: The authors perform a bottom-up deconstruction of agentic systems, treating each component as a dual-use interface. The analysis spans three levels of complexity: Single Agent, Multi-Agent System (MAS), and Interoperable Multi-Agent System (IMAS), identifying core architectural components and their inherent security risks at each level.
Result: The central finding is that agentic safety is an architectural principle rather than an add-on. The survey identifies specific vulnerabilities and derives mitigation principles at each level of the agentic stack, establishing a foundation for secure design.
Conclusion: This survey serves as a foundational guide for building capable, safe, and trustworthy AI needed to realize a secure Internet of Agents, emphasizing that security must be built into the architecture rather than added as an afterthought.
Abstract: Background: Autonomous agents powered by Large Language Models (LLMs) are driving a paradigm shift toward an “Internet of Agents” (IoA). While offering immense potential, this vision also introduces novel and systemic risks to safety and security. Objectives: Unlike common threat-centric taxonomies, our survey provides a principled, architectural framework for engineering safe and reliable agentic systems. We aim to identify the architectural sources of vulnerabilities to establish a foundation for secure design. Methods: We perform a bottom-up deconstruction of agentic systems, treating each component as a dual-use interface. The analysis spans three levels of complexity: the foundational Single Agent, the collaborative Multi-Agent System (MAS), and the visionary Interoperable Multi-Agent System (IMAS). At each level, we identify core architectural components and their inherent security risks. Results & Conclusions: Our central finding is that agentic safety is an architectural principle, not an add-on. By identifying specific vulnerabilities and deriving mitigation principles at each level of the agentic stack, this survey serves as a foundational guide for building the capable, safe, and trustworthy AI needed to realize a secure Internet of Agents.
[923] AgentODRL: A Large Language Model-based Multi-agent System for ODRL Generation
Wanle Zhong, Keman Huang, Xiaoyong Du
Main category: cs.MA
TL;DR: AgentODRL: A multi-agent LLM-based system that translates complex natural language data rights policies into ODRL format using an Orchestrator-Workers architecture with specialized components for generation, decomposition, and rewriting.
Details
Motivation: Current methods struggle to translate complex natural language authorization policies into ODRL format due to logical complexity and lack of high-quality training datasets, hindering automated data rights management.Method: AgentODRL uses an Orchestrator-Workers architecture with specialized agents: Generator (creates ODRL policies with validator-based syntax and semantic reflection), Decomposer (breaks down complex use cases), and Rewriter (simplifies nested logical relationships). The Orchestrator dynamically coordinates these workers based on input complexity.
Result: The system was tested on 770 data space use cases of varying complexity and achieved superior performance on ODRL generation tasks, as measured by ODRL syntax and semantic scores.
Conclusion: AgentODRL successfully addresses the challenge of translating complex natural language policies to ODRL by leveraging LLMs in a multi-agent architecture, enabling more efficient and accurate automated data rights management.
Abstract: The Open Digital Rights Language (ODRL) is a pivotal standard for automating data rights management. However, the inherent logical complexity of authorization policies, combined with the scarcity of high-quality “Natural Language-to-ODRL” training datasets, impedes the ability of current methods to efficiently and accurately translate complex rules from natural language into the ODRL format. To address this challenge, this research leverages the potent comprehension and generation capabilities of Large Language Models (LLMs) to achieve both automation and high fidelity in this translation process. We introduce AgentODRL, a multi-agent system based on an Orchestrator-Workers architecture. The architecture consists of specialized Workers, including a Generator for ODRL policy creation, a Decomposer for breaking down complex use cases, and a Rewriter for simplifying nested logical relationships. The Orchestrator agent dynamically coordinates these Workers, assembling an optimal pathway based on the complexity of the input use case. Specifically, we enhance the ODRL Generator by incorporating a validator-based syntax strategy and a semantic reflection mechanism powered by a LoRA-finetuned model, significantly elevating the quality of the generated policies. Extensive experiments were conducted on a newly constructed dataset comprising 770 use cases of varying complexity, all situated within the context of data spaces. The results, evaluated using ODRL syntax and semantic scores, demonstrate that our proposed Orchestrator-Workers system, enhanced with these strategies, achieves superior performance on the ODRL generation task.
[924] Hierarchical Decentralized Multi-Agent Coordination with Privacy-Preserving Knowledge Sharing: Extending AgentNet for Scalable Autonomous Systems
Goutham Nalagatla
Main category: cs.MA
TL;DR: AgentNet++ extends AgentNet with hierarchical organization, privacy-preserving techniques, and adaptive resource management to address scalability, communication overhead, and privacy limitations in decentralized multi-agent systems.
Details
Motivation: Address limitations of AgentNet including scalability challenges with large agent populations, communication overhead, lack of privacy guarantees, and suboptimal resource allocation in decentralized multi-agent systems.Method: Hierarchical decentralized framework with multilevel agent organization, privacy-preserving knowledge sharing via differential privacy and secure aggregation, adaptive resource management, and cluster-based hierarchies where agents self-organize into specialized groups.
Result: Achieves 23% higher task completion rates, 40% reduction in communication overhead, maintains strong privacy guarantees, scales effectively to 1000+ agents, and preserves emergent intelligence properties compared to AgentNet and other baselines.
Conclusion: AgentNet++ successfully addresses key limitations of decentralized multi-agent systems through hierarchical organization and privacy-preserving techniques while maintaining full decentralization and improving performance metrics significantly.
Abstract: Decentralized multi-agent systems have shown promise in enabling autonomous collaboration among LLM-based agents. While AgentNet demonstrated the feasibility of fully decentralized coordination through dynamic DAG topologies, several limitations remain: scalability challenges with large agent populations, communication overhead, lack of privacy guarantees, and suboptimal resource allocation. We propose AgentNet++, a hierarchical decentralized framework that extends AgentNet with multilevel agent organization, privacy-preserving knowledge sharing via differential privacy and secure aggregation, adaptive resource management, and theoretical convergence guarantees. Our approach introduces cluster-based hierarchies where agents self-organize into specialized groups, enabling efficient task routing and knowledge distillation while maintaining full decentralization. We provide formal analysis of convergence properties and privacy bounds, and demonstrate through extensive experiments on complex multi-agent tasks that AgentNet++ achieves 23% higher task completion rates, 40% reduction in communication overhead, and maintains strong privacy guarantees compared to AgentNet and other baselines. Our framework scales effectively to 1000+ agents while preserving the emergent intelligence properties of the original AgentNet.
[925] Augmented Runtime Collaboration for Self-Organizing Multi-Agent Systems: A Hybrid Bi-Criteria Routing Approach
Qingwen Yang, Feiyu Qu, Tiezheng Guo, Yanyi Liu, Yingyou Wen
Main category: cs.MA
TL;DR: BiRouter is a dual-criteria routing method for self-organizing multi-agent systems that enables agents to autonomously route tasks using only local information by balancing long-term importance and contextual continuity metrics.
Details
Motivation: Current LLM-based multi-agent systems use static topologies and centralized planning, which limits scalability and adaptability in open, decentralized networks. There's a need for effective collaboration planning in distributed systems using only local information.Method: BiRouter uses a dual-criteria routing approach where each agent autonomously executes “next-hop” task routing at runtime based on local information. It balances two metrics: ImpScore (evaluates candidate agent’s long-term importance to overall goal) and GapScore (assesses contextual continuity for current task state). Includes dynamically updated reputation mechanism for robustness in untrustworthy environments.
Result: BiRouter achieves superior performance and token efficiency over existing baselines, while maintaining strong robustness and effectiveness in information-limited, decentralized, and untrustworthy settings. A large-scale, cross-domain dataset with thousands of annotated task-routing paths was developed to enhance generalization.
Conclusion: BiRouter provides an effective solution for decentralized multi-agent collaboration by enabling autonomous task routing with local information, addressing scalability and adaptability limitations of centralized approaches while maintaining robustness in challenging environments.
Abstract: LLM-based multi-agent systems have demonstrated significant capabilities across diverse domains. However, the task performance and efficiency are fundamentally constrained by their collaboration strategies. Prevailing approaches rely on static topologies and centralized global planning, a paradigm that limits their scalability and adaptability in open, decentralized networks. Effective collaboration planning in distributed systems using only local information thus remains a formidable challenge. To address this, we propose BiRouter, a novel dual-criteria routing method for Self-Organizing Multi-Agent Systems (SO-MAS). This method enables each agent to autonomously execute ``next-hop’’ task routing at runtime, relying solely on local information. Its core decision-making mechanism is predicated on balancing two metrics: (1) the ImpScore, which evaluates a candidate agent’s long-term importance to the overall goal, and (2) the GapScore, which assesses its contextual continuity for the current task state. Furthermore, we introduce a dynamically updated reputation mechanism to bolster system robustness in untrustworthy environments and have developed a large-scale, cross-domain dataset, comprising thousands of annotated task-routing paths, to enhance the model’s generalization. Extensive experiments demonstrate that BiRouter achieves superior performance and token efficiency over existing baselines, while maintaining strong robustness and effectiveness in information-limited, decentralized, and untrustworthy settings.
[926] Chain of Unit-Physics: A Primitive-Centric Approach to Scientific Code Synthesis
Vansh Sharma, Venkat Raman
Main category: cs.MA
TL;DR: Chain of Unit-Physics framework enables reliable scientific code generation by embedding first-principles constraints through unit-physics tests, outperforming standard LLM approaches on combustion benchmark.
Details
Motivation: Current LLMs struggle with scientific code generation due to sparse domain code representation in training data and limited feasibility of RLHF with small expert communities, raising reliability concerns for high-stakes scientific computing problems.Method: Proposes Chain of Unit-Physics framework: a first-principles-centric, multi-agent system where human expert knowledge is encoded as unit-physics tests that explicitly constrain code generation, evaluated on a nontrivial combustion benchmark task.
Result: Framework converges in 5-6 iterations, matches human-expert implementation (mean error 3.1×10⁻³%), with ~33.4% faster runtime and ~30% more efficient memory usage at comparable cost to commercial APIs. Standard LLM approaches fail with four recurrent error classes.
Conclusion: Chain of Unit-Physics framework provides practical template for physics-grounded scientific code generation by embedding first-principles analysis, addressing fundamental limitations of current LLM approaches for scientific computing.
Abstract: Agentic large language models are proposed as autonomous code generators for scientific computing, yet their reliability in high-stakes problems remains unclear. Developing computational scientific software from natural-language queries remains challenging broadly due to (a) sparse representation of domain codes during training and (b) the limited feasibility of RLHF with a small expert community. To address these limitations, this work conceptualizes an inverse approach to code design, embodied in the Chain of Unit-Physics framework: a first-principles (or primitives)-centric, multi-agent system in which human expert knowledge is encoded as unit-physics tests that explicitly constrain code generation. The framework is evaluated on a nontrivial combustion task, used here as a representative benchmark for scientific problem with realistic physical constraints. Closed-weight systems and code-focused agentic variants fail to produce correct end-to-end solvers, despite tool and web access, exhibiting four recurrent error classes: interface (syntax/API) hallucinations, overconfident assumptions, numerical/physical incoherence, and configuration fragility. Open-weight models with chain-of-thought (CoT) decoding reduce interface errors but still yield incorrect solutions. On the benchmark task, the proposed framework converges within 5-6 iterations, matches the human-expert implementation (mean error of $3.1\times10^{-3}$ %), with a $\sim$33.4 % faster runtime and a $\sim$30 % efficient memory usage at a cost comparable to mid-sized commercial APIs, yielding a practical template for physics-grounded scientific code generation. As datasets and models evolve, zero-shot code accuracy will improve; however, the Chain of Unit-Physics framework goes further by embedding first-principles analysis that is foundational to scientific codes.
[927] SocialDriveGen: Generating Diverse Traffic Scenarios with Controllable Social Interactions
Jiaguo Tian, Zhengbang Zhu, Shenyu Zhang, Li Xu, Bo Zheng, Xu Liu, Weiji Peng, Shizeng Yao, Weinan Zhang
Main category: cs.MA
TL;DR: SocialDriveGen is a hierarchical framework that generates realistic traffic scenarios by integrating semantic reasoning, social preference modeling (egoism/altruism), and generative trajectory synthesis to create diverse driver behaviors.
Details
Motivation: Existing simulation frameworks use rule-based or simplified models that lack fidelity and diversity for real-world driving. Current generative models overlook how social preferences influence driving behavior, creating a gap in realistic traffic scenario generation.Method: Hierarchical framework combining semantic reasoning and social preference modeling with generative trajectory synthesis. Models egoism and altruism as complementary social dimensions to enable controllable diversity in driver personalities and interaction styles.
Result: Experiments on Argoverse 2 dataset show SocialDriveGen generates diverse, high-fidelity traffic scenarios spanning cooperative to adversarial behaviors, significantly enhancing policy robustness and generalization to rare or high-risk situations.
Conclusion: SocialDriveGen addresses the gap in realistic traffic scenario generation by incorporating social preference modeling, producing diverse and high-fidelity scenarios that improve autonomous driving system evaluation and development.
Abstract: The generation of realistic and diverse traffic scenarios in simulation is essential for developing and evaluating autonomous driving systems. However, most simulation frameworks rely on rule-based or simplified models for scene generation, which lack the fidelity and diversity needed to represent real-world driving. While recent advances in generative modeling produce more realistic and context-aware traffic interactions, they often overlook how social preferences influence driving behavior. SocialDriveGen addresses this gap through a hierarchical framework that integrates semantic reasoning and social preference modeling with generative trajectory synthesis. By modeling egoism and altruism as complementary social dimensions, our framework enables controllable diversity in driver personalities and interaction styles. Experiments on the Argoverse 2 dataset show that SocialDriveGen generates diverse, high-fidelity traffic scenarios spanning cooperative to adversarial behaviors, significantly enhancing policy robustness and generalization to rare or high-risk situations.
[928] Agent-Kernel: A MicroKernel Multi-Agent System Framework for Adaptive Social Simulation Powered by LLMs
Yuren Mao, Peigen Liu, Xinjian Wang, Rui Ding, Jing Miao, Hui Zou, Mingjie Qi, Wanxiang Luo, Longbin Lai, Kai Wang, Zhengping Qian, Peilun Yang, Yunjun Gao, Ying Zhang
Main category: cs.MA
TL;DR: Agent-Kernel is a novel society-centric modular microkernel framework for LLM-powered multi-agent simulations that addresses limitations in adaptability, configurability, reliability, and reusability of existing frameworks.
Details
Motivation: Existing Multi-Agent System frameworks fail to support large-scale social simulations due to limitations in adaptability, configurability, reliability, and code reusability. They cannot handle dynamic societies where agent populations and profiles change over time.Method: Proposes Agent-Kernel framework built on a society-centric modular microkernel architecture that decouples core system functions from simulation logic, and separates cognitive processes from physical environments and action execution.
Result: Agent-Kernel achieves superior adaptability, configurability, reliability, and reusability. Validated through two applications: Universe 25 (Mouse Utopia) experiment simulation demonstrating handling of rapid population dynamics, and large-scale simulation of Zhejiang University Campus Life coordinating 10,000 heterogeneous agents.
Conclusion: Agent-Kernel successfully addresses the limitations of existing MAS frameworks for LLM-powered social simulations, enabling large-scale, dynamic simulations with changing agent populations and profiles through its novel modular architecture.
Abstract: Multi-Agent System (MAS) developing frameworks serve as the foundational infrastructure for social simulations powered by Large Language Models (LLMs). However, existing frameworks fail to adequately support large-scale simulation development due to inherent limitations in adaptability, configurability, reliability, and code reusability. For example, they cannot simulate a society where the agent population and profiles change over time. To fill this gap, we propose Agent-Kernel, a framework built upon a novel society-centric modular microkernel architecture. It decouples core system functions from simulation logic and separates cognitive processes from physical environments and action execution. Consequently, Agent-Kernel achieves superior adaptability, configurability, reliability, and reusability. We validate the framework’s superiority through two distinct applications: a simulation of the Universe 25 (Mouse Utopia) experiment, which demonstrates the handling of rapid population dynamics from birth to death; and a large-scale simulation of the Zhejiang University Campus Life, successfully coordinating 10,000 heterogeneous agents, including students and faculty.
[929] R3DM: Enabling Role Discovery and Diversity Through Dynamics Models in Multi-agent Reinforcement Learning
Harsh Goel, Mohammad Omama, Behdad Chalaki, Vaishnav Tadiparthi, Ehsan Moradi Pari, Sandeep Chinchali
Main category: cs.MA
TL;DR: R3DM is a novel role-based MARL framework that learns emergent roles by maximizing mutual information between agents’ roles, observed trajectories, and expected future behaviors, outperforming state-of-the-art methods by up to 20% win rates.
Details
Motivation: Existing role-based MARL methods only derive roles from past experience, ignoring how roles should shape future behavior for effective coordination. Biological systems show that roles naturally emerge to enable coordination, suggesting roles should influence future trajectories.Method: Proposes R3DM (Role Discovery and Diversity through Dynamics Models) that learns emergent roles by maximizing mutual information between agents’ roles, observed trajectories, and expected future behaviors. Uses contrastive learning on past trajectories to derive intermediate roles, then shapes intrinsic rewards to promote diversity in future behaviors through a learned dynamics model.
Result: Benchmarking on SMAC and SMACv2 environments shows R3DM outperforms state-of-the-art MARL approaches, improving multi-agent coordination to increase win rates by up to 20%.
Conclusion: R3DM successfully addresses the limitation of existing role-based MARL methods by ensuring roles shape future behavior, leading to better coordination and performance in complex multi-agent tasks. The framework demonstrates the importance of considering future trajectories in role discovery for effective multi-agent coordination.
Abstract: Multi-agent reinforcement learning (MARL) has achieved significant progress in large-scale traffic control, autonomous vehicles, and robotics. Drawing inspiration from biological systems where roles naturally emerge to enable coordination, role-based MARL methods have been proposed to enhance cooperation learning for complex tasks. However, existing methods exclusively derive roles from an agent’s past experience during training, neglecting their influence on its future trajectories. This paper introduces a key insight: an agent’s role should shape its future behavior to enable effective coordination. Hence, we propose Role Discovery and Diversity through Dynamics Models (R3DM), a novel role-based MARL framework that learns emergent roles by maximizing the mutual information between agents’ roles, observed trajectories, and expected future behaviors. R3DM optimizes the proposed objective through contrastive learning on past trajectories to first derive intermediate roles that shape intrinsic rewards to promote diversity in future behaviors across different roles through a learned dynamics model. Benchmarking on SMAC and SMACv2 environments demonstrates that R3DM outperforms state-of-the-art MARL approaches, improving multi-agent coordination to increase win rates by up to 20%. The code is available at https://github.com/UTAustin-SwarmLab/R3DM}{https://github.com/UTAustin-SwarmLab/R3DM.
[930] DISPATCH – Decentralized Informed Spatial Planning and Assignment of Tasks for Cooperative Heterogeneous Agents
Yao Liu, Sampad Mohanty, Elizabeth Ondula, Bhaskar Krishnamachari
Main category: cs.MA
TL;DR: Connects Eisenberg-Gale equilibrium to decentralized multi-agent learning for fair spatial task allocation under partial observability, developing two algorithms that balance fairness and efficiency.
Details
Motivation: Greedy assignment policies in multi-robot/ride-sharing systems maximize efficiency but create inequities where some tasks get favorable service while others face long waits. Need fairness-aware approaches for heterogeneous multi-agent systems with partial observability.Method: Establishes connection between Eisenberg-Gale equilibrium and decentralized multi-agent learning. Develops two algorithms: (1) EG-MARL - multi-agent reinforcement learning guided by centralized fair assignment algorithms (EG and preference-aware Hungarian method), (2) stochastic online optimization with guided exploration and subset-based fair assignment.
Result: Both algorithms preserve fairness-efficiency balance of Eisenberg-Gale equilibrium under partial observability. EG-MARL achieves near-centralized coordination and reduced travel distances. Stochastic online mechanism enables real-time allocation with competitive fairness.
Conclusion: Spatially aware Eisenberg-Gale formulations can effectively guide decentralized coordination in agents with heterogeneous capabilities, balancing fairness and efficiency in partially observable environments.
Abstract: Spatial task allocation in systems such as multi-robot delivery or ride-sharing requires balancing efficiency with fair service across tasks. Greedy assignment policies that match each agent to its highest-preference or lowest-cost task can maximize efficiency but often create inequities: some tasks receive disproportionately favorable service (e.g., shorter delays or better matches), while others face long waits or poor allocations. We study fairness in heterogeneous multi-agent systems where tasks vary in preference alignment and urgency. Most existing approaches either assume centralized coordination or largely ignore fairness under partial observability. Distinct from this prior work, we establish a connection between the Eisenberg-Gale (EG) equilibrium convex program and decentralized, partially observable multi-agent learning. Building on this connection, we develop two equilibrium-informed algorithms that integrate fairness and efficiency: (i) a multi-agent reinforcement learning (MARL) framework, EG-MARL, whose training is guided by centralized fair assignment algorithms (EG and a preference-aware Hungarian method); and (ii) a stochastic online optimization mechanism that performs guided exploration and subset-based fair assignment as tasks are discovered. We evaluate our frameworks across a range of team sizes and assignment formulations against centralized EG, Hungarian, and Min-Max Distance baselines. Both algorithms preserve the fairness-efficiency balance of the Eisenberg-Gale equilibrium under partial observability. EG-MARL achieves near-centralized coordination and reduced travel distances, while the stochastic online mechanism enables real-time allocation with competitive fairness. Together, these results demonstrate that spatially aware EG formulations can effectively guide decentralized coordination in agents with heterogeneous capabilities.
[931] Tool-RoCo: An Agent-as-Tool Self-organization Large Language Model Benchmark in Multi-robot Cooperation
Ke Zhang, Xiaoning Zhao, Ce Zheng, Jiahong Ning, Dandan Zhu, Wenqi Zhang, Chen Sun, Toshiharu Sugawara
Main category: cs.MA
TL;DR: Tool-RoCo is a benchmark for evaluating LLMs in long-term multi-agent cooperation using tool usage as a mechanism for agent interaction and self-organization.
Details
Motivation: Existing LLM-based multi-agent systems rely on predefined orchestration and ignore agent autonomy. There's a need for benchmarks that evaluate how LLMs can autonomously cooperate and self-organize in multi-agent scenarios.Method: Tool-RoCo treats other agents as tools and introduces cooperative tools. It proposes four LLM paradigms with different autonomy levels: centralized cooperation, centralized self-organization, decentralized cooperation, and self-organization. The benchmark includes three multi-robot tasks (SORT, PACK, CABINET) to measure format/parameter accuracy and agent coordination through tool usage.
Result: Cooperative tools accounted for only 7.09% of all tools, indicating LLM-based agents rarely invoked others as assistants. Activation tools accounted for 96.42%, suggesting current LLMs tend to maintain active agents while seldom deactivating them for adaptive coordination.
Conclusion: Tool-RoCo provides a systematic benchmark to evaluate LLM autonomy and cooperation in multi-agent tasks, revealing that current LLMs have limited ability to use cooperative tools and adaptively coordinate agent activation/deactivation.
Abstract: This study proposes Tool-RoCo, a novel benchmark for evaluating large language models (LLMs) in long-term multi-agent cooperation based on RoCo, a multi-robot cooperative benchmark. Recent research on LLM-based multi-agent systems has relied on predefined orchestration, while ignoring agent autonomy. Tool-RoCo treats other agents as tools and introduces cooperative tools, leveraging tool usage to evaluate multi-agent cooperation and self-organization. Tool usage means that each agent (LLM) selects a tool from a candidate set based on the current state, receives feedback, and adjusts its selection in subsequent rounds. To evaluate different autonomy levels, we propose four LLM paradigms: (1) centralized cooperation, where a single LLM allocates tools to all agents; (2) centralized self-organization, where a central LLM autonomously activates agents while keeping others inactive; (3) decentralized cooperation, where each agent has its own LLM and calls tools based on local information; and (4) self-organization, where a randomly chosen initial agent can request collaboration, activating additional agents via tool calls. Tool-RoCo includes three multi-robot tasks, SORT, PACK, and CABINET, to measure format and parameter accuracy and agent coordination through tool usage. The results using several LLMs showed that cooperative tools accounted for only 7.09% of all tools, indicating that LLM-based agents rarely invoked others as assistants. Moreover, activation tools accounted for 96.42%, suggesting that current LLMs tend to maintain active agents while seldom deactivating them for adaptive coordination. Tool-RoCo provides a systematic benchmark to evaluate LLM autonomy and cooperation in multi-agent tasks. Code and Demo: https://github.com/ColaZhang22/Tool-Roco
cs.MM
[932] Audio-Visual World Models: Towards Multisensory Imagination in Sight and Sound
Jiahua Wang, Shannan Yan, Leqi Zheng, Jialong Wu, Yaoxin Mao
Main category: cs.MM
TL;DR: First formal framework for Audio-Visual World Models (AVWM) that jointly captures binaural spatial audio and visual dynamics with action control and reward prediction, validated through novel dataset and transformer architecture.
Details
Motivation: Real-world perception involves multiple sensory modalities, with audio providing crucial spatial/temporal cues. Existing world models focus primarily on visual observations, leaving audio integration largely unexplored. No prior work has formally defined audio-visual world models or how to jointly capture binaural spatial audio and visual dynamics with precise action control and task reward prediction.Method: 1) Formal framework for AVWM as partially observable Markov decision process with synchronized audio-visual observations, actions, and rewards. 2) Constructed AVW-4k dataset with 30 hours of binaural audio-visual trajectories across 76 indoor environments. 3) Proposed AV-CDiT (Audio-Visual Conditional Diffusion Transformer) with novel modality expert architecture balancing visual/auditory learning, using three-stage training strategy for multimodal integration.
Result: AV-CDiT achieves high-fidelity multimodal prediction across visual and auditory modalities with reward. Validated in continuous audio-visual navigation tasks where AVWM significantly enhances agent performance.
Conclusion: This work presents the first formal framework for Audio-Visual World Models, addressing the gap in multimodal environment simulation. The proposed approach successfully integrates audio and visual modalities for improved world modeling and demonstrates practical utility in navigation tasks.
Abstract: World models simulate environmental dynamics to enable agents to plan and reason about future states. While existing approaches have primarily focused on visual observations, real-world perception inherently involves multiple sensory modalities. Audio provides crucial spatial and temporal cues such as sound source localization and acoustic scene properties, yet its integration into world models remains largely unexplored. No prior work has formally defined what constitutes an audio-visual world model or how to jointly capture binaural spatial audio and visual dynamics under precise action control with task reward prediction. This work presents the first formal framework for Audio-Visual World Models (AVWM), formulating multimodal environment simulation as a partially observable Markov decision process with synchronized audio-visual observations, fine-grained actions, and task rewards. To address the lack of suitable training data, we construct AVW-4k, a dataset comprising 30 hours of binaural audio-visual trajectories with action annotations and reward signals across 76 indoor environments. We propose AV-CDiT, an Audio-Visual Conditional Diffusion Transformer with a novel modality expert architecture that balances visual and auditory learning, optimized through a three-stage training strategy for effective multimodal integration. Extensive experiments demonstrate that AV-CDiT achieves high-fidelity multimodal prediction across visual and auditory modalities with reward. Furthermore, we validate its practical utility in continuous audio-visual navigation tasks, where AVWM significantly enhances the agent’s performance.
[933] Augmenting Intra-Modal Understanding in MLLMs for Robust Multimodal Keyphrase Generation
Jiajun Cao, Qinggang Zhang, Yunbo Tang, Zhishang Xiang, Chang Yang, Jinsong Su
Main category: cs.MM
TL;DR: AimKP is a novel framework that enhances multimodal keyphrase generation by reinforcing intra-modal semantic learning in MLLMs while preserving cross-modal alignment, addressing modality bias and fine-grained feature extraction issues.
Details
Motivation: Multimodal keyphrase generation faces challenges with modality bias and fine-grained intra-modal feature extraction in MLLMs, leading to lack of robustness in real-world scenarios with noisy, incomplete, or misaligned multimedia data.Method: AimKP incorporates two core innovations: (1) Progressive Modality Masking that forces fine-grained feature extraction from corrupted inputs by progressively masking modality information during training, and (2) Gradient-based Filtering that identifies and discards noisy samples to prevent corruption of core cross-modal learning.
Result: Extensive experiments validate AimKP’s effectiveness in multimodal keyphrase generation and its robustness across different scenarios.
Conclusion: AimKP successfully addresses the limitations of existing MLLMs in multimodal keyphrase generation by explicitly reinforcing intra-modal semantic learning while maintaining cross-modal alignment, improving performance and robustness in real-world applications.
Abstract: Multimodal keyphrase generation (MKP) aims to extract a concise set of keyphrases that capture the essential meaning of paired image-text inputs, enabling structured understanding, indexing, and retrieval of multimedia data across the web and social platforms. Success in this task demands effectively bridging the semantic gap between heterogeneous modalities. While multimodal large language models (MLLMs) achieve superior cross-modal understanding by leveraging massive pretraining on image-text corpora, we observe that they often struggle with modality bias and fine-grained intra-modal feature extraction. This oversight leads to a lack of robustness in real-world scenarios where multimedia data is noisy, along with incomplete or misaligned modalities. To address this problem, we propose AimKP, a novel framework that explicitly reinforces intra-modal semantic learning in MLLMs while preserving cross-modal alignment. AimKP incorporates two core innovations: (i) Progressive Modality Masking, which forces fine-grained feature extraction from corrupted inputs by progressively masking modality information during training; (ii) Gradient-based Filtering, that identifies and discards noisy samples, preventing them from corrupting the model’s core cross-modal learning. Extensive experiments validate AimKP’s effectiveness in multimodal keyphrase generation and its robustness across different scenarios.
[934] ZO-ASR: Zeroth-Order Fine-Tuning of Speech Foundation Models without Back-Propagation
Yuezhang Peng, Yuxin Liu, Yao Li, Sheng Wang, Fei Wen, Xie Chen
Main category: cs.MM
TL;DR: ZO-ASR is a memory-efficient zeroth-order method for fine-tuning speech recognition models without backpropagation, using only inference memory by estimating gradients through forward passes.
Details
Motivation: Fine-tuning pre-trained speech foundation models for ASR requires substantial GPU memory due to backpropagation and activation storage. There's a need for memory-efficient methods that can work in resource-constrained environments or where gradients are inaccessible.Method: ZO-ASR uses zeroth-order optimization that avoids backpropagation by estimating gradients through forward passes only. It employs a multiple query mechanism for robustness and can be combined with SGD optimizer (ZO-ASR-SGD) to fine-tune models using only inference memory.
Result: For supervised domain adaptation on Whisper-Large-V3, ZO-ASR achieves up to 18.9% relative WER reduction over zero-shot baselines, outperforming existing ZO methods. For unsupervised test-time adaptation on Wav2Vec2-Base, it shows moderately lower performance than first-order Adam optimizer.
Conclusion: ZO-ASR provides a viable BP-free solution for fine-tuning ASR models in computationally resource-constrained or gradient-inaccessible scenarios, offering memory efficiency at the cost of some performance compared to first-order methods.
Abstract: Fine-tuning pre-trained speech foundation models for Automatic Speech Recognition (ASR) is prevalent, yet constrained by substantial GPU memory requirements. We introduce ZO-ASR, a memory-efficient Zeroth-Order (ZO) method that avoids Back-Propagation (BP) and activation memory by estimating gradients via forward passes. When combined with SGD optimizer, ZO-ASR-SGD fine-tunes ASR models using only inference memory. Our evaluation spans supervised and unsupervised tasks. For Supervised Domain Adaptation on Whisper-Large-V3, ZO-ASR’s multiple query mechanism enhances robustness and achieves up to an 18.9% relative Word Error Rate reduction over zero-shot baselines, outperforming existing ZO methods. For unsupervised Test-Time Adaptation on Wav2Vec2-Base, ZO-ASR exhibits moderately lower performance compared to first-order optimizer Adam. Our BP-free approach provides a viable solution for fine-tuning ASR models in computationally resource-constrained or gradient-inaccessible scenarios.
[935] PSA-MF: Personality-Sentiment Aligned Multi-Level Fusion for Multimodal Sentiment Analysis
Heng Xie, Kang Zhu, Zhengqi Wen, Jianhua Tao, Xuefei Liu, Ruibo Fu, Changsheng Li
Main category: cs.MM
TL;DR: A personality-sentiment aligned multi-level fusion framework for multimodal sentiment analysis that incorporates personality traits during feature extraction and uses multi-level fusion strategies to improve sentiment recognition.
Details
Motivation: Existing multimodal sentiment analysis methods have two key limitations: 1) they extract only shallow information from unimodal features, ignoring sentimental differences across different personalities, and 2) they directly merge features from different modalities without considering feature-level differences, which ultimately affects recognition performance.Method: Proposes a personality-sentiment aligned multi-level fusion framework with two main innovations: 1) Introduces personality traits during feature extraction phase with a novel personality-sentiment alignment method to obtain personalized sentiment embeddings from textual modality, and 2) Uses a novel multi-level fusion method that gradually integrates sentimental information through multimodal pre-fusion and multi-level enhanced fusion strategy across textual, visual, and audio modalities.
Result: The method was evaluated through multiple experiments on two commonly used datasets and achieved state-of-the-art results.
Conclusion: The proposed framework effectively addresses the limitations of existing multimodal sentiment analysis methods by incorporating personality traits and using multi-level fusion strategies, leading to improved sentiment recognition performance.
Abstract: Multimodal sentiment analysis (MSA) is a research field that recognizes human sentiments by combining textual, visual, and audio modalities. The main challenge lies in integrating sentiment-related information from different modalities, which typically arises during the unimodal feature extraction phase and the multimodal feature fusion phase. Existing methods extract only shallow information from unimodal features during the extraction phase, neglecting sentimental differences across different personalities. During the fusion phase, they directly merge the feature information from each modality without considering differences at the feature level. This ultimately affects the model’s recognition performance. To address this problem, we propose a personality-sentiment aligned multi-level fusion framework. We introduce personality traits during the feature extraction phase and propose a novel personality-sentiment alignment method to obtain personalized sentiment embeddings from the textual modality for the first time. In the fusion phase, we introduce a novel multi-level fusion method. This method gradually integrates sentimental information from textual, visual, and audio modalities through multimodal pre-fusion and a multi-level enhanced fusion strategy. Our method has been evaluated through multiple experiments on two commonly used datasets, achieving state-of-the-art results.
[936] An Efficient Recommendation System in E-commerce using Passer learning optimization based on Bi-LSTM
Hemn Barzan Abdalla, Awder Ahmed, Bahtiyar Mehmed, Mehdi Gheisari, Maryam Cheraghy, Yang Liu
Main category: cs.MM
TL;DR: A novel e-commerce recommendation system using Passer Learning Optimization-enhanced Bi-LSTM classifier achieves state-of-the-art performance with low MSE (1.24%) and high F1 scores (up to 92.51%) across multiple datasets.
Details
Motivation: Online reviews are crucial for consumer decisions but vary in quality, containing misleading or unhelpful information like ads, fake content, or irrelevant details, posing challenges for recommendation systems that rely on user-generated reviews.Method: Proposes a recommendation system based on Passer Learning Optimization-enhanced Bi-LSTM classifier, utilizing advanced graph embedding for effective knowledge extraction and fine-tuning of classifier parameters.
Result: Achieves 1.24% MSE on baby dataset, 88.58% improvement, with robust performance on digital music (88.46% F1) and patio lawn garden (92.51% F1) datasets, outperforming state-of-the-art models.
Conclusion: The proposed model demonstrates suitability for various e-commerce environments with improved accuracy and efficiency, effectively addressing challenges posed by unreliable online reviews in recommendation systems.
Abstract: Online reviews play a crucial role in shaping consumer decisions, especially in the context of e-commerce. However, the quality and reliability of these reviews can vary significantly. Some reviews contain misleading or unhelpful information, such as advertisements, fake content, or irrelevant details. These issues pose significant challenges for recommendation systems, which rely on user-generated reviews to provide personalized suggestions. This article introduces a recommendation system based on Passer Learning Optimization-enhanced Bi-LSTM classifier applicable to e-commerce recommendation systems with improved accuracy and efficiency compared to state-of-the-art models. It achieves as low as 1.24% MSE on the baby dataset. This lifts it as high as 88.58%. Besides, there is also robust performance of the system on digital music and patio lawn garden datasets at F1 of 88.46% and 92.51%, correspondingly. These results, made possible by advanced graph embedding for effective knowledge extraction and fine-tuning of classifier parameters, establish the suitability of the proposed model in various e-commerce environments.
[937] RDTF: Resource-efficient Dual-mask Training Framework for Multi-frame Animated Sticker Generation
Zhiqiang Yuan, Ting Zhang, Peixiang Luo, Ying Deng, Jiapei Zhang, Zexi Jia, Jinchao Zhang, Jie Zhou
Main category: cs.MM
TL;DR: RDTF is a resource-efficient framework for multi-frame animated sticker generation that trains compact models from scratch instead of using PEFT methods, achieving better performance with three core designs and a new million-level dataset.
Details
Motivation: Current video generation methods struggle with resource-constrained downstream tasks like animated sticker generation (ASG) which has low frame rates, abstract semantics, and long tail frame distributions. PEFT techniques suffer from insufficient fitting ability and source-domain knowledge interference.Method: Proposes RDTF with three core designs: 1) Discrete Frame Generation Network (DFGN) optimized for low-frame-rate ASG, 2) dual-mask data utilization strategy to enhance limited data availability and diversity, 3) difficulty-adaptive curriculum learning that decomposes sample entropy into static and adaptive components. Also constructs VSD2M dataset with million-level multi-modal animated stickers.
Result: RDTF demonstrates quantitative and qualitative superiority over state-of-the-art PEFT methods (I2V-Adapter, SimDA) on ASG tasks, verifying framework feasibility under resource constraints.
Conclusion: Training compact models from scratch with million-level samples outperforms PEFT on large models for resource-constrained ASG tasks, with RDTF providing an effective framework through specialized network design, data utilization strategies, and curriculum learning.
Abstract: Recently, significant advancements have been achieved in video generation technology, but applying it to resource-constrained downstream tasks like multi-frame animated sticker generation (ASG) characterized by low frame rates, abstract semantics, and long tail frame length distribution-remains challenging. Parameter-efficient fine-tuning (PEFT) techniques (e.g., Adapter, LoRA) for large pre-trained models suffer from insufficient fitting ability and source-domain knowledge interference. In this paper, we propose Resource-Efficient Dual-Mask Training Framework (RDTF), a dedicated solution for multi-frame ASG task under resource constraints. We argue that training a compact model from scratch with million-level samples outperforms PEFT on large models, with RDTF realizing this via three core designs: 1) a Discrete Frame Generation Network (DFGN) optimized for low-frame-rate ASG, ensuring parameter efficiency; 2) a dual-mask based data utilization strategy to enhance the availability and diversity of limited data; 3) a difficulty-adaptive curriculum learning method that decomposes sample entropy into static and adaptive components, enabling easy-to-difficult training convergence. To provide high-quality data support for RDTFs training from scratch, we construct VSD2M-a million-level multi-modal animated sticker dataset with rich annotations (static and animated stickers, action-focused text descriptions)-filling the gap of dedicated animated data for ASG task. Experiments demonstrate that RDTF is quantitatively and qualitatively superior to state-of-the-art PEFT methods (e.g., I2V-Adapter, SimDA) on ASG tasks, verifying the feasibility of our framework under resource constraints.
eess.AS
[938] Beyond Performance: Probing Representation Dynamics In Speech Enhancement Models
Yair Amar, Amir Ivry, Israel Cohen
Main category: eess.AS
TL;DR: The paper analyzes how noise levels affect internal representations in a speech enhancement model, showing encoder robustness vs latent/decoder sensitivity to SNR changes.
Details
Motivation: To understand how speech enhancement models internally process different noise conditions and how noise levels differentially activate model regions, which could inform better SNR-aware conditioning strategies.Method: Used MUSE transformer-convolutional model trained on VoiceBank DEMAND, analyzed activations across encoder, latent, decoder, and refinement blocks while sweeping SNRs from -10 to 30 dB. Applied Centered Kernel Alignment (CKA) for point-wise representation similarity and diffusion distance for distributional shifts.
Result: Encoder CKA between noisy and clean inputs remains stable, while latent and decoder CKA drop sharply with decreasing SNR. Linear fits show depth-dependent robustness-sensitivity trade-off. Diffusion distance varies incrementally with SNR within layers but differs strongly across layers, especially at low SNRs.
Conclusion: Noise levels differentially activate model regions and induce distinct inter-layer dynamics, motivating SNR-aware conditioning and refinement strategies for speech enhancement systems.
Abstract: We probe internal representations of a speech enhancement (SE) model across noise conditions. Using MUSE, a transformer-convolutional model trained on VoiceBank DEMAND, we analyze activations in encoder, latent, decoder, and refinement blocks while sweeping input signal-to-noise-ratios (SNRs) from -10 to 30 dB. We use Centered Kernel Alignment (CKA) to measure point-wise representation similarity and diffusion distance to capture distributional shifts across SNRs. Results show that the encoder CKA between noisy and clean inputs remains stable and latent and decoder CKA drop sharply as SNR decreases. Linear fits of CKA versus SNR reveal a depth-dependent robustness-sensitivity trade-off. The diffusion distance varies incrementally with SNR within each layer but differs strongly across layers, especially at low SNRs. Together, these findings indicate that noise levels differentially activate model regions and induce distinct inter-layer dynamics, motivating SNR-aware conditioning and refinement strategies for SE.
[939] A Low-Complexity Speech Codec Using Parametric Dithering for ASR
Ellison Murray, Morriel Kasher, Predrag Spasojevic
Main category: eess.AS
TL;DR: Dithering improves ASR input compression performance, with proposed parametric dithering showing 25-33.5% relative CER improvements at 1-3 bit resolutions while maintaining low complexity.
Details
Motivation: Dithering is known to improve perceptual quality in lossy compression, but its application to ASR input compression needs analytical and experimental justification to optimize speech recognition performance under compressed inputs.Method: Formalize understanding of optimal ASR performance under lossy input compression, then propose parametric dithering technique for low-complexity speech compression pipeline. The method is adaptable to meet performance targets or entropy constraints.
Result: 25% relative CER improvement at 1-bit resolution, 32.4% improvement at 2-bit, and 33.5% improvement at 3-bit resolution. Second dither choice yields reduced data rate. Codec is adaptable to performance targets or entropy constraints.
Conclusion: Dithering is analytically and experimentally justified for ASR input compression, with proposed parametric dithering technique showing significant performance improvements at low bit resolutions while maintaining low complexity and adaptability.
Abstract: Dithering is a technique commonly used to improve the perceptual quality of lossy data compression. In this work, we analytically and experimentally justify the use of dithering for ASR input compression. We formalize an understanding of optimal ASR performance under lossy input compression and leverage this to propose a parametric dithering technique for a low-complexity speech compression pipeline. The method performs well at 1-bit resolution, showing a 25% relative CER improvement, while also demonstrating improvements of 32.4% and 33.5% at 2- and 3-bit resolution, respectively, with our second dither choice yielding a reduced data rate. The proposed codec is adaptable to meet performance targets or stay within entropy constraints.
[940] Arabic TTS with FastPitch: Reproducible Baselines, Adversarial Training, and Oversmoothing Analysis
Lars Nippert
Main category: eess.AS
TL;DR: The paper presents reproducible Arabic TTS baselines using FastPitch architecture, introduces cepstral-domain metrics to analyze oversmoothing, addresses it with adversarial spectrogram loss, and explores multi-speaker TTS with synthetic voices.
Details
Motivation: Arabic TTS faces challenges due to limited resources and complex phonological patterns, and existing methods suffer from oversmoothing issues in mel-spectrogram prediction.Method: Built on FastPitch architecture, introduced cepstral-domain metrics to analyze oversmoothing, incorporated lightweight adversarial spectrogram loss to reduce oversmoothing, and explored multi-speaker TTS by augmenting FastPitch with XTTSv2 synthetic voices.
Result: The adversarial spectrogram loss trains stably and substantially reduces oversmoothing. Multi-speaker Arabic TTS with synthetic voices improves prosodic diversity without loss of stability.
Conclusion: The paper provides reproducible Arabic TTS baselines with improved quality through adversarial training and multi-speaker augmentation, making code, models, and recipes publicly available.
Abstract: Arabic text-to-speech (TTS) remains challenging due to limited resources and complex phonological patterns. We present reproducible baselines for Arabic TTS built on the FastPitch architecture and introduce cepstral-domain metrics for analyzing oversmoothing in mel-spectrogram prediction. While traditional Lp reconstruction losses yield smooth but over-averaged outputs, the proposed metrics reveal their temporal and spectral effects throughout training. To address this, we incorporate a lightweight adversarial spectrogram loss, which trains stably and substantially reduces oversmoothing. We further explore multi-speaker Arabic TTS by augmenting FastPitch with synthetic voices generated using XTTSv2, resulting in improved prosodic diversity without loss of stability. The code, pretrained models, and training recipes are publicly available at: https://github.com/nipponjo/tts-arabic-pytorch.
[941] Identifiability Conditions for Acoustic Feedback Cancellation with the Two-Channel Adaptive Feedback Canceller Algorithm
Arnout Roebben, Toon van Waterschoot, Jan Wouters, Marc Moonen
Main category: eess.AS
TL;DR: The paper analyzes identifiability conditions for acoustic feedback cancellation, showing that for the 2ch-AFC algorithm, identifiability can be achieved when the forward path feedforward filter order exceeds the AR model order, and proposes using the correlation matrix condition number to monitor identifiability.
Details
Motivation: In audio systems with microphone-loudspeaker setups, acoustic feedback creates instability. While PEM-based feedback cancellation algorithms exist, their identifiability conditions need to be better understood and generalized for practical applications.Method: The paper analyzes the two-channel adaptive feedback canceller (2ch-AFC) algorithm, generalizing the delay-based identifiability condition to an invertibility-based condition. It examines the relationship between forward path feedforward filter order and AR model order, and proposes using the condition number of the correlation matrix inversion as an identifiability monitoring measure.
Result: The paper demonstrates that identifiability can be achieved when the order of the forward path feedforward filter exceeds the order of the AR model. The condition number of the correlation matrix inversion serves as an effective measure for monitoring identifiability in the 2ch-AFC algorithm.
Conclusion: The invertibility-based condition generalizes previous delay-based conditions for feedback path identifiability, providing a more comprehensive understanding of when the 2ch-AFC algorithm can successfully identify acoustic feedback paths, with practical monitoring through correlation matrix condition number analysis.
Abstract: In audio signal processing applications with a microphone and a loudspeaker within the same acoustic environment, the loudspeaker signals can feed back into the microphone, thereby creating a closed-loop system that potentially leads to system instability. To remove this acoustic coupling, prediction error method (PEM) feedback cancellation algorithms aim to identify the feedback path between the loudspeaker and the microphone by assuming that the input signal can be modelled by means of an autoregressive (AR) model. It has previously been shown that this PEM framework and resulting algorithms can identify the feedback path correctly in cases where the forward path from microphone to loudspeaker is sufficiently time-varying or non-linear, or when the forward path delay equals or exceeds the order of the AR model. In this paper, it is shown that this delay-based condition can be generalised for one particular PEM-based algorithm, the so-called two-channel adaptive feedback canceller (2ch-AFC), to an invertibility-based condition, for which it is shown that identifiability can be achieved when the order of the forward path feedforward filter exceeds the order of the AR model. Additionally, the condition number of inversion of the correlation matrix as used in the 2ch-AFC algorithm can serve as a measure for monitoring the identifiability.
[942] RIFT: Entropy-Optimised Fractional Wavelet Constellations for Ideal Time-Frequency Estimation
James M. Cozens, Simon J. Godsill
Main category: eess.AS
TL;DR: RIFT method estimates Ideal Time-Frequency Representation using fractional wavelet transforms, entropy-based sparsity, and deconvolution to achieve WVD-like resolution without cross-terms.
Details
Motivation: Need for high-resolution time-frequency analysis of complex nonstationary signals that achieves Wigner-Ville Distribution's auto-term resolution while suppressing cross-terms, which has been a long-standing challenge in signal processing.Method: 1) Compute constellation of Continuous Fractional Wavelet Transforms aligned to different local time-frequency curvatures; 2) Combine via localized entropy-based sparsity measure to resolve auto-terms and attenuate cross-terms; 3) Apply positivity-constrained Lucy-Richardson deconvolution with total-variation regularization to estimate ITFR; 4) Derive Cohen’s class convolutional kernels for CFWT constellations; 5) Extract Instantaneous Phase Direction field for curvature visualization and Kalman tracking.
Result: Achieves auto-term resolution comparable to Wigner-Ville Distribution while effectively suppressing cross-terms; yields high-resolution RIFT representation and enables extraction of signal component trajectories via Spline-RIFT variant; demonstrates superior time-frequency precision on synthetic and real-world signals.
Conclusion: RIFT provides a significant advance in time-frequency analysis, offering cross-term-free high-resolution representations with potential applications across various fields requiring precise time-frequency analysis of complex signals.
Abstract: We introduce a new method for estimating the Ideal Time-Frequency Representation (ITFR) of complex nonstationary signals. The Reconstructive Ideal Fractional Transform (RIFT) computes a constellation of Continuous Fractional Wavelet Transforms (CFWTs) aligned to different local time-frequency curvatures. This constellation is combined into a single optimised time-frequency energy representation via a localised entropy-based sparsity measure, designed to resolve auto-terms and attenuate cross-terms. Finally, a positivity-constrained Lucy-Richardson deconvolution with total-variation regularisation is applied to estimate the ITFR, achieving auto-term resolution comparable to that of the Wigner-Ville Distribution (WVD), yielding the high-resolution RIFT representation. The required Cohen’s class convolutional kernels are fully derived in the paper for the chosen CFWT constellations. Additionally, the optimisation yields an Instantaneous Phase Direction (IPD) field, which allows the localised curvature in speech or music extracts to be visualised and utilised within a Kalman tracking scheme, enabling the extraction of signal component trajectories and the construction of the Spline-RIFT variant. Evaluation on synthetic and real-world signals demonstrates the algorithm’s ability to effectively suppress cross-terms and achieve superior time-frequency precision relative to competing methods. This advance holds significant potential for a wide range of applications requiring high-resolution cross-term-free time-frequency analysis.
[943] Safeguarding Privacy in Edge Speech Understanding with Tiny Foundation Models
Afsara Benazir, Felix Xiaozhu Lin
Main category: eess.AS
TL;DR: SpeechShield: Edge/cloud privacy-preserving speech inference engine that filters sensitive entities on-device using tiny speech foundation models without compromising transcription accuracy.
Details
Motivation: Current speech recognition systems rely on cloud providers, risking exposure of sensitive content. Need privacy-preserving solution that maintains transcription accuracy while filtering sensitive entities on resource-constrained devices.Method: Uses timestamp-based on-device masking with token-to-entity prediction model to filter sensitive entities. Masked input sent to trusted cloud/local hub. Recovery uses confidence score to choose best prediction between cloud and on-device models.
Result: Achieves SOTA transcription performance while filtering ~83% of private entities on-device. 16x smaller memory, 3.3x faster, 17x more compute efficient than prior frameworks. 38.8-77.5% relative reduction in WER compared to existing offline services.
Conclusion: SpeechShield enables robust speech recognition with privacy preservation on resource-constrained devices, demonstrating practical deployment on Raspberry Pi 4B with efficient performance.
Abstract: Robust speech recognition systems rely on cloud service providers for inference. It needs to ensure that an untrustworthy provider cannot deduce the sensitive content in speech. Sanitization can be done on speech content keeping in mind that it has to avoid compromising transcription accuracy. Realizing the under utilized capabilities of tiny speech foundation models (FMs), for the first time, we propose a novel use: enhancing speech privacy on resource-constrained devices. We introduce SpeechShield, an edge/cloud privacy preserving speech inference engine that can filter sensitive entities without compromising transcript accuracy. We utilize a timestamp based on-device masking approach that utilizes a token to entity prediction model to filter sensitive entities. Our choice of mask strategically conceals parts of the input and hides sensitive data. The masked input is sent to a trusted cloud service or to a local hub to generate the masked output. The effectiveness of SpeechShield hinges on how well the entity time segments are masked. Our recovery is a confidence score based approach that chooses the best prediction between cloud and on-device model. We implement SpeechShield on a 64 bit Raspberry Pi 4B. Experiments show that our solution leads to robust speech recognition without forsaking privacy. SpeechShield with < 100 MB memory, achieves state-of-the-art (SOTA) speech transcription performance while filtering about 83% of private entities directly on-device. SpeechShield is 16x smaller in memory, 3.3x faster and 17x more compute efficient than prior privacy preserving speech frameworks and has a relative reduction in word error rate (WER) by 38.8-77.5% when compared to existing offline transcription services.
[944] Discrete Optimal Transport and Voice Conversion
Anton Selitskiy, Maitreya Kocharekar
Main category: eess.AS
TL;DR: Voice conversion using vector-based interface with discrete optimal transport mapping for speaker alignment, achieving high quality and showing that this approach can make synthetic audio indistinguishable from real.
Details
Motivation: To improve voice conversion by creating a method that can effectively align audio embeddings between different speakers while maintaining high audio quality.Method: Uses a vector-based interface for voice conversion and employs discrete optimal transport mapping to align audio embeddings between speakers.
Result: The method demonstrates high quality and effectiveness in voice conversion. Additionally, applying discrete optimal transport as post-processing can cause synthetic audio to be incorrectly classified as real.
Conclusion: Discrete optimal transport mapping is an effective approach for voice conversion that not only produces high-quality results but can also make synthetic audio indistinguishable from real audio, raising potential concerns about audio authenticity.
Abstract: In this work, we address the voice conversion (VC) task using a vector-based interface. To align audio embeddings between speakers, we employ discrete optimal transport mapping. Our evaluation results demonstrate the high quality and effectiveness of this method. Additionally, we show that applying discrete optimal transport as a post-processing step in audio generation can lead to the incorrect classification of synthetic audio as real.
[945] The Extended SONICOM HRTF Dataset and Spatial Audio Metrics Toolbox
Katarina C. Poole, Julie Meyer, Vincent Martin, Rapolas Daugintis, Nils Marggraf-Turley, Jack Webb, Ludovic Pirard, Nicola La Magna, Oliver Turvey, Lorenzo Picinali
Main category: eess.AS
TL;DR: Extended SONICOM HRTF dataset now includes 300 subjects with demographic data, synthesized HRTFs for 200 subjects, and optimized 3D scans, plus a new Python toolbox for HRTF analysis.
Details
Motivation: HRTFs are personalized based on individual anatomy, but existing datasets are limited in size and lack demographic context, hindering development of personalized spatial audio systems and machine learning approaches.Method: Expanded dataset from 300 measured subjects with demographic data, used Mesh2HRTF to synthesize HRTFs for 200 subjects, pre-processed 3D head/ear scans optimized for HRTF synthesis, and developed Spatial Audio Metrics (SAM) Python toolbox for analysis.
Result: Created comprehensive HRTF resource with 300 subjects (largest available), synthesized HRTFs for 200 subjects, optimized 3D scans enabling morphological studies, and open-source analysis toolbox, facilitating rapid HRTF synthesis algorithm development and ML research.
Conclusion: The extended SONICOM dataset and SAM toolbox provide a complete ecosystem for advancing personalized spatial audio research, enabling better understanding of anatomy-HRTF relationships and accelerating development of individualized audio systems.
Abstract: Headphone-based spatial audio uses head-related transfer functions (HRTFs) to simulate real-world acoustic environments. HRTFs are unique to everyone, due to personal morphology, shaping how sound waves interact with the body before reaching the eardrums. Here we present the extended SONICOM HRTF dataset which expands on the previous version released in 2023. The total number of measured subjects has now been increased to 300, with demographic information for a subset of the participants, providing context for the dataset’s population and relevance. The dataset incorporates synthesised HRTFs for 200 of the 300 subjects, generated using Mesh2HRTF, alongside pre-processed 3D scans of the head and ears, optimised for HRTF synthesis. This rich dataset facilitates rapid and iterative optimisation of HRTF synthesis algorithms, allowing the automatic generation of large data. The optimised scans enable seamless morphological modifications, providing insights into how anatomical changes impact HRTFs, and the larger sample size enhances the effectiveness of machine learning approaches. To support analysis, we also introduce the Spatial Audio Metrics (SAM) Toolbox, a Python package designed for efficient analysis and visualisation of HRTF data, offering customisable tools for advanced research. Together, the extended dataset and toolbox offer a comprehensive resource for advancing personalised spatial audio research and development.
[946] AHAMask: Reliable Task Specification for Large Audio Language Models without Instructions
Yiwei Guo, Bohan Li, Hankun Wang, Zhihan Li, Shuai Wang, Xie Chen, Kai Yu
Main category: eess.AS
TL;DR: AHAMask improves audio language models by masking specific attention heads to trigger acoustic tasks without instructions, solving prompt sensitivity issues.
Details
Motivation: Current large audio language models (LALMs) suffer from prompt sensitivity - different instructions with the same intention yield drastically different outcomes, making them unreliable.Method: Propose AHAMask which masks specific attention heads in the decoder-only LLM backbone of LALMs. These masks are efficiently trained with parameters equal to attention head count, triggering acoustic task functionalities without instructions.
Result: Applying selective attention head masks achieves comparable or better performance than using instructions, both on single and composite tasks. This reveals LALMs have “functional pathways” in their attention heads.
Conclusion: AHAMask provides reliable acoustic task specification for LALMs while solving prompt sensitivity, and reveals interesting insights about functional organization in attention heads.
Abstract: Although current large audio language models (LALMs) extend text large language models (LLMs) with generic acoustic understanding abilities, they usually suffer from prompt sensitivity, where different instructions of the same intention can yield drastically different outcomes. In this work, we propose AHAMask, where we simply mask some of the attention heads in the decoder-only LLM backbone of LALMs, to trigger specific acoustic task functionalities without instructions. These masks are efficiently obtained by training on an LALM, with the number of trainable parameters equal to the attention head count in its LLM backbone. We show by experiments that applying such selective attention head masks achieves comparable or even better performance than using instructions, either on single or composite tasks. Besides achieving reliable acoustic task specification for LALMs, this also reveals that LALMs exhibit certain “functional pathways” in their attention heads.
[947] Semantic-VAE: Semantic-Alignment Latent Representation for Better Speech Synthesis
Zhikang Niu, Shujie Hu, Jeongsoo Choi, Yushen Chen, Peining Chen, Pengcheng Zhu, Yunting Yang, Bowen Zhang, Jian Zhao, Chunhui Wang, Xie Chen
Main category: eess.AS
TL;DR: Semantic-VAE improves zero-shot TTS by using semantic alignment regularization to overcome the trade-off between reconstruction quality and intelligibility in VAE latent spaces.
Details
Motivation: Current VAE-based latent representations for zero-shot TTS face a fundamental optimization dilemma: higher-dimensional spaces improve reconstruction quality and speaker similarity but degrade intelligibility, while lower-dimensional spaces improve intelligibility at the expense of reconstruction fidelity. Mel-spectrograms are inefficient due to redundancy.Method: Proposes Semantic-VAE, a novel VAE framework that utilizes semantic alignment regularization in the latent space to capture semantic structure in high-dimensional latent representations, alleviating the reconstruction-generation trade-off.
Result: When integrated into F5-TTS, achieves 2.10% WER and 0.64 speaker similarity on LibriSpeech-PC, outperforming mel-based systems (2.23%, 0.60) and vanilla acoustic VAE baselines (2.65%, 0.59). Shows significant improvements in synthesis quality and training efficiency.
Conclusion: Semantic-VAE effectively overcomes the reconstruction-intelligibility trade-off in VAE-based TTS systems, providing better performance than both mel-spectrogram and vanilla VAE approaches while improving training efficiency.
Abstract: While mel-spectrograms have been widely utilized as intermediate representations in zero-shot text-to-speech (TTS), their inherent redundancy leads to inefficiency in learning text-speech alignment. Compact VAE-based latent representations have recently emerged as a stronger alternative, but they also face a fundamental optimization dilemma: higher-dimensional latent spaces improve reconstruction quality and speaker similarity, but degrade intelligibility, while lower-dimensional spaces improve intelligibility at the expense of reconstruction fidelity. To overcome this dilemma, we propose Semantic-VAE, a novel VAE framework that utilizes semantic alignment regularization in the latent space. This design alleviates the reconstruction-generation trade-off by capturing semantic structure in high-dimensional latent representations. Extensive experiments demonstrate that Semantic-VAE significantly improves synthesis quality and training efficiency. When integrated into F5-TTS, our method achieves 2.10% WER and 0.64 speaker similarity on LibriSpeech-PC, outperforming mel-based systems (2.23%, 0.60) and vanilla acoustic VAE baselines (2.65%, 0.59). We also release the code and models to facilitate further research.
[948] MeanVC: Lightweight and Streaming Zero-Shot Voice Conversion via Mean Flows
Guobin Ma, Jixun Yao, Ziqian Ning, Yuepeng Jiang, Lingxin Xiong, Lei Xie, Pengcheng Zhu
Main category: eess.AS
TL;DR: MeanVC is a lightweight streaming zero-shot voice conversion model that combines AR and NAR approaches using diffusion transformers with chunk-wise processing and mean flows for single-step inference.
Details
Motivation: There's a growing need for streaming voice conversion models that are fast, lightweight, and high-fidelity, but existing streaming methods either require large parameters for good performance or struggle with unseen speakers.Method: Proposes MeanVC with: 1) diffusion transformer with chunk-wise autoregressive denoising for streaming, 2) mean flows that regress average velocity field for single-step inference, and 3) diffusion adversarial post-training to reduce over-smoothing.
Result: Outperforms existing zero-shot streaming VC systems with superior conversion quality, higher efficiency, and significantly fewer parameters.
Conclusion: MeanVC successfully addresses the need for lightweight, streaming zero-shot voice conversion by combining AR/NAR strengths and enabling single-step inference through mean flows.
Abstract: Zero-shot voice conversion (VC) aims to transfer timbre from a source speaker to any unseen target speaker while preserving linguistic content. Growing application scenarios demand models with streaming inference capabilities. This has created a pressing need for models that are simultaneously fast, lightweight, and high-fidelity. However, existing streaming methods typically rely on either autoregressive (AR) or non-autoregressive (NAR) frameworks, which either require large parameter sizes to achieve strong performance or struggle to generalize to unseen speakers. In this study, we propose MeanVC, a lightweight and streaming zero-shot VC approach. MeanVC introduces a diffusion transformer with a chunk-wise autoregressive denoising strategy, combining the strengths of both AR and NAR paradigms for efficient streaming processing. By introducing mean flows, MeanVC regresses the average velocity field during training, enabling zero-shot VC with superior speech quality and speaker similarity in a single sampling step by directly mapping from the start to the endpoint of the flow trajectory. Additionally, we incorporate diffusion adversarial post-training to mitigate over-smoothing and further enhance speech quality. Experimental results demonstrate that MeanVC significantly outperforms existing zero-shot streaming VC systems, achieving superior conversion quality with higher efficiency and significantly fewer parameters. Audio demos and code are publicly available at https://aslp-lab.github.io/MeanVC.
eess.IV
[949] Comparative Evaluation of Generative AI Models for Chest Radiograph Report Generation in the Emergency Department
Woo Hyeon Lim, Ji Young Lee, Jong Hyuk Lee, Saehoon Kim, Hyungjin Kim
Main category: eess.IV
TL;DR: Medical VLMs for CXR report generation show variable performance; AIRead outperforms other models and approaches radiologist quality in some metrics, while other VLMs have higher error rates.
Details
Motivation: To benchmark open-source and commercial medical image-specific Vision-Language Models (VLMs) against real-world radiologist-written reports to evaluate their clinical utility and performance.Method: Retrospective study of 478 emergency department patients with same-day CXR and CT. Reports from five VLMs (AIRead, Lingshu, MAIRA-2, MedGemma, MedVersa) and radiologists were blindly evaluated by three thoracic radiologists using RADPEER, clinical acceptability, hallucination, and language clarity criteria. Statistical analysis used generalized linear mixed models with radiologist reports as reference.
Result: AIRead performed best: lowest RADPEER 3b rate (5.3% vs 13.9%), highest clinical acceptability (84.5% vs 74.3%), minimal hallucinations comparable to radiologists (0.3% vs 0.1%). Other VLMs showed higher disagreement rates (16.8-43.0%), lower clinical acceptability (41.1-71.4%), and frequent hallucinations (5.4-17.4%). Language clarity was higher for AIRead, Lingshu, and MedVersa than radiologists. Sensitivity for common findings varied widely across models (2.4-86.7%).
Conclusion: Medical VLMs for CXR report generation exhibit variable performance in report quality and diagnostic measures, with AIRead demonstrating the best overall performance approaching radiologist quality, while other models show significant limitations.
Abstract: Purpose: To benchmark open-source or commercial medical image-specific VLMs against real-world radiologist-written reports. Methods: This retrospective study included adult patients who presented to the emergency department between January 2022 and April 2025 and underwent same-day CXR and CT for febrile or respiratory symptoms. Reports from five VLMs (AIRead, Lingshu, MAIRA-2, MedGemma, and MedVersa) and radiologist-written reports were randomly presented and blindly evaluated by three thoracic radiologists using four criteria: RADPEER, clinical acceptability, hallucination, and language clarity. Comparative performance was assessed using generalized linear mixed models, with radiologist-written reports treated as the reference. Finding-level analyses were also performed with CT as the reference. Results: A total of 478 patients (median age, 67 years [interquartile range, 50-78]; 282 men [59.0%]) were included. AIRead demonstrated the lowest RADPEER 3b rate (5.3% [76/1434] vs. radiologists 13.9% [200/1434]; P<.001), whereas other VLMs showed higher disagreement rates (16.8-43.0%; P<.05). Clinical acceptability was the highest with AIRead (84.5% [1212/1434] vs. radiologists 74.3% [1065/1434]; P<.001), while other VLMs performed worse (41.1-71.4%; P<.05). Hallucinations were rare with AIRead, comparable to radiologists (0.3% [4/1425]) vs. 0.1% [1/1425]; P=.21), but frequent with other models (5.4-17.4%; P<.05). Language clarity was higher with AIRead (82.9% [1189/1434]), Lingshu (88.0% [1262/1434]), and MedVersa (88.4% [1268/1434]) compared with radiologists (78.1% [1120/1434]; P<.05). Sensitivity varied substantially across VLMs for the common findings: AIRead, 15.5-86.7%; Lingshu, 2.4-86.7%; MAIRA-2, 6.0-72.0%; MedGemma, 4.8-76.7%; and MedVersa, 20.2-69.3%. Conclusion: Medical VLMs for CXR report generation exhibited variable performance in report quality and diagnostic measures.
[950] MedCondDiff: Lightweight, Robust, Semantically Guided Diffusion for Medical Image Segmentation
Ruirui Huang, Jiacheng Li
Main category: eess.IV
TL;DR: MedCondDiff is a lightweight diffusion framework for multi-organ medical image segmentation that uses semantic priors from a PVT backbone to guide the denoising process, improving efficiency and robustness.
Details
Motivation: To address the computational inefficiency and anatomical inconsistency of conventional diffusion models for medical image segmentation, particularly for multi-organ, multi-modality scenarios.Method: A diffusion-based framework that conditions the denoising process on semantic priors extracted by a Pyramid Vision Transformer (PVT) backbone, creating a semantically guided and lightweight architecture.
Result: The model demonstrates competitive performance across anatomical regions and imaging modalities while reducing inference time and VRAM usage compared to conventional diffusion models.
Conclusion: Semantically guided diffusion models represent an effective class of architectures for medical imaging tasks, offering improved efficiency and robustness for multi-organ segmentation.
Abstract: We introduce MedCondDiff, a diffusion-based framework for multi-organ medical image segmentation that is efficient and anatomically grounded. The model conditions the denoising process on semantic priors extracted by a Pyramid Vision Transformer (PVT) backbone, yielding a semantically guided and lightweight diffusion architecture. This design improves robustness while reducing both inference time and VRAM usage compared to conventional diffusion models. Experiments on multi-organ, multi-modality datasets demonstrate that MedCondDiff delivers competitive performance across anatomical regions and imaging modalities, underscoring the potential of semantically guided diffusion models as an effective class of architectures for medical imaging tasks.
[951] Large-field-of-view lensless imaging with miniaturized sensors
Yu Ren, Xiaoling Zhang, Xu Zhan, Xiangdong Ma, Yunqi Wang, Edmund Y. Lam, Tianjiao Zeng
Main category: eess.IV
TL;DR: A hierarchical local-to-global framework for lensless imaging that addresses PSF spatial variation and sensor truncation issues, enabling better reconstruction quality with smaller sensors.
Details
Motivation: Existing lensless imaging methods rely on idealized models assuming globally shift-invariant PSF and large sensors, but real-world PSF varies spatially across FOV and finite sensor boundaries truncate modulated light, degrading peripheral reconstruction quality and limiting effective FOV, especially as sensors shrink.Method: Proposes a local-to-global hierarchical framework using a locally shift-invariant convolution model that accounts for PSF variation and sensor truncation. First performs patch-wise learned deconvolution to adaptively estimate local PSFs and reconstruct regions independently, then uses a hierarchical enhancement network that progressively expands receptive field from small patches through intermediate blocks to full image, integrating local details with global context.
Result: Achieves superior reconstruction quality over larger effective FOV with significantly reduced sensor sizes. Under extreme miniaturization (sensors reduced to 8% of original area), achieves improvements of 2 dB PSNR and 5% SSIM, with notable gains in structural fidelity.
Conclusion: The proposed hierarchical local-to-global framework effectively addresses PSF spatial variation and sensor truncation challenges in lensless imaging, enabling high-quality reconstruction with smaller sensors and expanding practical applications of compact imaging systems.
Abstract: Lensless cameras replace bulky optics with thin modulation masks, enabling compact imaging systems. However, existing methods rely on an idealized model that assumes a globally shift-invariant point spread function (PSF) and sufficiently large sensors. In reality, the PSF varies spatially across the field of view (FOV), and finite sensor boundaries truncate modulated light–effects that intensify as sensors shrink, degrading peripheral reconstruction quality and limiting the effective FOV. We address these limitations through a local-to-global hierarchical framework grounded in a locally shift-invariant convolution model that explicitly accounts for PSF variation and sensor truncation. Patch-wise learned deconvolution first adaptively estimates local PSFs and reconstructs regions independently. A hierarchical enhancement network then progressively expands its receptive field–from small patches through intermediate blocks to the full image–integrating fine local details with global contextual information. Experiments on public datasets show that our method achieves superior reconstruction quality over a larger effective FOV with significantly reduced sensor sizes. Under extreme miniaturization–sensors reduced to 8% of the original area–we achieve improvements of 2 dB (PSNR) and 5% (SSIM), with particularly notable gains in structural fidelity. Code is available at https://github.com/KB504-public/l2g_lensless_imaging .
[952] Diffusion-Based Synthesis of 3D T1w MPRAGE Images from Multi-Echo GRE with Multi-Parametric MRI Integration
Sizhe Fang, Deqiang Qiu
Main category: eess.IV
TL;DR: Deep learning framework synthesizes T1w MPRAGE images from multi-echo GRE data using diffusion model with QSM and R2* priors, achieving biologically plausible results comparable to ground truth.
Details
Motivation: Current neuroimaging requires separate T1-weighted MPRAGE acquisitions for structural morphometry alongside mGRE sequences for quantitative parametric maps, leading to prolonged scan times. There's a need to streamline protocols by synthesizing structural images from existing quantitative data.Method: Developed a multi-parametric conditional diffusion model based on Fast-DDPM architecture that integrates iron-sensitive QSM and R2* maps as physical priors to address contrast ambiguity in iron-rich deep gray matter. Trained and validated on 175 healthy subjects, compared against U-Net and GAN baselines.
Result: Significantly outperformed baselines in perceptual quality and segmentation accuracy, especially in subcortical regions (thalamus, pallidum). Synthesized images preserved biological dependencies: high concordance in age-related atrophy rates, aging effect sizes, and sexual dimorphism patterns compared to ground truth.
Conclusion: The diffusion-based method generates biologically plausible T1w images suitable for clinical morphometric analysis, offering a promising pathway to reduce acquisition time by deriving structural contrasts retrospectively from quantitative mGRE sequences.
Abstract: Multi-echo Gradient Echo (mGRE) sequences provide valuable quantitative parametric maps, such as Quantitative Susceptibility Mapping (QSM) and transverse relaxation rate (R2*), sensitive to tissue iron and myelin. However, structural morphometry typically relies on separate T1-weighted MPRAGE acquisitions, prolonging scan times. We propose a deep learning framework to synthesize high-contrast 3D T1w MPRAGE images directly from mGRE data, streamlining neuroimaging protocols. We developed a novel multi-parametric conditional diffusion model based on the Fast-DDPM architecture. Unlike conventional intensity-based synthesis, our approach integrates iron-sensitive QSM and R2* maps as physical priors to address contrast ambiguity in iron-rich deep gray matter. We trained and validated the model on 175 healthy subjects. Performance was evaluated against established U-Net and GAN-based baselines using perceptual metrics and downstream segmentation accuracy. Uniquely, we assessed the biological plausibility of synthesized images by replicating population-level statistical associations with age and sex. The proposed framework significantly outperformed baselines, achieving superior perceptual quality and segmentation accuracy, particularly in subcortical regions like the thalamus and pallidum. Crucially, synthesized images preserved essential biological dependencies: regression analyses showed high concordance in age-related atrophy rates, aging effect sizes, and sexual dimorphism patterns compared to ground truth. By effectively leveraging quantitative MRI priors, our diffusion-based method generates strictly biologically plausible T1w images suitable for reliable clinical morphometric analysis. This approach offers a promising pathway to reduce acquisition time by deriving structural contrasts retrospectively from quantitative mGRE sequences.
[953] Disentangling Progress in Medical Image Registration: Beyond Trend-Driven Architectures towards Domain-Specific Strategies
Bailiang Jian, Jiazhen Pan, Rohit Jena, Morteza Ghahremani, Hongwei Bran Li, Daniel Rueckert, Christian Wachinger, Benedikt Wiestler
Main category: eess.IV
TL;DR: The paper finds that domain-specific registration designs outperform generic architectural trends in medical image registration, advocating for a shift in research focus toward domain-specific principles.
Details
Motivation: To determine whether future advances in medical image registration should focus on importing generic architectural trends (like Transformers, large-kernel CNNs) or refining domain-specific design principles, as current methods combine both but their relative contributions are unclear and entangled.Method: Developed a modular framework spanning brain, lung, cardiac, and abdominal registration to systematically disentangle the influence of generic architectural trends vs. domain-specific designs. Created a transparent, modular benchmark (rethink-reg) for plug-and-play comparison of architectures.
Result: Low-level “trend-driven” computational blocks offer only marginal or inconsistent gains, while high-level registration-specific designs consistently deliver more accurate, smoother, and more robust deformations. Domain priors significantly elevate U-Net baseline performance (~3% average relative improvement) far more than trend-driven blocks.
Conclusion: Research should shift emphasis from following architectural trends to embracing domain-specific design principles as the true drivers of progress in learning-based medical image registration. The modular benchmark provides a common ground for reproducible evaluation.
Abstract: Medical image registration drives quantitative analysis across organs, modalities, and patient populations. Recent deep learning methods often combine low-level “trend-driven” computational blocks from computer vision, such as large-kernel CNNs, Transformers, and state-space models, with high-level registration-specific designs like motion pyramids, correlation layers, and iterative refinement. Yet, their relative contributions remain unclear and entangled. This raises a central question: should future advances in registration focus on importing generic architectural trends or on refining domain-specific design principles? Through a modular framework spanning brain, lung, cardiac, and abdominal registration, we systematically disentangle the influence of these two paradigms. Our evaluation reveals that low-level “trend-driven” computational blocks offer only marginal or inconsistent gains, while high-level registration-specific designs consistently deliver more accurate, smoother, and more robust deformations. These domain priors significantly elevate the performance of a standard U-Net baseline, far more than variants incorporating “trend-driven” blocks, achieving an average relative improvement of $\sim3%$. All models and experiments are released within a transparent, modular benchmark that enables plug-and-play comparison for new architectures and registration tasks (https://github.com/BailiangJ/rethink-reg). This dynamic and extensible platform establishes a common ground for reproducible and fair evaluation, inviting the community to isolate genuine methodological contributions from domain priors. Our findings advocate a shift in research emphasis: from following architectural trends to embracing domain-specific design principles as the true drivers of progress in learning-based medical image registration.
[954] Beyond Subspace Isolation: Many-to-Many Transformer for Light Field Image Super-resolution
Zeke Zexi Hu, Xiaoming Chen, Vera Yuk Ying Chung, Yiran Shen
Main category: eess.IV
TL;DR: Proposes Many-to-Many Transformer (M2MT) for light field image super-resolution to overcome subspace isolation in existing methods, enabling comprehensive access to all spatial-angular information with better efficiency.
Details
Motivation: Existing LFSR methods using Transformers decompose 4D light field data into subspaces, restricting self-attention to limited subsets (One-to-One scheme), preventing comprehensive optimization of spatial and angular cues - identified as "subspace isolation".Method: Introduces Many-to-Many Transformer (M2MT) that aggregates angular information in spatial subspace before self-attention, enabling complete access to all sub-aperture images. Builds M2MT network for LFSR with favorable memory/computation efficiency.
Result: Achieves state-of-the-art performance across public datasets with better balance between performance and efficiency (lower memory/computation demands). Local attribution maps validate M2MT captures truly non-local context in both spatial and angular subspaces.
Conclusion: M2MT effectively mitigates subspace isolation by enabling Many-to-Many self-attention across all spatial-angular information, acquiring superior spatial-angular representation for LFSR with practical efficiency advantages.
Abstract: The effective extraction of spatial-angular features plays a crucial role in light field image super-resolution (LFSR) tasks, and the introduction of convolution and Transformers leads to significant improvement in this area. Nevertheless, due to the large 4D data volume of light field images, many existing methods opted to decompose the data into a number of lower-dimensional subspaces and perform Transformers in each sub-space individually. As a side effect, these methods inadvertently restrict the self-attention mechanisms to a One-to-One scheme accessing only a limited subset of LF data, explicitly preventing comprehensive optimization on all spatial and angular cues. In this paper, we identify this limitation as subspace isolation and introduce a novel Many-to-Many Transformer (M2MT) to address it. M2MT aggregates angular information in the spatial subspace before performing the self-attention mechanism. It enables complete access to all information across all sub-aperture images (SAIs) in a light field image. Consequently, M2MT is enabled to comprehensively capture long-range correlation dependencies. With M2MT as the foundational component, we develop a simple yet effective M2MT network for LFSR. Our experimental results demonstrate that M2MT achieves state-of-the-art performance across various public datasets, and it offers a favorable balance between model performance and efficiency, yielding higher-quality LFSR results with substantially lower demand for memory and computation. We further conduct in-depth analysis using local attribution maps (LAM) to obtain visual interpretability, and the results validate that M2MT is empowered with a truly non-local context in both spatial and angular subspaces to mitigate subspace isolation and acquire effective spatial-angular representation.
[955] GuideGen: A Text-Guided Framework for Paired Full-torso Anatomy and CT Volume Generation
Linrui Dai, Rongzhao Zhang, Yongrui Yu, Xiaofan Zhang
Main category: eess.IV
TL;DR: GuideGen is a controllable framework that generates full-torso CT volumes and anatomical masks from text prompts, enabling data synthesis for segmentation tasks without requiring large medical imaging datasets.
Details
Motivation: Existing 3D CT generation methods focus on specific organs with fixed contrast and don't fully leverage both semantic and textual conditions. There's a need for a more comprehensive approach to generate full-torso CT data using easily-acquired text prompts to reduce the labor and expenses of building large 3D medical imaging datasets.Method: Three core components: 1) text-conditional semantic synthesizer for creating realistic full-torso anatomies, 2) anatomy-aware high-dynamic-range (HDR) autoencoder for high-fidelity feature extraction across varying intensity levels, and 3) latent feature generator ensuring alignment between CT images, anatomical semantics, and input prompts.
Result: GuideGen demonstrates superiority over existing CT generation methods in comprehensive evaluations across generation quality, cross-modality alignment, and data usability on multi-organ and tumor segmentation tasks, using a dataset compiled from 12 public TCIA datasets and one private real-world dataset.
Conclusion: GuideGen provides an effective framework for generating full-torso CT volumes and anatomical masks from text prompts, enabling data synthesis for segmentation tasks and addressing the challenges of building large 3D medical imaging datasets.
Abstract: The recently emerging conditional diffusion models seem promising for mitigating the labor and expenses in building large 3D medical imaging datasets. However, previous studies on 3D CT generation primarily focus on specific organs characterized by a local structure and fixed contrast and have yet to fully capitalize on the benefits of both semantic and textual conditions. In this paper, we present GuideGen, a controllable framework based on easily-acquired text prompts to generate anatomical masks and corresponding CT volumes for the entire torso-from chest to pelvis. Our approach includes three core components: a text-conditional semantic synthesizer for creating realistic full-torso anatomies; an anatomy-aware high-dynamic-range (HDR) autoencoder for high-fidelity feature extraction across varying intensity levels; and a latent feature generator that ensures alignment between CT images, anatomical semantics and input prompts. Combined, these components enable data synthesis for segmentation tasks from only textual instructions. To train and evaluate GuideGen, we compile a multi-modality cancer imaging dataset with paired CT and clinical descriptions from 12 public TCIA datasets and one private real-world dataset. Comprehensive evaluations across generation quality, cross-modality alignment, and data usability on multi-organ and tumor segmentation tasks demonstrate GuideGen’s superiority over existing CT generation methods. Relevant materials are available at https://github.com/OvO1111/GuideGen.
[956] Self-Supervised One-Step Diffusion Refinement for Snapshot Compressive Imaging
Shaoguang Huang, Yunzhen Wang, Haijin Zeng, Hongyu Chen, Hongyan Zhang
Main category: eess.IV
TL;DR: OSD: A self-supervised one-step diffusion framework for SCI reconstruction that eliminates iterative denoising, uses equivariant learning from raw measurements, and distills RGB priors to overcome MSI data limitations.
Details
Motivation: Existing diffusion-based SCI reconstruction methods face three critical limitations: lack of large-scale MSI training data, domain shifts from RGB-pretrained models, and inefficient multi-step sampling that restricts real-world practicality.Method: Proposes OSD framework with: 1) single-step diffusion refiner to correct initial reconstruction without iterative denoising, 2) self-supervised equivariant learning to train predictor and refiner directly from raw 2D measurements, and 3) band-selection-driven distillation to transfer generative priors from RGB datasets to bridge domain gap.
Result: Sets new benchmarks with PSNR gains of 3.44 dB (Harvard), 1.61 dB (NTIRE), and 0.28 dB (ICVL) while reducing reconstruction time by 97.5% compared to existing methods.
Conclusion: OSD represents a significant advancement in SCI reconstruction by combining accuracy and practicality through efficient one-step diffusion, self-supervised learning from raw measurements, and effective domain adaptation, enabling real-world deployment.
Abstract: Snapshot compressive imaging (SCI) captures multispectral images (MSIs) using a single coded two-dimensional (2-D) measurement, but reconstructing high-fidelity MSIs from these compressed inputs remains a fundamentally ill-posed challenge. While diffusion-based reconstruction methods have recently raised the bar for quality, they face critical limitations: a lack of large-scale MSI training data, adverse domain shifts from RGB-pretrained models, and inference inefficiencies due to multi-step sampling. These drawbacks restrict their practicality in real-world applications. In contrast to existing methods, which either follow costly iterative refinement or adapt subspace-based embeddings for diffusion models (e.g. DiffSCI, PSR-SCI), we introduce a fundamentally different paradigm: a self-supervised One-Step Diffusion (OSD) framework specifically designed for SCI. The key novelty lies in using a single-step diffusion refiner to correct an initial reconstruction, eliminating iterative denoising entirely while preserving generative quality. Moreover, we adopt a self-supervised equivariant learning strategy to train both the predictor and refiner directly from raw 2-D measurements, enabling generalization to unseen domains without the need for ground-truth MSI. To further address the challenge of limited MSI data, we design a band-selection-driven distillation strategy that transfers core generative priors from large-scale RGB datasets, effectively bridging the domain gap. Extensive experiments confirm that our approach sets a new benchmark, yielding PSNR gains of 3.44 dB, 1.61 dB, and 0.28 dB on the Harvard, NTIRE, and ICVL datasets, respectively, while reducing reconstruction time by 97.5%. This remarkable improvement in efficiency and adaptability makes our method a significant advancement in SCI reconstruction, combining both accuracy and practicality for real-world deployment.
[957] 3D MedDiffusion: A 3D Medical Latent Diffusion Model for Controllable and High-quality Medical Image Generation
Haoshen Wang, Zhentao Liu, Kaicong Sun, Xiaodong Wang, Dinggang Shen, Zhiming Cui
Main category: eess.IV
TL;DR: 3D MedDiffusion: A novel 3D medical image generation framework using patch-volume autoencoder and improved noise estimator for high-resolution (up to 512³) controllable generation across CT/MRI modalities.
Details
Motivation: Existing methods struggle with high-quality 3D medical image generation due to high-resolution requirements and lack of universal generative frameworks for medical imaging.Method: Proposes 3D MedDiffusion with: 1) Patch-Volume Autoencoder for efficient compression/decoding, 2) Novel noise estimator capturing both local details and global structure during diffusion denoising.
Result: Generates high-resolution images up to 512x512x512, surpasses SOTA methods in quality, and shows strong generalizability across tasks including sparse-view CT reconstruction, fast MRI reconstruction, and data augmentation.
Conclusion: 3D MedDiffusion provides a universal generative framework for high-quality 3D medical imaging with demonstrated effectiveness across multiple modalities and downstream applications.
Abstract: The generation of medical images presents significant challenges due to their high-resolution and three-dimensional nature. Existing methods often yield suboptimal performance in generating high-quality 3D medical images, and there is currently no universal generative framework for medical imaging. In this paper, we introduce a 3D Medical Latent Diffusion (3D MedDiffusion) model for controllable, high-quality 3D medical image generation. 3D MedDiffusion incorporates a novel, highly efficient Patch-Volume Autoencoder that compresses medical images into latent space through patch-wise encoding and recovers back into image space through volume-wise decoding. Additionally, we design a new noise estimator to capture both local details and global structural information during diffusion denoising process. 3D MedDiffusion can generate fine-detailed, high-resolution images (up to 512x512x512) and effectively adapt to various downstream tasks as it is trained on large-scale datasets covering CT and MRI modalities and different anatomical regions (from head to leg). Experimental results demonstrate that 3D MedDiffusion surpasses state-of-the-art methods in generative quality and exhibits strong generalizability across tasks such as sparse-view CT reconstruction, fast MRI reconstruction, and data augmentation for segmentation and classification. Source code and checkpoints are available at https://github.com/ShanghaiTech-IMPACT/3D-MedDiffusion.
[958] Full-scale Representation Guided Network for Retinal Vessel Segmentation
Sunyong Seo, Sangwook Yoo, Huisu Yoon
Main category: eess.IV
TL;DR: FSG-Net introduces a novel Full-Scale Guided Network for retinal vessel segmentation that captures full-scale structural information and refines it using attention-guided filters, achieving competitive performance with SOTA methods despite compact architecture.
Details
Motivation: U-Net variants have dominated retinal vessel segmentation for a decade, but there's room for improvement in capturing full-scale structural information and refining fine vascular details. The authors aim to develop a more effective approach that maintains flexibility and scalability.Method: Proposes FSG-Net with: 1) Feature representation module using modernized convolution blocks to capture full-scale structural information, 2) Guided convolution block with attention-guided filter (similar to unsharp masking) to enhance fine vascular structures, 3) Attention mechanism that uses full-scale information to generate contextually relevant attention maps, 4) Flexible architecture where the front-end can be replaced by any U-Net variant.
Result: FSG-Net achieves competitive performance with state-of-the-art methods across multiple public datasets despite its compact architecture. The authors re-implemented recent studies for fair comparison and demonstrated scalability and reproducibility. Ablation studies confirm meaningful contributions from each proposed component.
Conclusion: FSG-Net provides an effective, flexible solution for retinal vessel segmentation that captures full-scale information and refines fine details through attention-guided filtering, offering competitive performance while maintaining compact architecture and scalability across various segmentation tasks.
Abstract: The U-Net architecture and its variants have remained state-of-the-art (SOTA) for retinal vessel segmentation over the past decade. In this study, we introduce a Full-Scale Guided Network (FSG-Net), where a novel feature representation module using modernized convolution blocks effectively captures full-scale structural information, while a guided convolution block subsequently refines this information. Specifically, we introduce an attention-guided filter within the guided convolution block, leveraging its similarity to unsharp masking to enhance fine vascular structures. Passing full-scale information to the attention block facilitates the generation of more contextually relevant attention maps, which are then passed to the attention-guided filter, providing further refinement to the segmentation performance. The structure preceding the guided convolution block can be replaced by any U-Net variant, ensuring flexibility and scalability across various segmentation tasks. For a fair comparison, we re-implemented recent studies available in public repositories to evaluate their scalability and reproducibility. Our experiments demonstrate that, despite its compact architecture, FSG-Net delivers performance competitive with SOTA methods across multiple public datasets. Ablation studies further demonstrate that each proposed component meaningfully contributes to this competitive performance. Our code is available on https://github.com/ZombaSY/FSG-Net-pytorch.
[959] GBT-SAM: A Parameter-Efficient Depth-Aware Model for Generalizable Brain tumour Segmentation on mp-MRI
Cecilia Diana-Albelda, Roberto Alcover-Couso, Álvaro García-Martín, Jesus Bescos, Marcos Escudero-Viñolo
Main category: eess.IV
TL;DR: GBT-SAM is a parameter-efficient deep learning framework that adapts Segment Anything Model (SAM) for brain tumor segmentation in multi-parametric MRI, achieving high accuracy with minimal computational resources.
Details
Motivation: Manual glioma segmentation is time-consuming and variable, while existing deep learning methods fail to fully exploit multi-parametric MRI information, require heavy computation, and lack robustness across tumor types.Method: Adapts SAM to volumetric mp-MRI by selecting <2.6% of slices per scan, uses two-step fine-tuning with depth-aware module for inter-slice correlations, and lightweight adaptation layers (only 6.5M trainable parameters).
Result: Achieves Dice Score of 93.54 on BraTS Adult Glioma dataset and demonstrates robust performance on Meningioma, Pediatric Glioma, and Sub-Saharan Glioma datasets.
Conclusion: GBT-SAM is a computationally efficient and domain-robust framework for brain tumor segmentation using mp-MRI, with potential for clinical application.
Abstract: Gliomas are aggressive brain tumors that require accurate imaging-based diagnosis, with segmentation playing a critical role in evaluating morphology and treatment decisions. Manual delineation of gliomas is time-consuming and prone to variability, motivating the use of deep learning to improve consistency and alleviate clinical workload. However, existing methods often fail to fully exploit the information available in multi-parametric MRI (mp-MRI), particularly inter-slice contextual features, and typically require considerable computational resources while lacking robustness across tumor type variations. We present GBT-SAM, a parameter-efficient deep learning framework that adapts the Segment Anything Model (SAM), a large-scale vision model, to volumetric mp-MRI data. GBT-SAM reduces input complexity by selecting fewer than 2.6% of slices per scan while incorporating all four MRI modalities, preserving essential tumor-related information with minimal cost. Furthermore, our model is trained by a two-step fine-tuning strategy that incorporates a depth-aware module to capture inter-slice correlations and lightweight adaptation layers, resulting in just 6.5M trainable parameters, which is the lowest among SAM-based approaches. GBT-SAM achieves a Dice Score of 93.54 on the BraTS Adult Glioma dataset and demonstrates robust performance on Meningioma, Pediatric Glioma, and Sub-Saharan Glioma datasets. These results highlight GBT-SAM’s potential as a computationally efficient and domain-robust framework for brain tumor segmentation using mp-MRI. Our code and models are available at https://github.com/vpulab/med-sam-brain .
[960] SCOPE-MRI: Bankart Lesion Detection as a Case Study in Data Curation and Deep Learning for Challenging Diagnoses
Sahil Sethi, Sai Reddy, Mansi Sakarvadia, Jordan Serotte, Darlington Nwaudo, Nicholas Maassen, Lewis Shi
Main category: eess.IV
TL;DR: ScopeMRI introduces the first public expert-annotated dataset for shoulder pathologies and a deep learning framework that achieves radiologist-level performance in detecting challenging Bankart lesions on standard MRIs and MRAs.
Details
Motivation: Bankart lesions (anterior-inferior glenoid labral tears) are difficult to diagnose on standard MRIs due to subtle imaging features, often requiring invasive MRI arthrograms. Prior deep learning work has focused on straightforward musculoskeletal conditions, leaving challenging diagnostic problems underexplored.Method: Created ScopeMRI dataset with shoulder MRIs from arthroscopy patients (gold standard ground truth). Trained separate CNN- and transformer-based models for MRIs and MRAs, with predictions ensembled across multiple imaging planes. Used modular codebase for training and evaluation.
Result: Models achieved radiologist-level performance, with accuracy on standard MRIs surpassing radiologists interpreting MRAs. External validation on independent hospital data demonstrated initial generalizability across imaging protocols.
Conclusion: By releasing ScopeMRI dataset and modular codebase, the authors aim to accelerate musculoskeletal imaging research and foster development of datasets/models for clinically challenging diagnostic tasks, showing deep learning can match expert performance on difficult conditions.
Abstract: Deep learning has shown strong performance in musculoskeletal imaging, but prior work has largely targeted conditions where diagnosis is relatively straightforward. More challenging problems remain underexplored, such as detecting Bankart lesions (anterior-inferior glenoid labral tears) on standard MRIs. These lesions are difficult to diagnose due to subtle imaging features, often necessitating invasive MRI arthrograms (MRAs). We introduce ScopeMRI, the first publicly available, expert-annotated dataset for shoulder pathologies, and present a deep learning framework for Bankart lesion detection on both standard MRIs and MRAs. ScopeMRI contains shoulder MRIs from patients who underwent arthroscopy, providing ground-truth labels from intraoperative findings, the diagnostic gold standard. Separate models were trained for MRIs and MRAs using CNN- and transformer-based architectures, with predictions ensembled across multiple imaging planes. Our models achieved radiologist-level performance, with accuracy on standard MRIs surpassing radiologists interpreting MRAs. External validation on independent hospital data demonstrated initial generalizability across imaging protocols. By releasing ScopeMRI and a modular codebase for training and evaluation, we aim to accelerate research in musculoskeletal imaging and foster development of datasets and models that address clinically challenging diagnostic tasks.
[961] Zero-shot Denoising via Neural Compression: Theoretical and algorithmic framework
Ali Zafari, Xi Chen, Shirin Jalali
Main category: eess.IV
TL;DR: ZS-NCD: A zero-shot denoising method using neural compression as an untrained model optimized directly on patches from a single noisy image, achieving SOTA performance without training data or clean references.
Details
Motivation: Zero-shot denoising is crucial for practical imaging in specialized domains like medical imaging or biology where training samples or clean reference images are unavailable. Existing methods often require manual regularization or early stopping to prevent overfitting.Method: Treats a neural compression network as an untrained model optimized directly on patches from a single noisy image. Uses built-in entropy constraints of compression architectures to naturally avoid overfitting without manual regularization. Final reconstruction aggregates outputs over overlapping patches.
Result: Achieves state-of-the-art performance among zero-shot denoisers for both Gaussian and Poisson noise. Generalizes well to both natural and non-natural images. Provides new finite-sample theoretical results characterizing upper bounds on reconstruction error for compression-based denoisers.
Conclusion: ZS-NCD offers an effective zero-shot denoising framework that leverages neural compression’s inherent regularization properties, eliminating the need for training data or manual regularization while establishing theoretical foundations for compression-based denoising.
Abstract: Zero-shot denoising aims to denoise observations without access to training samples or clean reference images. This setting is particularly relevant in practical imaging scenarios involving specialized domains such as medical imaging or biology. In this work, we propose the Zero-Shot Neural Compression Denoiser (ZS-NCD), a novel denoising framework based on neural compression. ZS-NCD treats a neural compression network as an untrained model, optimized directly on patches extracted from a single noisy image. The final reconstruction is then obtained by aggregating the outputs of the trained model over overlapping patches. Thanks to the built-in entropy constraints of compression architectures, our method naturally avoids overfitting and does not require manual regularization or early stopping. Through extensive experiments, we show that ZS-NCD achieves state-of-the-art performance among zero-shot denoisers for both Gaussian and Poisson noise, and generalizes well to both natural and non-natural images. Additionally, we provide new finite-sample theoretical results that characterize upper bounds on the achievable reconstruction error of general maximum-likelihood compression-based denoisers. These results further establish the theoretical foundations of compression-based denoising. Our code is available at: https://github.com/Computational-Imaging-RU/ZS-NCDenoiser.
[962] Cross-Cancer Knowledge Transfer in WSI-based Prognosis Prediction
Pei Liu, Luping Ji, Jiaxiang Gou, Xiangxiang Zeng
Main category: eess.IV
TL;DR: CROPKT is the first systematic study on cross-cancer prognosis knowledge transfer in Whole-Slide Images, shifting from cancer-specific models to knowledge transfer paradigms.
Details
Motivation: Current cancer prognosis models are cancer-specific, struggling with rare tumors and unable to leverage knowledge from other cancers. Multi-task learning requires extensive computational resources and large datasets.Method: Three-part approach: (1) Curate UNI2-h-DSS dataset with 26 cancers to measure transferability, (2) Design experiments to understand transferability mechanisms, (3) Propose ROUPKT - a routing-based baseline approach for efficient knowledge transfer from off-the-shelf models.
Result: The study establishes foundational work for cross-cancer knowledge transfer in WSI-based prognosis prediction, demonstrating utility of transferred knowledge and providing systematic evaluation framework.
Conclusion: CROPKT serves as an inception for the nascent paradigm of WSI-based prognosis prediction with cross-cancer knowledge transfer, offering a more scalable approach that can handle rare tumors and leverage knowledge across cancers.
Abstract: Whole-Slide Image (WSI) is an important tool for estimating cancer prognosis. Current studies generally follow a conventional cancer-specific paradigm in which each cancer corresponds to a single model. However, this paradigm naturally struggles to scale to rare tumors and cannot leverage knowledge from other cancers. While multi-task learning frameworks have been explored recently, they often place high demands on computational resources and require extensive training on ultra-large, multi-cancer WSI datasets. To this end, this paper shifts the paradigm to knowledge transfer and presents the first preliminary yet systematic study on cross-cancer prognosis knowledge transfer in WSIs, called CROPKT. It comprises three major parts. (1) We curate a large dataset (UNI2-h-DSS) with 26 cancers and use it to measure the transferability of WSI-based prognostic knowledge across different cancers (including rare tumors). (2) Beyond a simple evaluation merely for benchmarking, we design a range of experiments to gain deeper insights into the underlying mechanism behind transferability. (3) We further show the utility of cross-cancer knowledge transfer, by proposing a routing-based baseline approach (ROUPKT) that could often efficiently utilize the knowledge transferred from off-the-shelf models of other cancers. CROPKT could serve as an inception that lays the foundation for this nascent paradigm, i.e., WSI-based prognosis prediction with cross-cancer knowledge transfer. Our source code is available at https://github.com/liupei101/CROPKT.
[963] TomoGraphView: 3D Medical Image Classification with Omnidirectional Slice Representations and Graph Neural Networks
Johannes Kiechle, Stefan M. Fischer, Daniel M. Lang, Cosmin I. Bercea, Matthew J. Nyflot, Lina Felsner, Julia A. Schnabel, Jan C. Peeken
Main category: eess.IV
TL;DR: TomoGraphView: A framework using omnidirectional volume slicing with spherical graph-based feature aggregation for 3D medical image analysis, overcoming limitations of standard 2D slice-based approaches.
Details
Motivation: The rise in medical tomography exams demands automated feature extraction systems. While 3D volumes contain richer information than 2D slices, effective 3D classification is difficult due to complex spatial dependencies and scarcity of large-scale 3D datasets. Current 2D foundation model approaches applied to 3D scans via slice decomposition are fundamentally limited - they fail to capture true spatial extent when structures don't align with canonical views and lose spatial coherence by aggregating slice features independently.Method: Proposes TomoGraphView framework integrating omnidirectional volume slicing with spherical graph-based feature aggregation. Instead of restricting to axial, sagittal, or coronal planes, samples both canonical and non-canonical cross-sections from uniformly distributed points on a sphere enclosing the volume. Uses graph-based approach to aggregate features while preserving 3D spatial relationships.
Result: The authors share their accessible code base and provide a user-friendly library for omnidirectional volume slicing (OmniSlicer). The method addresses limitations of current approaches by capturing true spatial extent regardless of orientation and maintaining spatial coherence across slices.
Conclusion: TomoGraphView represents a novel approach to 3D medical image analysis that overcomes fundamental limitations of current 2D foundation model methods by integrating omnidirectional slicing with graph-based feature aggregation, better capturing 3D spatial relationships and structure orientation.
Abstract: The sharp rise in medical tomography examinations has created a demand for automated systems that can reliably extract informative features for downstream tasks such as tumor characterization. Although 3D volumes contain richer information than individual slices, effective 3D classification remains difficult: volumetric data encode complex spatial dependencies, and the scarcity of large-scale 3D datasets has constrained progress toward 3D foundation models. As a result, many recent approaches rely on 2D vision foundation models trained on natural images, repurposing them as feature extractors for medical scans with surprisingly strong performance. Despite their practical success, current methods that apply 2D foundation models to 3D scans via slice-based decomposition remain fundamentally limited. Standard slicing along axial, sagittal, and coronal planes often fails to capture the true spatial extent of a structure when its orientation does not align with these canonical views. More critically, most approaches aggregate slice features independently, ignoring the underlying 3D geometry and losing spatial coherence across slices. To overcome these limitations, we propose TomoGraphView, a novel framework that integrates omnidirectional volume slicing with spherical graph-based feature aggregation. Instead of restricting the model to axial, sagittal, or coronal planes, our method samples both canonical and non-canonical cross-sections generated from uniformly distributed points on a sphere enclosing the volume. We publicly share our accessible code base at http://github.com/compai-lab/2025-MedIA-kiechle and provide a user-friendly library for omnidirectional volume slicing at https://pypi.org/project/OmniSlicer.
[964] MRI Super-Resolution with Deep Learning: A Comprehensive Survey
Mohammad Khateri, Serge Vasylechko, Morteza Ghahremani, Liam Timms, Deniz Kocanaogullari, Simon K. Warfield, Camilo Jaimes, Davood Karimi, Alejandra Sierra, Jussi Tohka, Sila Kurugol, Onur Afacan
Main category: eess.IV
TL;DR: Survey paper reviewing deep learning-based MRI super-resolution techniques, covering methods, challenges, and providing resources for the community.
Details
Motivation: High-resolution MRI is clinically important but expensive and technically constrained. Super-resolution offers a computational solution using deep learning to generate HR images from affordable LR scans without additional hardware.Method: Systematic review and taxonomy of DL-based MRI SR methods from multiple perspectives: computer vision, computational imaging, inverse problems, and MR physics. Includes analysis of theoretical foundations, architectures, learning strategies, datasets, and metrics.
Result: Comprehensive survey categorizing established and emerging MRI SR techniques, highlighting unique clinical/research challenges, open problems, and providing GitHub repository with resources, tools, and tutorials.
Conclusion: Deep learning-based MRI super-resolution is a promising computational approach to overcome hardware limitations, but requires addressing specific clinical challenges and advancing methods for practical adoption.
Abstract: High-resolution (HR) magnetic resonance imaging (MRI) is crucial for many clinical and research applications. However, achieving it remains costly and constrained by technical trade-offs and experimental limitations. Super-resolution (SR) presents a promising computational approach to overcome these challenges by generating HR images from more affordable low-resolution (LR) scans, potentially improving diagnostic accuracy and efficiency without requiring additional hardware. This survey reviews recent advances in MRI SR techniques, with a focus on deep learning (DL) approaches. It examines DL-based MRI SR methods from the perspectives of computer vision, computational imaging, inverse problems, and MR physics, covering theoretical foundations, architectural designs, learning strategies, benchmark datasets, and performance metrics. We propose a systematic taxonomy to categorize these methods and present an in-depth study of both established and emerging SR techniques applicable to MRI, considering unique challenges in clinical and research contexts. We also highlight open challenges and directions that the community needs to address. Additionally, we provide a collection of essential open-access resources, tools, and tutorials, available on our GitHub: https://github.com/mkhateri/Awesome-MRI-Super-Resolution. IEEE keywords: MRI, Super-Resolution, Deep Learning, Computational Imaging, Inverse Problem, Survey.
[965] Deep Learning for Restoring MPI System Matrices Using Simulated Training Data
Artyom Tsanda, Sarah Reiss, Konrad Scheffler, Marija Boberg, Tobias Knopp
Main category: eess.IV
TL;DR: Deep learning models trained on physics-based simulated system matrices can generalize to real magnetic particle imaging data for various restoration tasks, addressing data scarcity issues.
Details
Motivation: Magnetic particle imaging relies on system matrices obtained through time-consuming, noisy calibration measurements. Deep learning methods need curated training data, which is scarce. The study investigates whether physics-based simulated system matrices can train models that generalize to real measured data.Method: Generated a large dataset of system matrices using an equilibrium magnetization model extended with uniaxial anisotropy, spanning various particle, scanner, and calibration parameters for 2D/3D trajectories. Included background noise from empty-frame measurements. Compared deep learning models (DnCNN, RDN, SwinIR, SMRnet, PConvUNet) with classical non-learning baselines for four restoration tasks: denoising, accelerated calibration, upsampling, and inpainting.
Result: Models trained solely on simulated data generalized to measured data across all tasks: For denoising, DnCNN/RDN/SwinIR outperformed DCT-F baseline by >10 dB PSNR and up to 0.1 SSIM on simulations, leading to perceptually better reconstructions of real data. For 2D upsampling, SMRnet exceeded bicubic by 20 dB PSNR and 0.08 SSIM at ×2-×4 scales, though qualitative transfer to real measurements was limited. For 3D accelerated calibration, SMRnet matched tricubic in noiseless cases and was more robust under noise. For 3D inpainting, biharmonic inpainting was superior when noise-free but degraded with noise, while PConvUNet maintained quality and yielded less blurry reconstructions.
Conclusion: Deep learning models trained on physics-based simulations can effectively transfer to real magnetic particle imaging measurements, mitigating the data-scarcity problem and enabling development of new methods beyond current measurement capabilities.
Abstract: Magnetic particle imaging reconstructs tracer distributions using a system matrix obtained through time-consuming, noise-prone calibration measurements. Methods for addressing imperfections in measured system matrices increasingly rely on deep neural networks, yet curated training data remain scarce. This study evaluates whether physics-based simulated system matrices can be used to train deep learning models for different system matrix restoration tasks, i.e., denoising, accelerated calibration, upsampling, and inpainting, that generalize to measured data. A large system matrices dataset was generated using an equilibrium magnetization model extended with uniaxial anisotropy. The dataset spans particle, scanner, and calibration parameters for 2D and 3D trajectories, and includes background noise injected from empty-frame measurements. For each restoration task, deep learning models were compared with classical non-learning baseline methods. The models trained solely on simulated system matrices generalized to measured data across all tasks: for denoising, DnCNN/RDN/SwinIR outperformed DCT-F baseline by >10 dB PSNR and up to 0.1 SSIM on simulations and led to perceptually better reconstuctions of real data; for 2D upsampling, SMRnet exceeded bicubic by 20 dB PSNR and 0.08 SSIM at $\times 2$-$\times 4$ which did not transfer qualitatively to real measurements. For 3D accelerated calibration, SMRnet matched tricubic in noiseless cases and was more robust under noise, and for 3D inpainting, biharmonic inpainting was superior when noise-free but degraded with noise, while a PConvUNet maintained quality and yielded less blurry reconstructions. The demonstrated transferability of deep learning models trained on simulations to real measurements mitigates the data-scarcity problem and enables the development of new methods beyond current measurement capabilities.