Archives AI News

ToMA: Token Merge with Attention for Diffusion Models

arXiv:2509.10918v2 Announce Type: replace Abstract: Diffusion models excel in high-fidelity image generation but face scalability limits due to transformers’ quadratic attention complexity. Plug-and-play token reduction methods like ToMeSD and ToFu reduce FLOPs by merging redundant tokens in generated images but…

NurseSchedRL: Attention-Guided Reinforcement Learning for Nurse-Patient Assignment

arXiv:2509.18125v1 Announce Type: new Abstract: Healthcare systems face increasing pressure to allocate limited nursing resources efficiently while accounting for skill heterogeneity, patient acuity, staff fatigue, and continuity of care. Traditional optimization and heuristic scheduling methods struggle to capture these dynamic,…

Adaptive Kernel Design for Bayesian Optimization Is a Piece of CAKE with LLMs

arXiv:2509.17998v2 Announce Type: replace Abstract: The efficiency of Bayesian optimization (BO) relies heavily on the choice of the Gaussian process (GP) kernel, which plays a central role in balancing exploration and exploitation under limited evaluation budgets. Traditional BO methods often…

Variational decision diagrams for quantum-inspired machine learning applications

arXiv:2502.04271v2 Announce Type: replace-cross Abstract: Decision diagrams (DDs) have emerged as an efficient tool for simulating quantum circuits due to their capacity to exploit data redundancies in quantum states and quantum operations, enabling the efficient computation of probability amplitudes. However,…

Meta-Semantics Augmented Few-Shot Relational Learning

arXiv:2505.05684v3 Announce Type: replace-cross Abstract: Few-shot relational learning on knowledge graph (KGs) aims to perform reasoning over relations with only a few training examples. While current methods have focused primarily on leveraging specific relational information, rich semantics inherent in KGs…

Training Language Model Agents to Find Vulnerabilities with CTF-Dojo

arXiv:2508.18370v2 Announce Type: replace-cross Abstract: Large language models (LLMs) have demonstrated exceptional capabilities when trained within executable runtime environments, notably excelling at software engineering tasks through verified feedback loops. Yet, scalable and generalizable execution-grounded environments remain scarce, limiting progress in…