Archives AI News

Thin Keys, Full Values: Reducing KV Cache via Low-Dimensional Attention Selection

arXiv:2603.04427v4 Announce Type: replace Abstract: Standard Transformer attention uses identical dimensionality for queries, keys, and values, yet these components serve different roles: queries and keys produce scalar attention weights (selection), while values carry rich representations (value transfer). We show that…

Strategic Candidacy in Generative AI Arenas

arXiv:2603.26891v1 Announce Type: new Abstract: AI arenas, which rank generative models from pairwise preferences of users, are a popular method for measuring the relative performance of models in the course of their organic use. Because rankings are computed from noisy…

Scaling Attention via Feature Sparsity

arXiv:2603.22300v2 Announce Type: replace Abstract: Scaling Transformers to ultra-long contexts is bottlenecked by the $O(n^2 d)$ cost of self-attention. Existing methods reduce this cost along the sequence axis through local windows, kernel approximations, or token-level sparsity, but these approaches consistently…

Tunable Domain Adaptation Using Unfolding

arXiv:2603.26931v1 Announce Type: new Abstract: Machine learning models often struggle to generalize across domains with varying data distributions, such as differing noise levels, leading to degraded performance. Traditional strategies like personalized training, which trains separate models per domain, and joint…

Empirical Likelihood for Nonsmooth Functionals

arXiv:2603.27743v1 Announce Type: cross Abstract: Empirical likelihood is an attractive inferential framework that respects natural parameter boundaries, but existing approaches typically require smoothness of the functional and miscalibrate substantially when these assumptions are violated. For the optimal-value functional central to…

On the Hardness of Reinforcement Learning with Transition Look-Ahead

arXiv:2510.19372v2 Announce Type: replace-cross Abstract: We study reinforcement learning (RL) with transition look-ahead, where the agent may observe which states would be visited upon playing any sequence of $ell$ actions before deciding its course of action. While such predictive information…

Noise in Photonic Quantum Machine Learning: Models, Impacts, and Mitigation Strategies

arXiv:2603.09645v2 Announce Type: replace-cross Abstract: Photonic Quantum Machine Learning (PQML) is an emerging method to implement scalable, energy-efficient quantum information processing by combining photonic quantum computing technologies with machine learning techniques. The features of photonic technologies offer several benefits: room-temperature…