Archives AI News

When do spectral gradient updates help in deep learning?

arXiv:2512.04299v1 Announce Type: new Abstract: Spectral gradient methods, such as the recently popularized Muon optimizer, are a promising alternative to standard Euclidean gradient descent for training deep neural networks and transformers, but it is still unclear in which regimes they…

Triangle Multiplication Is All You Need For Biomolecular Structure Representations

arXiv:2510.18870v2 Announce Type: replace-cross Abstract: AlphaFold has transformed protein structure prediction, but emerging applications such as virtual ligand screening, proteome-wide folding, and de novo binder design demand predictions at a massive scale, where runtime and memory costs become prohibitive. A…

Evaluating Long-Context Reasoning in LLM-Based WebAgents

arXiv:2512.04307v1 Announce Type: new Abstract: As large language model (LLM)-based agents become increasingly integrated into daily digital interactions, their ability to reason across long interaction histories becomes crucial for providing personalized and contextually aware assistance. However, the performance of these…

Arbitrage: Efficient Reasoning via Advantage-Aware Speculation

arXiv:2512.05033v1 Announce Type: cross Abstract: Modern Large Language Models achieve impressive reasoning capabilities with long Chain of Thoughts, but they incur substantial computational cost during inference, and this motivates techniques to improve the performance-cost ratio. Among these techniques, Speculative Decoding…

RNNs perform task computations by dynamically warping neural representations

arXiv:2512.04310v1 Announce Type: new Abstract: Analysing how neural networks represent data features in their activations can help interpret how they perform tasks. Hence, a long line of work has focused on mathematically characterising the geometry of such “neural representations.” In…

DraCo: Draft as CoT for Text-to-Image Preview and Rare Concept Generation

arXiv:2512.05112v1 Announce Type: cross Abstract: Recent unified multimodal large language models (MLLMs) have shown impressive capabilities, incorporating chain-of-thought (CoT) reasoning for enhanced text-to-image generation. However, existing approaches remain limited, either treating the model merely as a standalone generator or relying…

Data-regularized Reinforcement Learning for Diffusion Models at Scale

arXiv:2512.04332v1 Announce Type: new Abstract: Aligning generative diffusion models with human preferences via reinforcement learning (RL) is critical yet challenging. Most existing algorithms are often vulnerable to reward hacking, such as quality degradation, over-stylization, or reduced diversity. Our analysis demonstrates…