Archives AI News

EvoSyn: Generalizable Evolutionary Data Synthesis for Verifiable Learning

arXiv:2510.17928v1 Announce Type: new Abstract: Reliable verifiable data has become a key driver of capability gains in modern language models, enabling stable reinforcement learning with verifiable rewards and effective distillation that transfers competence across math, coding, and agentic tasks. Yet…

Understanding Differential Transformer Unchains Pretrained Self-Attentions

arXiv:2505.16333v3 Announce Type: replace Abstract: Differential Transformer has recently gained significant attention for its impressive empirical performance, often attributed to its ability to perform noise canceled attention. However, precisely how differential attention achieves its empirical benefits remains poorly understood. Moreover,…

TPP-SD: Accelerating Transformer Point Process Sampling with Speculative Decoding

arXiv:2507.09252v3 Announce Type: replace Abstract: We propose TPP-SD, a novel approach that accelerates Transformer temporal point process (TPP) sampling by adapting speculative decoding (SD) techniques from language models. By identifying the structural similarities between thinning algorithms for TPPs and speculative…

Deep Edge Filter: Return of the Human-Crafted Layer in Deep Learning

arXiv:2510.13865v3 Announce Type: replace Abstract: We introduce the Deep Edge Filter, a novel approach that applies high-pass filtering to deep neural network features to improve model generalizability. Our method is motivated by our hypothesis that neural networks encode task-relevant semantic…

Demystifying Transition Matching: When and Why It Can Beat Flow Matching

arXiv:2510.17991v1 Announce Type: new Abstract: Flow Matching (FM) underpins many state-of-the-art generative models, yet recent results indicate that Transition Matching (TM) can achieve higher quality with fewer sampling steps. This work answers the question of when and why TM outperforms…