Archives AI News

Dispatch-Aware Ragged Attention for Pruned Vision Transformers

arXiv:2604.15408v1 Announce Type: new Abstract: Token pruning methods for Vision Transformers (ViTs) promise quadratic reductions in attention FLOPs by dropping uninformative patches. Yet when pruned sequences are executed with state-of-the-art variable-length attention APIs — including FlashAttention-2’s varlen and PyTorch’s NestedTensor…

Jailbreak Scaling Laws for Large Language Models: Polynomial-Exponential Crossover

arXiv:2603.11331v2 Announce Type: replace Abstract: Adversarial attacks can reliably steer safety-aligned large language models toward unsafe behavior. Empirically, we find that strong adversarial prompt-injection attacks can amplify attack success rate from the slow polynomial growth observed without injection to exponential…

Teaching Language Models Mechanistic Explainability Through MechSMILES

arXiv:2512.05722v2 Announce Type: replace Abstract: Chemical reaction mechanisms are the foundation of how chemists evaluate reactivity and feasibility, yet current Computer-Assisted Synthesis Planning (CASP) systems operate without this mechanistic reasoning. We introduce a computational framework that teaches language models to…