Archives AI News

Tool Zero: Training Tool-Augmented LLMs via Pure RL from Scratch

arXiv:2511.01934v1 Announce Type: new Abstract: Training tool-augmented LLMs has emerged as a promising approach to enhancing language models’ capabilities for complex tasks. The current supervised fine-tuning paradigm relies on constructing extensive domain-specific datasets to train models. However, this approach often…

Optimizing Kernel Discrepancies via Subset Selection

arXiv:2511.02706v1 Announce Type: cross Abstract: Kernel discrepancies are a powerful tool for analyzing worst-case errors in quasi-Monte Carlo (QMC) methods. Building on recent advances in optimizing such discrepancy measures, we extend the subset selection problem to the setting of kernel…

Remasking Discrete Diffusion Models with Inference-Time Scaling

arXiv:2503.00307v3 Announce Type: replace Abstract: Part of the success of diffusion models stems from their ability to perform iterative refinement, i.e., repeatedly correcting outputs during generation. However, modern masked discrete diffusion lacks this capability: when a token is generated, it…

The Geometry of Grokking: Norm Minimization on the Zero-Loss Manifold

arXiv:2511.01938v1 Announce Type: new Abstract: Grokking is a puzzling phenomenon in neural networks where full generalization occurs only after a substantial delay following the complete memorization of the training data. Previous research has linked this delayed generalization to representation learning…