Archives AI News

Bayesian Double Descent

arXiv:2507.07338v3 Announce Type: replace-cross Abstract: Double descent is a phenomenon of over-parameterized statistical models such as deep neural networks which have a re-descending property in their risk function. As the complexity of the model increases, risk exhibits a U-shaped region…

Randomness and Interpolation Improve Gradient Descent

arXiv:2510.13040v1 Announce Type: new Abstract: Based on Stochastic Gradient Descent (SGD), the paper introduces two optimizers, named Interpolational Accelerating Gradient Descent (IAGD) as well as Noise-Regularized Stochastic Gradient Descent (NRSGD). IAGD leverages second-order Newton Interpolation to expedite the convergence process…

On Pretraining for Project-Level Code Completion

arXiv:2510.13697v1 Announce Type: cross Abstract: Repository-level pretraining is commonly used to enable large language models for code to leverage codebase-wide context. This enhances their ability to generate accurate and context-aware code completions. In this work, we investigate how different repository-processing…

Time-Varying Optimization for Streaming Data Via Temporal Weighting

arXiv:2510.13052v1 Announce Type: new Abstract: Classical optimization theory deals with fixed, time-invariant objective functions. However, time-varying optimization has emerged as an important subject for decision-making in dynamic environments. In this work, we study the problem of learning from streaming data…

PriorGuide: Test-Time Prior Adaptation for Simulation-Based Inference

arXiv:2510.13763v1 Announce Type: cross Abstract: Amortized simulator-based inference offers a powerful framework for tackling Bayesian inference in computational fields such as engineering or neuroscience, increasingly leveraging modern generative methods like diffusion models to map observed data to model parameters or…

Achieving Logarithmic Regret in KL-Regularized Zero-Sum Markov Games

arXiv:2510.13060v1 Announce Type: new Abstract: Reverse Kullback-Leibler (KL) divergence-based regularization with respect to a fixed reference policy is widely used in modern reinforcement learning to preserve the desired traits of the reference policy and sometimes to promote exploration (using uniform…

Do LLM Agents Have Regret? A Case Study in Online Learning and Games

arXiv:2403.16843v5 Announce Type: replace Abstract: Large language models (LLMs) have been increasingly employed for (interactive) decision-making, via the development of LLM-based autonomous agents. Despite their emerging successes, the performance of LLM agents in decision-making has not been fully investigated through…