Archives AI News

On Pretraining for Project-Level Code Completion

arXiv:2510.13697v1 Announce Type: cross Abstract: Repository-level pretraining is commonly used to enable large language models for code to leverage codebase-wide context. This enhances their ability to generate accurate and context-aware code completions. In this work, we investigate how different repository-processing…

Time-Varying Optimization for Streaming Data Via Temporal Weighting

arXiv:2510.13052v1 Announce Type: new Abstract: Classical optimization theory deals with fixed, time-invariant objective functions. However, time-varying optimization has emerged as an important subject for decision-making in dynamic environments. In this work, we study the problem of learning from streaming data…

PriorGuide: Test-Time Prior Adaptation for Simulation-Based Inference

arXiv:2510.13763v1 Announce Type: cross Abstract: Amortized simulator-based inference offers a powerful framework for tackling Bayesian inference in computational fields such as engineering or neuroscience, increasingly leveraging modern generative methods like diffusion models to map observed data to model parameters or…

Achieving Logarithmic Regret in KL-Regularized Zero-Sum Markov Games

arXiv:2510.13060v1 Announce Type: new Abstract: Reverse Kullback-Leibler (KL) divergence-based regularization with respect to a fixed reference policy is widely used in modern reinforcement learning to preserve the desired traits of the reference policy and sometimes to promote exploration (using uniform…

Do LLM Agents Have Regret? A Case Study in Online Learning and Games

arXiv:2403.16843v5 Announce Type: replace Abstract: Large language models (LLMs) have been increasingly employed for (interactive) decision-making, via the development of LLM-based autonomous agents. Despite their emerging successes, the performance of LLM agents in decision-making has not been fully investigated through…

NeuroRVQ: Multi-Scale EEG Tokenization for Generative Large Brainwave Models

arXiv:2510.13068v1 Announce Type: new Abstract: Electroencephalography (EEG) captures neural activity across multiple temporal and spectral scales, yielding signals that are rich but complex for representation learning. Recently, EEG foundation models trained to predict masked signal-tokens have shown promise for learning…

Random Scaling for Emergent Capabilities

arXiv:2502.17356v4 Announce Type: replace Abstract: Language models famously improve under a smooth scaling law, but some specific capabilities exhibit sudden breakthroughs in performance. While advocates of “emergence” view breakthroughs as unlocked capabilities, others attribute them to thresholding effects on noncontinuous…