Archives AI News

Generating Full-field Evolution of Physical Dynamics from Irregular Sparse Observations

arXiv:2505.09284v3 Announce Type: replace-cross Abstract: Modeling and reconstructing multidimensional physical dynamics from sparse and off-grid observations presents a fundamental challenge in scientific research. Recently, diffusion-based generative modeling shows promising potential for physical simulation. However, current approaches typically operate on on-grid…

A Greedy PDE Router for Blending Neural Operators and Classical Methods

arXiv:2509.24814v1 Announce Type: cross Abstract: When solving PDEs, classical numerical solvers are often computationally expensive, while machine learning methods can suffer from spectral bias, failing to capture high-frequency components. Designing an optimal hybrid iterative solver–where, at each iteration, a solver…

A quantitative Robbins-Siegmund theorem

arXiv:2410.15986v2 Announce Type: replace-cross Abstract: The Robbins-Siegmund theorem is one of the most important results in stochastic optimization, where it is widely used to prove the convergence of stochastic algorithms. We provide a quantitative version of the theorem, establishing a…

Leveraging Coordinate Momentum in SignSGD and Muon: Memory-Optimized Zero-Order

arXiv:2506.04430v3 Announce Type: replace Abstract: Fine-tuning Large Language Models (LLMs) is essential for adapting pre-trained models to downstream tasks. Yet traditional first-order optimizers such as Stochastic Gradient Descent (SGD) and Adam incur prohibitive memory and computational costs that scale poorly…

Type-Compliant Adaptation Cascades: Adapting Programmatic LM Workflows to Data

arXiv:2508.18244v2 Announce Type: replace Abstract: Reliably composing Large Language Models (LLMs) for complex, multi-step workflows remains a significant challenge. The dominant paradigm — optimizing discrete prompts in a pipeline — is notoriously brittle and struggles to enforce the formal compliance…

Neighborhood Sampling Does Not Learn the Same Graph Neural Network

arXiv:2509.22868v1 Announce Type: new Abstract: Neighborhood sampling is an important ingredient in the training of large-scale graph neural networks. It suppresses the exponential growth of the neighborhood size across network layers and maintains feasible memory consumption and time costs. While…

Meta-Learning to Explore via Memory Density Feedback

arXiv:2503.02831v2 Announce Type: replace Abstract: Exploration algorithms for reinforcement learning typically replace or augment the reward function with an additional “intrinsic” reward that trains the agent to seek previously unseen states of the environment. Here, we consider an exploration algorithm…

Adaptive Margin RLHF via Preference over Preferences

arXiv:2509.22851v1 Announce Type: new Abstract: Margin-based optimization is fundamental to improving generalization and robustness in classification tasks. In the context of reward model learning from preferences within Reinforcement Learning from Human Feedback (RLHF), existing methods typically rely on no margins,…