Archives AI News

Flow Matching Neural Processes

arXiv:2512.23853v1 Announce Type: new Abstract: Neural processes (NPs) are a class of models that learn stochastic processes directly from data and can be used for inference, sampling and conditional sampling. We introduce a new NP model based on flow matching,…

Deep sequence models tend to memorize geometrically; it is unclear why

arXiv:2510.26745v2 Announce Type: replace Abstract: Deep sequence models are said to store atomic facts predominantly in the form of associative memory: a brute-force lookup of co-occurring entities. We identify a dramatically different form of storage of atomic facts that we…

Tazza: Shuffling Neural Network Parameters for Secure and Private Federated Learning

arXiv:2412.07454v3 Announce Type: replace Abstract: Federated learning enables decentralized model training without sharing raw data, preserving data privacy. However, its vulnerability towards critical security threats, such as gradient inversion and model poisoning by malicious clients, remain unresolved. Existing solutions often…

Learning Network Dismantling Without Handcrafted Inputs

arXiv:2508.00706v2 Announce Type: replace Abstract: The application of message-passing Graph Neural Networks has been a breakthrough for important network science problems. However, the competitive performance often relies on using handcrafted structural features as inputs, which increases computational cost and introduces…

Nonlinear Noise2Noise for Efficient Monte Carlo Denoiser Training

arXiv:2512.24794v1 Announce Type: cross Abstract: The Noise2Noise method allows for training machine learning-based denoisers with pairs of input and target images where both the input and target can be noisy. This removes the need for training with clean target images,…

Optimal Approximation — Smoothness Tradeoffs for Soft-Max Functions

arXiv:2010.11450v2 Announce Type: replace Abstract: A soft-max function has two main efficiency measures: (1) approximation – which corresponds to how well it approximates the maximum function, (2) smoothness – which shows how sensitive it is to changes of its input.…

Training Language Models to Explain Their Own Computations

arXiv:2511.08579v2 Announce Type: replace-cross Abstract: Can language models (LMs) learn to faithfully describe their internal computations? Are they better able to describe themselves than other models? We study the extent to which LMs’ privileged access to their own internals can…