Archives AI News

Training Language Models to Explain Their Own Computations

arXiv:2511.08579v2 Announce Type: replace-cross Abstract: Can language models (LMs) learn to faithfully describe their internal computations? Are they better able to describe themselves than other models? We study the extent to which LMs’ privileged access to their own internals can…

HINTS: Extraction of Human Insights from Time-Series Without External Sources

arXiv:2512.23755v1 Announce Type: new Abstract: Human decision-making, emotions, and collective psychology are complex factors that shape the temporal dynamics observed in financial and economic systems. Many recent time series forecasting models leverage external sources (e.g., news and social media) to…

Geometric Scaling of Bayesian Inference in LLMs

arXiv:2512.23752v1 Announce Type: new Abstract: Recent work has shown that small transformers trained in controlled “wind-tunnel” settings can implement exact Bayesian inference, and that their training dynamics produce a geometric substrate — low-dimensional value manifolds and progressively orthogonal keys —…

NeuroPMD: Neural Fields for Density Estimation on Product Manifolds

arXiv:2501.02994v2 Announce Type: replace-cross Abstract: We propose a novel deep neural network methodology for density estimation on product Riemannian manifold domains. In our approach, the network directly parameterizes the unknown density function and is trained using a penalized maximum likelihood…

Drift-Based Dataset Stability Benchmark

arXiv:2512.23762v1 Announce Type: new Abstract: Machine learning (ML) represents an efficient and popular approach for network traffic classification. However, network traffic classification is a challenging domain, and trained models may degrade soon after deployment due to the obsolete datasets and…

PERK: Long-Context Reasoning as Parameter-Efficient Test-Time Learning

arXiv:2507.06415v2 Announce Type: replace-cross Abstract: Long-context reasoning requires accurately identifying relevant information in extensive, noisy input contexts. Previous research shows that using test-time learning to encode context directly into model parameters can effectively enable reasoning over noisy information. However, meta-learning…