Archives AI News

Support Vector Machine Kernels as Quantum Propagators

arXiv:2502.11153v3 Announce Type: replace-cross Abstract: Selecting optimal kernels for regression in physical systems remains a challenge, often relying on trial-and-error with standard functions. In this work, we establish a mathematical correspondence between support vector machine kernels and quantum propagators, demonstrating…

The Curse of Depth in Large Language Models

arXiv:2502.05795v3 Announce Type: replace Abstract: In this paper, we introduce the Curse of Depth, a concept that highlights, explains, and addresses the recent observation in modern Large Language Models (LLMs) where nearly half of the layers are less effective than…

Flattening Hierarchies with Policy Bootstrapping

arXiv:2505.14975v3 Announce Type: replace Abstract: Offline goal-conditioned reinforcement learning (GCRL) is a promising approach for pretraining generalist policies on large datasets of reward-free trajectories, akin to the self-supervised objectives used to train foundation models for computer vision and natural language…

Generative Conditional Missing Imputation Networks

arXiv:2601.00517v1 Announce Type: cross Abstract: In this study, we introduce a sophisticated generative conditional strategy designed to impute missing values within datasets, an area of considerable importance in statistical analysis. Specifically, we initially elucidate the theoretical underpinnings of the Generative…

Support Vector Machine Kernels as Quantum Propagators

arXiv:2502.11153v3 Announce Type: replace-cross Abstract: Selecting optimal kernels for regression in physical systems remains a challenge, often relying on trial-and-error with standard functions. In this work, we establish a mathematical correspondence between support vector machine kernels and quantum propagators, demonstrating…

The Curse of Depth in Large Language Models

arXiv:2502.05795v3 Announce Type: replace Abstract: In this paper, we introduce the Curse of Depth, a concept that highlights, explains, and addresses the recent observation in modern Large Language Models (LLMs) where nearly half of the layers are less effective than…