Archives AI News

PTQTP: Post-Training Quantization to Trit-Planes for Large Language Models

arXiv:2509.16989v3 Announce Type: replace Abstract: Post-training quantization (PTQ) of large language models (LLMs) to extremely low bit-widths remains challenging due to the fundamental trade-off between computational efficiency and representational capacity. While existing ultra-low-bit methods rely on binary approximations or quantization-aware…

Causality-Inspired Safe Residual Correction for Multivariate Time Series

arXiv:2512.22428v2 Announce Type: replace Abstract: While modern multivariate forecasters such as Transformers and GNNs achieve strong benchmark performance, they often suffer from systematic errors at specific variables or horizons and, critically, lack guarantees against performance degradation in deployment. Existing post-hoc…

MCD: Marginal Contrastive Discrimination for conditional density estimation

arXiv:2206.01592v2 Announce Type: replace-cross Abstract: We consider the problem of conditional density estimation, which is a major topic of interest in the fields of statistical and machine learning. Our method, called Marginal Contrastive Discrimination, MCD, reformulates the conditional density function…

Can Optimal Transport Improve Federated Inverse Reinforcement Learning?

arXiv:2601.00309v1 Announce Type: new Abstract: In robotics and multi-agent systems, fleets of autonomous agents often operate in subtly different environments while pursuing a common high-level objective. Directly pooling their data to learn a shared reward function is typically impractical due…

Mitigating optimistic bias in entropic risk estimation and optimization

arXiv:2409.19926v4 Announce Type: replace-cross Abstract: The entropic risk measure is widely used in high-stakes decision-making across economics, management science, finance, and safety-critical control systems because it captures tail risks associated with uncertain losses. However, when data are limited, the empirical…

The Curse of Depth in Large Language Models

arXiv:2502.05795v3 Announce Type: replace Abstract: In this paper, we introduce the Curse of Depth, a concept that highlights, explains, and addresses the recent observation in modern Large Language Models (LLMs) where nearly half of the layers are less effective than…

Flattening Hierarchies with Policy Bootstrapping

arXiv:2505.14975v3 Announce Type: replace Abstract: Offline goal-conditioned reinforcement learning (GCRL) is a promising approach for pretraining generalist policies on large datasets of reward-free trajectories, akin to the self-supervised objectives used to train foundation models for computer vision and natural language…