Archives AI News

Compute-Optimal Quantization-Aware Training

arXiv:2509.22935v1 Announce Type: new Abstract: Quantization-aware training (QAT) is a leading technique for improving the accuracy of quantized neural networks. Previous work has shown that decomposing training into a full-precision (FP) phase followed by a QAT phase yields superior accuracy…

Understanding SOAP from the Perspective of Gradient Whitening

arXiv:2509.22938v1 Announce Type: new Abstract: Shampoo with Adam in the Preconditioner’s eigenbasis (SOAP) has recently emerged as a promising optimization algorithm for neural network training, achieving superior training efficiency over both Adam and Shampoo in language modeling tasks. In this…

LVT: Large-Scale Scene Reconstruction via Local View Transformers

arXiv:2509.25001v1 Announce Type: cross Abstract: Large transformer models are proving to be a powerful tool for 3D vision and novel view synthesis. However, the standard Transformer’s well-known quadratic complexity makes it difficult to scale these methods to large scenes. To…