Archives AI News

Denoising the US Census: Succinct Block Hierarchical Regression

arXiv:2603.10099v1 Announce Type: new Abstract: The US Census Bureau Disclosure Avoidance System (DAS) balances confidentiality and utility requirements for the decennial US Census (Abowd et al., 2022). The DAS was used in the 2020 Census to produce demographic datasets critically…

Self-Scaled Broyden Family of Quasi-Newton Methods in JAX

arXiv:2603.10599v1 Announce Type: cross Abstract: We present a JAX implementation of the Self-Scaled Broyden family of quasi-Newton methods, fully compatible with JAX and building on the Optimistix~cite{rader_optimistix_2024} optimisation library. The implementation includes BFGS, DFP, Broyden and their Self-Scaled variants(SSBFGS, SSDFP,…

Hardware Efficient Approximate Convolution with Tunable Error Tolerance for CNNs

arXiv:2603.10100v1 Announce Type: new Abstract: Modern CNNs’ high computational demands hinder edge deployment, as traditional “hard” sparsity (skipping mathematical zeros) loses effectiveness in deep layers or with smooth activations like Tanh. We propose a “soft sparsity” paradigm using a hardware…

CLIPO: Contrastive Learning in Policy Optimization Generalizes RLVR

arXiv:2603.10101v1 Announce Type: new Abstract: Reinforcement Learning with Verifiable Rewards (RLVR) has significantly advanced the reasoning capacity of Large Language Models (LLMs). However, RLVR solely relies on final answers as outcome rewards, neglecting the correctness of intermediate reasoning steps. Training…

V2M-Zero: Zero-Pair Time-Aligned Video-to-Music Generation

arXiv:2603.11042v1 Announce Type: cross Abstract: Generating music that temporally aligns with video events is challenging for existing text-to-music models, which lack fine-grained temporal control. We introduce V2M-Zero, a zero-pair video-to-music generation approach that outputs time-aligned music for video. Our method…

Mamba Neural Operator: Who Wins? Transformers vs. State-Space Models for PDEs

arXiv:2410.02113v3 Announce Type: replace Abstract: Partial differential equations (PDEs) are widely used to model complex physical systems, but solving them efficiently remains a significant challenge. Recently, Transformers have emerged as the preferred architecture for PDEs due to their ability to…

Losing dimensions: Geometric memorization in generative diffusion

arXiv:2410.08727v2 Announce Type: replace-cross Abstract: Diffusion models power leading generative AI, but when and how they memorize training data, especially on low-dimensional manifolds, remains unclear. We find memorization emerges gradually, not abruptly: as data become scarce, diffusion models experience a…