Archives AI News

CodeGEMM: A Codebook-Centric Approach to Efficient GEMM in Quantized LLMs

arXiv:2512.17970v1 Announce Type: new Abstract: Weight-only quantization is widely used to mitigate the memory-bound nature of LLM inference. Codebook-based methods extend this trend by achieving strong accuracy in the extremely low-bit regime (e.g., 2-bit). However, current kernels rely on dequantization,…

Convolutional-neural-operator-based transfer learning for solving PDEs

arXiv:2512.17969v1 Announce Type: new Abstract: Convolutional neural operator is a CNN-based architecture recently proposed to enforce structure-preserving continuous-discrete equivalence and enable the genuine, alias-free learning of solution operators of PDEs. This neural operator was demonstrated to outperform for certain cases…

MOORL: A Framework for Integrating Offline-Online Reinforcement Learning

arXiv:2506.09574v2 Announce Type: replace Abstract: Sample efficiency and exploration remain critical challenges in Deep Reinforcement Learning (DRL), particularly in complex domains. Offline RL, which enables agents to learn optimal policies from static, pre-collected datasets, has emerged as a promising alternative.…

GenUQ: Predictive Uncertainty Estimates via Generative Hyper-Networks

arXiv:2509.21605v2 Announce Type: replace Abstract: Operator learning is a recently developed generalization of regression to mappings between functions. It promises to drastically reduce expensive numerical integration of PDEs to fast evaluations of mappings between functional states of a system, i.e.,…

PEDESTRIAN: An Egocentric Vision Dataset for Obstacle Detection on Pavements

arXiv:2512.19190v1 Announce Type: cross Abstract: Walking has always been a primary mode of transportation and is recognized as an essential activity for maintaining good health. Despite the need for safe walking conditions in urban environments, sidewalks are frequently obstructed by…

Trajectory-Aware Eligibility Traces for Off-Policy Reinforcement Learning

arXiv:2301.11321v3 Announce Type: replace Abstract: Off-policy learning from multistep returns is crucial for sample-efficient reinforcement learning, but counteracting off-policy bias without exacerbating variance is challenging. Classically, off-policy bias is corrected in a per-decision manner: past temporal-difference errors are re-weighted by…