Archives AI News

Operationalizing Quantized Disentanglement

arXiv:2511.20927v1 Announce Type: new Abstract: Recent theoretical work established the unsupervised identifiability of quantized factors under any diffeomorphism. The theory assumes that quantization thresholds correspond to axis-aligned discontinuities in the probability density of the latent factors. By constraining a learned…

No Request Left Behind: Tackling Heterogeneity in Long-Context LLM Inference with Medha

arXiv:2409.17264v5 Announce Type: replace Abstract: Deploying million-token Large Language Models (LLMs) is challenging because production workloads are highly heterogeneous, mixing short queries and long documents. This heterogeneity, combined with the quadratic complexity of attention, creates severe convoy effects where long-running…

F-INR: Functional Tensor Decomposition for Implicit Neural Representations

arXiv:2503.21507v2 Announce Type: replace Abstract: Implicit Neural Representations (INRs) model signals as continuous, differentiable functions. However, monolithic INRs scale poorly with data dimensionality, leading to excessive training costs. We propose F-INR, a framework that addresses this limitation by factorizing a…

Dataset Poisoning Attacks on Behavioral Cloning Policies

arXiv:2511.20992v1 Announce Type: new Abstract: Behavior Cloning (BC) is a popular framework for training sequential decision policies from expert demonstrations via supervised learning. As these policies are increasingly being deployed in the real world, their robustness and potential vulnerabilities are…

Inference-Time Alignment of Diffusion Models via Evolutionary Algorithms

arXiv:2506.00299v2 Announce Type: replace Abstract: Diffusion models are state-of-the-art generative models, yet their samples often fail to satisfy application objectives such as safety constraints or domain-specific validity. Existing techniques for alignment require gradients, internal model access, or large computational budgets…

Subgoal Graph-Augmented Planning for LLM-Guided Open-World Reinforcement Learning

arXiv:2511.20993v1 Announce Type: new Abstract: Large language models (LLMs) offer strong high-level planning capabilities for reinforcement learning (RL) by decomposing tasks into subgoals. However, their practical utility is limited by poor planning-execution alignment, which reflects a critical gap between abstract…

A Gray-box Attack against Latent Diffusion Model-based Image Editing by Posterior Collapse

arXiv:2408.10901v4 Announce Type: replace-cross Abstract: Recent advancements in Latent Diffusion Models (LDMs) have revolutionized image synthesis and manipulation, raising significant concerns about data misappropriation and intellectual property infringement. While adversarial attacks have been extensively explored as a protective measure against…

QiMeng-SALV: Signal-Aware Learning for Verilog Code Generation

arXiv:2510.19296v3 Announce Type: replace Abstract: The remarkable progress of Large Language Models (LLMs) presents promising opportunities for Verilog code generation which is significantly important for automated circuit design. The lacking of meaningful functional rewards hinders the preference optimization based on…

scipy.spatial.transform: Differentiable Framework-Agnostic 3D Transformations in Python

arXiv:2511.18157v2 Announce Type: replace Abstract: Three-dimensional rigid-body transforms, i.e. rotations and translations, are central to modern differentiable machine learning pipelines in robotics, vision, and simulation. However, numerically robust and mathematically correct implementations, particularly on SO(3), are error-prone due to issues…