Archives AI News

Finding Culture-Sensitive Neurons in Vision-Language Models

arXiv:2510.24942v2 Announce Type: replace Abstract: Despite their impressive performance, vision-language models (VLMs) still struggle on culturally situated inputs. To understand how VLMs process culturally grounded information, we study the presence of culture-sensitive neurons, i.e., neurons whose activations show preferential sensitivity…

Dimensional Criticality at Grokking Across MLPs and Transformers

arXiv:2604.16431v1 Announce Type: new Abstract: Abrupt transitions between distinct dynamical regimes are a hallmark of complex systems. Grokking in deep neural networks provides a striking example — an abrupt transition from memorization to generalization long after training accuracy saturates —…

Rate-Distortion Optimization for Transformer Inference

arXiv:2601.22002v3 Announce Type: replace Abstract: Transformers achieve superior performance on many tasks, but impose heavy compute and memory requirements during inference. This inference can be made more efficient by partitioning the process across multiple devices, which, in turn, requires compressing…

Instance-Adaptive Parametrization for Amortized Variational Inference

arXiv:2604.06796v2 Announce Type: replace Abstract: Variational autoencoders (VAEs) rely on amortized variational inference to enable efficient posterior approximation, but this efficiency comes at the cost of a shared parametrization, giving rise to the amortization gap. We propose the instance-adaptive variational…

Non-Stationarity in the Embedding Space of Time Series Foundation Models

arXiv:2604.16428v1 Announce Type: new Abstract: Time series foundation models (TSFMs) are widely used as generic feature extractors, yet the notion of non-stationarity in their embedding spaces remains poorly understood. Recent work often conflates non-stationarity with distribution shift, blurring distinctions fundamental…