Language Models Learn Universal Representations of Numbers and Here’s Why You Should Care
arXiv:2510.26285v2 Announce Type: replace-cross Abstract: Prior work has shown that large language models (LLMs) often converge to accurate input embedding for numbers, based on sinusoidal representations. In this work, we quantify that these representations are in fact strikingly systematic, to…
