Archives AI News

The Path Not Taken: Duality in Reasoning about Program Execution

arXiv:2604.20917v1 Announce Type: new Abstract: Large language models (LLMs) have shown remarkable capabilities across diverse coding tasks. However, their adoption requires a true understanding of program execution rather than relying on surface-level patterns. Existing benchmarks primarily focus on predicting program…

Absorber LLM: Harnessing Causal Synchronization for Test-Time Training

arXiv:2604.20915v1 Announce Type: new Abstract: Transformers suffer from a high computational cost that grows with sequence length for self-attention, making inference in long streams prohibited by memory consumption. Constant-memory alternatives such as RNNs and SSMs compress history into states with…

FairyFuse: Multiplication-Free LLM Inference on CPUs via Fused Ternary Kernels

arXiv:2604.20913v1 Announce Type: new Abstract: Large language models are increasingly deployed on CPU-only platforms where memory bandwidth is the primary bottleneck for autoregressive generation. Weight quantization to four bits or below reduces memory pressure, yet existing systems still dequantize weights…

Fixation Sequences as Time Series: A Topological Approach to Dyslexia Detection

arXiv:2604.21698v1 Announce Type: cross Abstract: Persistent homology, a method from topological data analysis, extracts robust, multi-scale features from data. It produces stable representations of time series by applying varying thresholds to their values (a process known as a textit{filtration}). We…

ILDR: Geometric Early Detection of Grokking

arXiv:2604.20923v1 Announce Type: new Abstract: Grokking describes a delayed generalization phenomenon in which a neural network achieves perfect training accuracy long before validation accuracy improves, followed by an abrupt transition to strong generalization. Existing detection signals are indirect: weight norm…