Archives AI News

RL in the Wild: Characterizing RLVR Training in LLM Deployment

arXiv:2509.25279v1 Announce Type: new Abstract: Large Language Models (LLMs) are now widely used across many domains. With their rapid development, Reinforcement Learning with Verifiable Rewards (RLVR) has surged in recent months to enhance their reasoning and understanding abilities. However, its…

Language Model Planning from an Information Theoretic Perspective

arXiv:2509.25260v1 Announce Type: new Abstract: The extent to which decoder-only language models (LMs) engage in planning, that is, organizing intermediate computations to support coherent long-range generation, remains an open and important question, with implications for interpretability, reliability, and principled model…

Memory Management and Contextual Consistency for Long-Running Low-Code Agents

arXiv:2509.25250v1 Announce Type: new Abstract: The rise of AI-native Low-Code/No-Code (LCNC) platforms enables autonomous agents capable of executing complex, long-duration business processes. However, a fundamental challenge remains: memory management. As agents operate over extended periods, they face “memory inflation” and…

A Formal Comparison Between Chain-of-Thought and Latent Thought

arXiv:2509.25239v1 Announce Type: new Abstract: Chain-of-Thought (CoT) elicits reasoning in large language models by explicitly generating intermediate steps in natural language. In contrast, Latent Thought in looped models operates directly in the continuous latent space, enabling computation beyond discrete linguistic…

Muon Outperforms Adam in Tail-End Associative Memory Learning

arXiv:2509.26030v1 Announce Type: cross Abstract: The Muon optimizer is consistently faster than Adam in training Large Language Models (LLMs), yet the mechanism underlying its success remains unclear. This paper demystifies this mechanism through the lens of associative memory. By ablating…

Toward Causal-Visual Programming: Enhancing Agentic Reasoning in Low-Code Environments

arXiv:2509.25282v1 Announce Type: new Abstract: Large language model (LLM) agents are increasingly capable of orchestrating complex tasks in low-code environments. However, these agents often exhibit hallucinations and logical inconsistencies because their inherent reasoning mechanisms rely on probabilistic associations rather than…