Archives AI News

GPO: Learning from Critical Steps to Improve LLM Reasoning

arXiv:2509.16456v1 Announce Type: new Abstract: Large language models (LLMs) are increasingly used in various domains, showing impressive potential on different tasks. Recently, reasoning LLMs have been proposed to improve the textit{reasoning} or textit{thinking} capabilities of LLMs to solve complex problems.…

Full-History Graphs with Edge-Type Decoupled Networks for Temporal Reasoning

arXiv:2508.03251v2 Announce Type: replace Abstract: Modeling evolving interactions among entities is critical in many real-world tasks. For example, predicting driver maneuvers in traffic requires tracking how neighboring vehicles accelerate, brake, and change lanes relative to one another over consecutive frames.…

Checking extracted rules in Neural Networks

arXiv:2509.16547v1 Announce Type: new Abstract: In this paper we investigate formal verification of extracted rules for Neural Networks under a complexity theoretic point of view. A rule is a global property or a pattern concerning a large portion of the…

Bayesian scaling laws for in-context learning

arXiv:2410.16531v4 Announce Type: replace-cross Abstract: In-context learning (ICL) is a powerful technique for getting language models to perform complex tasks with no training updates. Prior work has established strong correlations between the number of in-context examples provided and the accuracy…

SalaMAnder: Shapley-based Mathematical Expression Attribution and Metric for Chain-of-Thought Reasoning

arXiv:2509.16561v1 Announce Type: new Abstract: Chain-of-Thought (CoT) prompting enhances the math reasoning capability of large language models (LLMs) to a large margin. However, the mechanism underlying such improvements remains unexplored. In this paper, we present textbf{SalaMAnder} (textbf{S}htextbf{a}ptextbf{l}ey-btextbf{a}sed textbf{M}athematical Expression textbf{A}ttribution…

Can Language Models Follow Multiple Turns of Entangled Instructions?

arXiv:2503.13222v3 Announce Type: replace-cross Abstract: Despite significant achievements in improving the instruction-following capabilities of large language models (LLMs), the ability to process multiple potentially entangled or conflicting instructions remains a considerable challenge. Real-world scenarios often require consistency across multiple instructions…

Question Answering with LLMs and Learning from Answer Sets

arXiv:2509.16590v1 Announce Type: new Abstract: Large Language Models (LLMs) excel at understanding natural language but struggle with explicit commonsense reasoning. A recent trend of research suggests that the combination of LLM with robust symbolic reasoning systems can overcome this problem…