Disentangling Recall and Reasoning in Transformer Models through Layer-wise Attention and Activation Analysis
arXiv:2510.03366v2 Announce Type: replace Abstract: Transformer-based language models excel at both recall (retrieving memorized facts) and reasoning (performing multi-step inference), but whether these abilities rely on distinct internal mechanisms remains unclear. Distinguishing recall from reasoning is crucial for predicting model…
