Archives AI News

SCALAR: Benchmarking SAE Interaction Sparsity in Toy LLMs

arXiv:2511.07572v1 Announce Type: new Abstract: Mechanistic interpretability aims to decompose neural networks into interpretable features and map their connecting circuits. The standard approach trains sparse autoencoders (SAEs) on each layer’s activations. However, SAEs trained in isolation don’t encourage sparse cross-layer…

LLM Output Drift: Cross-Provider Validation & Mitigation for Financial Workflows

arXiv:2511.07585v1 Announce Type: new Abstract: Financial institutions deploy Large Language Models (LLMs) for reconciliations, regulatory reporting, and client communications, but nondeterministic outputs (output drift) undermine auditability and trust. We quantify drift across five model architectures (7B-120B parameters) on regulated financial…

Evolutionary Profiles for Protein Fitness Prediction

arXiv:2510.07286v2 Announce Type: replace Abstract: Predicting the fitness impact of mutations is central to protein engineering but constrained by limited assays relative to the size of sequence space. Protein language models (pLMs) trained with masked language modeling (MLM) exhibit strong…

Partial Action Replacement: Tackling Distribution Shift in Offline MARL

arXiv:2511.07629v1 Announce Type: new Abstract: Offline multi-agent reinforcement learning (MARL) is severely hampered by the challenge of evaluating out-of-distribution (OOD) joint actions. Our core finding is that when the behavior policy is factorized – a common scenario where agents act…

Synergy over Discrepancy: A Partition-Based Approach to Multi-Domain LLM Fine-Tuning

arXiv:2511.07198v2 Announce Type: replace Abstract: Large language models (LLMs) demonstrate impressive generalization abilities, yet adapting them effectively across multiple heterogeneous domains remains challenging due to inter-domain interference. To overcome this challenge, we propose a partition-based multi-stage fine-tuning framework designed to…