Archives AI News

N-ReLU: Zero-Mean Stochastic Extension of ReLU

arXiv:2511.07559v1 Announce Type: new Abstract: Activation functions are fundamental for enabling nonlinear representations in deep neural networks. However, the standard rectified linear unit (ReLU) often suffers from inactive or “dead” neurons caused by its hard zero cutoff. To address this…

Hierarchical Deep Counterfactual Regret Minimization

arXiv:2305.17327v3 Announce Type: replace Abstract: Imperfect Information Games (IIGs) offer robust models for scenarios where decision-makers face uncertainty or lack complete information. Counterfactual Regret Minimization (CFR) has been one of the most successful family of algorithms for tackling IIGs. The…

SCALAR: Benchmarking SAE Interaction Sparsity in Toy LLMs

arXiv:2511.07572v1 Announce Type: new Abstract: Mechanistic interpretability aims to decompose neural networks into interpretable features and map their connecting circuits. The standard approach trains sparse autoencoders (SAEs) on each layer’s activations. However, SAEs trained in isolation don’t encourage sparse cross-layer…

LLM Output Drift: Cross-Provider Validation & Mitigation for Financial Workflows

arXiv:2511.07585v1 Announce Type: new Abstract: Financial institutions deploy Large Language Models (LLMs) for reconciliations, regulatory reporting, and client communications, but nondeterministic outputs (output drift) undermine auditability and trust. We quantify drift across five model architectures (7B-120B parameters) on regulated financial…

Evolutionary Profiles for Protein Fitness Prediction

arXiv:2510.07286v2 Announce Type: replace Abstract: Predicting the fitness impact of mutations is central to protein engineering but constrained by limited assays relative to the size of sequence space. Protein language models (pLMs) trained with masked language modeling (MLM) exhibit strong…

Partial Action Replacement: Tackling Distribution Shift in Offline MARL

arXiv:2511.07629v1 Announce Type: new Abstract: Offline multi-agent reinforcement learning (MARL) is severely hampered by the challenge of evaluating out-of-distribution (OOD) joint actions. Our core finding is that when the behavior policy is factorized – a common scenario where agents act…

Synergy over Discrepancy: A Partition-Based Approach to Multi-Domain LLM Fine-Tuning

arXiv:2511.07198v2 Announce Type: replace Abstract: Large language models (LLMs) demonstrate impressive generalization abilities, yet adapting them effectively across multiple heterogeneous domains remains challenging due to inter-domain interference. To overcome this challenge, we propose a partition-based multi-stage fine-tuning framework designed to…