Archives AI News

ProtoTTA: Prototype-Guided Test-Time Adaptation

arXiv:2604.15494v1 Announce Type: new Abstract: Deep networks that rely on prototypes-interpretable representations that can be related to the model input-have gained significant attention for balancing high accuracy with inherent interpretability, which makes them suitable for critical domains such as healthcare.…

ChemAmp: Amplified Chemistry Tools via Composable Agents

arXiv:2505.21569v3 Announce Type: replace Abstract: Although LLM-based agents are proven to master tool orchestration in scientific fields, particularly chemistry, their single-task performance remains limited by underlying tool constraints. To this end, we propose tool amplification, a novel paradigm that enhances…

Sentiment Analysis of German Sign Language Fairy Tales

arXiv:2604.16138v1 Announce Type: cross Abstract: We present a dataset and a model for sentiment analysis of German sign language (DGS) fairy tales. First, we perform sentiment analysis for three levels of valence (negative, neutral, positive) on German fairy tales text…

Optimizing Stochastic Gradient Push under Broadcast Communications

arXiv:2604.15549v1 Announce Type: new Abstract: We consider the problem of minimizing the convergence time for decentralized federated learning (DFL) in wireless networks under broadcast communications, with focus on mixing matrix design. The mixing matrix is a critical hyperparameter for DFL…

FSPO: Few-Shot Optimization of Synthetic Preferences Personalizes to Real Users

arXiv:2502.19312v2 Announce Type: replace Abstract: Effective personalization of LLMs is critical for a broad range of user-interfacing applications such as virtual assistants and content curation. Inspired by the strong in-context capabilities of LLMs, we propose few-shot preference optimization (FSPO), an…

Natural gradient descent with momentum

arXiv:2604.15554v1 Announce Type: new Abstract: We consider the problem of approximating a function by an element of a nonlinear manifold which admits a differentiable parametrization, typical examples being neural networks with differentiable activation functions or tensor networks. Natural gradient descent…