Archives AI News

ShapeX: Shapelet-Driven Post Hoc Explanations for Time Series Classification Models

arXiv:2510.20084v1 Announce Type: new Abstract: Explaining time series classification models is crucial, particularly in high-stakes applications such as healthcare and finance, where transparency and trust play a critical role. Although numerous time series classification methods have identified key subsequences, known…

Hierarchical Dual-Head Model for Suicide Risk Assessment via MentalRoBERTa

arXiv:2510.20085v1 Announce Type: new Abstract: Social media platforms have become important sources for identifying suicide risk, but automated detection systems face multiple challenges including severe class imbalance, temporal complexity in posting patterns, and the dual nature of risk levels as…

Depth-Bounds for Neural Networks via the Braid Arrangement

arXiv:2502.09324v2 Announce Type: replace Abstract: We contribute towards resolving the open question of how many hidden layers are required in ReLU networks for exactly representing all continuous and piecewise linear functions on $mathbb{R}^d$. While the question has been resolved in…

MIRA: Medical Time Series Foundation Model for Real-World Health Data

arXiv:2506.07584v5 Announce Type: replace Abstract: A unified foundation model for medical time series — pretrained on open access and ethics board-approved medical corpora — offers the potential to reduce annotation burdens, minimize model customization, and enable robust transfer across clinical…

Don’t be lazy: CompleteP enables compute-efficient deep transformers

arXiv:2505.01618v3 Announce Type: replace Abstract: We study compute efficiency of LLM training when using different parameterizations, i.e., rules for adjusting model and optimizer hyperparameters (HPs) as model size changes. Some parameterizations fail to transfer optimal base HPs (such as learning…

Your Pre-trained LLM is Secretly an Unsupervised Confidence Calibrator

arXiv:2505.16690v4 Announce Type: replace Abstract: Post-training of large language models is essential for adapting pre-trained language models (PLMs) to align with human preferences and downstream tasks. While PLMs typically exhibit well-calibrated confidence, post-trained language models (PoLMs) often suffer from over-confidence,…

A Principle of Targeted Intervention for Multi-Agent Reinforcement Learning

arXiv:2510.17697v2 Announce Type: replace-cross Abstract: Steering cooperative multi-agent reinforcement learning (MARL) towards desired outcomes is challenging, particularly when the global guidance from a human on the whole multi-agent system is impractical in a large-scale MARL. On the other hand, designing…

Pre-training Epidemic Time Series Forecasters with Compartmental Prototypes

arXiv:2502.03393v5 Announce Type: replace Abstract: Accurate epidemic forecasting is crucial for outbreak preparedness, but existing data-driven models are often brittle. Typically trained on a single pathogen, they struggle with data scarcity during new outbreaks and fail under distribution shifts caused…