Archives AI News

Causal-EPIG: A Prediction-Oriented Active Learning Framework for CATE Estimation

arXiv:2509.21866v1 Announce Type: new Abstract: Estimating the Conditional Average Treatment Effect (CATE) is often constrained by the high cost of obtaining outcome measurements, making active learning essential. However, conventional active learning strategies suffer from a fundamental objective mismatch. They are…

SADA: Safe and Adaptive Inference with Multiple Black-Box Predictions

arXiv:2509.21707v1 Announce Type: new Abstract: Real-world applications often face scarce labeled data due to the high cost and time requirements of gold-standard experiments, whereas unlabeled data are typically abundant. With the growing adoption of machine learning techniques, it has become…

Effective continuous equations for adaptive SGD: a stochastic analysis view

arXiv:2509.21614v1 Announce Type: new Abstract: We present a theoretical analysis of some popular adaptive Stochastic Gradient Descent (SGD) methods in the small learning rate regime. Using the stochastic modified equations framework introduced by Li et al., we derive effective continuous…

IndiSeek learns information-guided disentangled representations

arXiv:2509.21584v1 Announce Type: new Abstract: Learning disentangled representations is a fundamental task in multi-modal learning. In modern applications such as single-cell multi-omics, both shared and modality-specific features are critical for characterizing cell states and supporting downstream analyses. Ideally, modality-specific features…

A Nonparametric Discrete Hawkes Model with a Collapsed Gaussian-Process Prior

arXiv:2509.21996v1 Announce Type: new Abstract: Hawkes process models are used in settings where past events increase the likelihood of future events occurring. Many applications record events as counts on a regular grid, yet discrete-time Hawkes models remain comparatively underused and…

Tricks and Plug-ins for Gradient Boosting with Transformers

arXiv:2508.02924v3 Announce Type: replace-cross Abstract: Transformer architectures dominate modern NLP but often demand heavy computational resources and intricate hyperparameter tuning. To mitigate these challenges, we propose a novel framework, BoostTransformer, that augments transformers with boosting principles through subgrid token selection…

SHAKE-GNN: Scalable Hierarchical Kirchhoff-Forest Graph Neural Network

arXiv:2509.22100v1 Announce Type: cross Abstract: Graph Neural Networks (GNNs) have achieved remarkable success across a range of learning tasks. However, scaling GNNs to large graphs remains a significant challenge, especially for graph-level tasks. In this work, we introduce SHAKE-GNN, a…