Archives AI News

Collaborative Unlabeled Data Optimization

arXiv:2505.14117v2 Announce Type: replace Abstract: This paper pioneers a novel data-centric paradigm to maximize the utility of unlabeled data, tackling a critical question: How can we enhance the efficiency and sustainability of deep learning training by optimizing the data itself?…

Graph Diffusion Transformers are In-Context Molecular Designers

arXiv:2510.08744v1 Announce Type: new Abstract: In-context learning allows large models to adapt to new tasks from a few demonstrations, but it has shown limited success in molecular design. Existing databases such as ChEMBL contain molecular properties spanning millions of biological…

RFOD: Random Forest-based Outlier Detection for Tabular Data

arXiv:2510.08747v1 Announce Type: new Abstract: Outlier detection in tabular data is crucial for safeguarding data integrity in high-stakes domains such as cybersecurity, financial fraud detection, and healthcare, where anomalies can cause serious operational and economic impacts. Despite advances in both…

Conformal Risk Training: End-to-End Optimization of Conformal Risk Control

arXiv:2510.08748v1 Announce Type: new Abstract: While deep learning models often achieve high predictive accuracy, their predictions typically do not come with any provable guarantees on risk or reliability, which are critical for deployment in high-stakes applications. The framework of conformal…

Untangling Component Imbalance in Hybrid Linear Attention Conversion Methods

arXiv:2510.05901v2 Announce Type: replace Abstract: Transformers’ quadratic computational complexity limits their scalability despite remarkable performance. While linear attention reduces this to linear complexity, pre-training such models from scratch remains, in most cases, prohibitively expensive. Recent post-training linearisation methods convert pre-trained…

LOTION: Smoothing the Optimization Landscape for Quantized Training

arXiv:2510.08757v1 Announce Type: new Abstract: Optimizing neural networks for quantized objectives is fundamentally challenging because the quantizer is piece-wise constant, yielding zero gradients everywhere except at quantization thresholds where the derivative is undefined. Most existing methods deal with this issue…