Archives AI News

Thompson Sampling via Fine-Tuning of LLMs

arXiv:2510.13328v2 Announce Type: replace Abstract: Bayesian optimization in large unstructured discrete spaces is often hindered by the computational cost of maximizing acquisition functions due to the absence of gradients. We propose a scalable alternative based on Thompson sampling that eliminates…

LTR-ICD: A Learning-to-Rank Approach for Automatic ICD Coding

arXiv:2510.13922v1 Announce Type: new Abstract: Clinical notes contain unstructured text provided by clinicians during patient encounters. These notes are usually accompanied by a sequence of diagnostic codes following the International Classification of Diseases (ICD). Correctly assigning and ordering ICD codes…

Uncertainty Quantification with the Empirical Neural Tangent Kernel

arXiv:2502.02870v2 Announce Type: replace-cross Abstract: While neural networks have demonstrated impressive performance across various tasks, accurately quantifying uncertainty in their predictions is essential to ensure their trustworthiness and enable widespread adoption in critical systems. Several Bayesian uncertainty quantification (UQ) methods…

Distributional Consistency Loss: Beyond Pointwise Data Terms in Inverse Problems

arXiv:2510.13972v1 Announce Type: new Abstract: Recovering true signals from noisy measurements is a central challenge in inverse problems spanning medical imaging, geophysics, and signal processing. Current solutions balance prior assumptions regarding the true signal (regularization) with agreement to noisy measured…

LLM-guided Chemical Process Optimization with a Multi-Agent Approach

arXiv:2506.20921v2 Announce Type: replace Abstract: Chemical process optimization maximizes production efficiency and economic performance, but optimization algorithms, including gradient-based solvers, numerical methods, and parameter grid searches, become impractical when operating constraints are ill-defined or unavailable. We present a multi-agent LLM…

Beyond Linear Probes: Dynamic Safety Monitoring for Language Models

arXiv:2509.26238v2 Announce Type: replace Abstract: Monitoring large language models’ (LLMs) activations is an effective way to detect harmful requests before they lead to unsafe outputs. However, traditional safety monitors often require the same amount of compute for every query. This…