Archives AI News

Exploring the Secondary Risks of Large Language Models

arXiv:2506.12382v3 Announce Type: replace Abstract: Ensuring the safety and alignment of Large Language Models is a significant challenge with their growing integration into critical applications and societal functions. While prior research has primarily focused on jailbreak attacks, less attention has…

What Makes a Reward Model a Good Teacher? An Optimization Perspective

arXiv:2503.15477v2 Announce Type: replace Abstract: The success of Reinforcement Learning from Human Feedback (RLHF) critically depends on the quality of the reward model. However, while this quality is primarily evaluated through accuracy, it remains unclear whether accuracy fully captures what…

Supervised Graph Contrastive Learning for Gene Regulatory Networks

arXiv:2505.17786v4 Announce Type: replace Abstract: Graph Contrastive Learning (GCL) is a powerful self-supervised learning framework that performs data augmentation through graph perturbations, with growing applications in the analysis of biological networks such as Gene Regulatory Networks (GRNs). The artificial perturbations…

Data-driven Neural Networks for Windkessel Parameter Calibration

arXiv:2509.21206v1 Announce Type: cross Abstract: In this work, we propose a novel method for calibrating Windkessel (WK) parameters in a dimensionally reduced 1D-0D coupled blood flow model. To this end, we design a data-driven neural network (NN)trained on simulated blood…

Myosotis: structured computation for attention like layer

arXiv:2509.20503v1 Announce Type: new Abstract: Attention layers apply a sequence-to-sequence mapping whose parameters depend on the pairwise interactions of the input elements. However, without any structural assumptions, memory and compute scale quadratically with the sequence length. The two main ways…

Offline Goal-conditioned Reinforcement Learning with Quasimetric Representations

arXiv:2509.20478v1 Announce Type: new Abstract: Approaches for goal-conditioned reinforcement learning (GCRL) often use learned state representations to extract goal-reaching policies. Two frameworks for representation structure have yielded particularly effective GCRL algorithms: (1) *contrastive representations*, in which methods learn “successor features”…