Archives AI News

Mind the Prompt: Self-adaptive Generation of Task Plan Explanations via LLMs

arXiv:2604.21092v1 Announce Type: new Abstract: Integrating Large Language Models (LLMs) into complex software systems enables the generation of human-understandable explanations of opaque AI processes, such as automated task planning. However, the quality and reliability of these explanations heavily depend on…

RIFT: Repurposing Negative Samples via Reward-Informed Fine-Tuning

arXiv:2601.09253v2 Announce Type: replace-cross Abstract: While Supervised Fine-Tuning (SFT) and Rejection Sampling Fine-Tuning (RFT) are standard for LLM alignment, they either rely on costly expert data or discard valuable negative samples, leading to data inefficiency. To address this, we propose…

Propensity Inference: Environmental Contributors to LLM Behaviour

arXiv:2604.21098v1 Announce Type: new Abstract: Motivated by loss of control risks from misaligned AI systems, we develop and apply methods for measuring language models’ propensity for unsanctioned behaviour. We contribute three methodological improvements: analysing effects of changes to environmental factors…

Mind the Prompt: Self-adaptive Generation of Task Plan Explanations via LLMs

arXiv:2604.21092v1 Announce Type: new Abstract: Integrating Large Language Models (LLMs) into complex software systems enables the generation of human-understandable explanations of opaque AI processes, such as automated task planning. However, the quality and reliability of these explanations heavily depend on…