Archives AI News

A Partition Cover Approach to Tokenization

arXiv:2501.06246v3 Announce Type: replace-cross Abstract: Tokenization is the process of encoding strings into tokens of a fixed vocabulary size, and is widely utilized in Natural Language Processing applications. The leading tokenization algorithm today is Byte-Pair Encoding (BPE), which formulates the…

On The Sample Complexity Bounds In Bilevel Reinforcement Learning

arXiv:2503.17644v4 Announce Type: replace-cross Abstract: Bilevel reinforcement learning (BRL) has emerged as a powerful framework for aligning generative models, yet its theoretical foundations, especially sample complexity bounds, remain underexplored. In this work, we present the first sample complexity bound for…

AutoEP: LLMs-Driven Automation of Hyperparameter Evolution for Metaheuristic Algorithms

arXiv:2509.23189v1 Announce Type: new Abstract: Dynamically configuring algorithm hyperparameters is a fundamental challenge in computational intelligence. While learning-based methods offer automation, they suffer from prohibitive sample complexity and poor generalization. We introduce AutoEP, a novel framework that bypasses training entirely…

$p$-less Sampling: A Robust Hyperparameter-Free Approach for LLM Decoding

arXiv:2509.23234v1 Announce Type: new Abstract: Obtaining high-quality outputs from Large Language Models (LLMs) often depends upon the choice of a sampling-based decoding strategy to probabilistically choose the next token at each generation step. While a variety of such sampling methods…

Communication-Efficient Desire Alignment for Embodied Agent-Human Adaptation

arXiv:2505.22503v2 Announce Type: replace-cross Abstract: While embodied agents have made significant progress in performing complex physical tasks, real-world applications demand more than pure task execution. The agents must collaborate with unfamiliar agents and human users, whose goals are often vague…

GRAF: Multi-turn Jailbreaking via Global Refinement and Active Fabrication

arXiv:2506.17881v2 Announce Type: replace-cross Abstract: Large Language Models (LLMs) have demonstrated remarkable performance across diverse tasks. Nevertheless, they still pose notable safety risks due to potential misuse for malicious purposes. Jailbreaking, which seeks to induce models to generate harmful content…