Sampling for Quality: Training-Free Reward-Guided LLM Decoding via Sequential Monte Carlo
arXiv:2604.16453v1 Announce Type: new Abstract: We introduce a principled probabilistic framework for reward-guided decoding in large language models, addressing the limitations of standard decoding methods that optimize token-level likelihood rather than sequence-level quality. Our method defines a reward-augmented target distribution…
