Ratio-Variance Regularized Policy Optimization for Efficient LLM Fine-tuning
arXiv:2601.03320v1 Announce Type: new Abstract: On-policy reinforcement learning (RL), particularly Proximal Policy Optimization (PPO) and Group Relative Policy Optimization (GRPO), has become the dominant paradigm for fine-tuning large language models (LLMs). While policy ratio clipping stabilizes training, this heuristic hard…
