Trajectory-Aware Eligibility Traces for Off-Policy Reinforcement Learning
arXiv:2301.11321v3 Announce Type: replace Abstract: Off-policy learning from multistep returns is crucial for sample-efficient reinforcement learning, but counteracting off-policy bias without exacerbating variance is challenging. Classically, off-policy bias is corrected in a per-decision manner: past temporal-difference errors are re-weighted by…
