<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Reinforcement Learning | Hejie Cui</title><link>https://hejiecui.com/tag/reinforcement-learning/</link><atom:link href="https://hejiecui.com/tag/reinforcement-learning/index.xml" rel="self" type="application/rss+xml"/><description>Reinforcement Learning</description><generator>Source Themes Academic (https://sourcethemes.com/academic/)</generator><language>en-us</language><lastBuildDate>Thu, 30 Apr 2026 01:00:00 +0000</lastBuildDate><image><url>https://hejiecui.com/images/logo_hu55b4809d0d762654adf09f4071918d91_2611179_300x300_fit_lanczos_2.png</url><title>Reinforcement Learning</title><link>https://hejiecui.com/tag/reinforcement-learning/</link></image><item><title>T²PO: Uncertainty-Guided Exploration Control for Stable Multi-Turn Agentic Reinforcement Learning</title><link>https://hejiecui.com/publication/t2po/</link><pubDate>Thu, 30 Apr 2026 01:00:00 +0000</pubDate><guid>https://hejiecui.com/publication/t2po/</guid><description>&lt;p>Recent progress in multi-turn reinforcement learning (RL) has significantly improved reasoning LLMs&amp;rsquo; performances on complex interactive tasks. Despite advances in stabilization techniques such as fine-grained credit assignment and trajectory filtering, instability remains pervasive and often leads to training collapse. We argue that this instability stems from inefficient exploration in multi-turn settings, where policies continue to generate low-information actions that neither reduce uncertainty nor advance task progress. To address this issue, we propose Token- and Turn-level Policy Optimization (T²PO), an uncertainty-aware framework that explicitly controls exploration at fine-grained levels. At the token level, T²PO monitors uncertainty dynamics and triggers a thinking intervention once the marginal uncertainty change falls below a threshold. At the turn level, T²PO identifies interactions with negligible exploration progress and dynamically resamples such turns to avoid wasted rollouts. We evaluate T²PO in diverse environments, including WebShop, ALFWorld, and Search QA, demonstrating substantial gains in training stability and performance improvements with better exploration efficiency.&lt;/p>
&lt;p>&lt;sup>*&lt;/sup> Corresponding author.&lt;/p></description></item></channel></rss>