Reinforcement learning with verifiable rewards faces a major hurdle in its exploration capabilities, as policies tend to converge on a limited set of solutions. To address this issue, researchers have traditionally relied on entropy regularization to encourage exploration, but this approach has proven inconsistent. A new method, bidirectional entropy modulation, offers a refinement over traditional entropy regularization, allowing for more effective exploration in reinforcement learning with verifiable rewards1. This development has significant implications for large language models, which have seen substantial advancements in reasoning capabilities thanks to RLVR. As these models continue to evolve, their ability to explore and adapt will be crucial in determining their overall capability and potential risks. The security implications of these developments are substantial, as more advanced language models can potentially be used for malicious purposes, making it essential for practitioners to stay informed about the latest advancements in RLVR.