Researchers have made a breakthrough in post-training for large reasoning models by developing Lightning OPD, an efficient method that enables offline on-policy distillation. This approach eliminates the need for a live teacher inference server, significantly reducing infrastructure overhead. By precomputing teacher log-probabilities, Lightning OPD achieves comparable performance to standard on-policy distillation while minimizing computational costs. The method has significant implications for the development and deployment of large language models, as it reduces the barriers to entry for organizations with limited resources. This advancement is particularly noteworthy as it can be applied to various domains, including natural language processing and computer vision1. The ability to perform on-policy distillation offline matters to practitioners because it enables more widespread adoption of large reasoning models, which can have far-reaching consequences for industries and societies alike.