A new training method called RLSD from JD.com and academic researchers aims to solve a common enterprise bottleneck: reasoning models are expensive to train and often get only sparse feedback. RLSD keeps reinforcement learning’s reliable direction while using self-distillation only for credit assignment, avoiding “privileged information leakage.” In tests, it beat baseline GRPO and standard OPSD with faster convergence.
Swipe through stories, personalise your feed, and save articles for later — all on the app.