The purpose of the looking at the future...is to disturb the present!

Gaston Berger (1896-1960), francuski futurolog

Microsoft and China AI Research Possible Reinforcement Pre-Training Breakthrough

Reinforcement Pre-Training (RPT) is a new method for training large language models (LLMs) by reframing the standard task of predicting the next token in a sequence as a reasoning problem solved using reinforcement learning (RL). Unlike traditional RL methods for LLMs that need expensive human data or limited annotated data, RPT uses verifiable rewards based ...

Read more


Link :
https://www.nextbigfuture.com/2025/06/microsoft-and-china-ai-research-possible-reinforcement-pre-training-breakthrough.html