News

Per benchmarks shared, the technique scales Qwen2.5-Math-7B from 58.8% to 90.0% and Phi3-mini-3.8B from 41.4% to 86.4%. Interestingly, this allows the SMLs to surpass OpenAI's o1 reasoning model ...
Renowned for its exceptional reasoning capabilities, particularly in complex fields like mathematics and coding ... a ...
Alibaba used the older Qwen2.5-Math and Qwen2.5-Coder models to generate synthetic training data. The training took place in two phases: the first with a context length of 4K and 30 trillion ...