The series contains four models with 1.5, 7, 14 and 32 billion parameters. All of them are created on the basis of a large model of 671 Milliard Deepseek, but they are much more compact, which allows you to run them on ordinary game graphics cards.

For learning, Nvidia used 5 million tasks in the field of mathematics, science and programming created using the NEMO Skills platform.

Models have been examined only with the use of controlled learning – without the use of RLHF, which has made them a suitable basis for further research.

According to the test results, the largest 32B model reaches 89.2 points in AIME24 and 73.8 Olympics and the youngest 1.5B at HMMT. In parallel, the performance of the 32B model can be compared or even over Openai, in the Gensect mode, which produces a few answers and chooses the best.

All four NVIDIA models are already available on the face of hugging for loading and local use.

Source: Ferra

Previous articleAmerican Conifer, Chinaavto on July 21, 2025, 08:15 due to unusual electric motors.
Next articleGSMarena, Xiaomi Smart Band recognized the best in the series: bright screen, right pulsometer and 10 -day work and technology 21 July 2025, 08:25
I am a professional journalist and content creator with extensive experience writing for news websites. I currently work as an author at Gadget Onus, where I specialize in covering hot news topics. My written pieces have been published on some of the biggest media outlets around the world, including The Guardian and BBC News.

LEAVE A REPLY

Please enter your comment!
Please enter your name here