"We are living in a timeline where a non-US company is keeping the original mission of OpenAI alive—truly open, frontier research that empowers all," said Jim Fan, senior research manager and lead of embodied AI (GEAR Lab) at NVIDIA. Besides the flagship model, DeepSeek also released six distilled versions, ranging from 1.5 billion to 70 billion parameters, optimised for math, coding, and reasoning. The lab is betting big on DeepSeek-R1-Zero, a model built entirely on reinforcement learning, which means it develops reasoning capabilities without any supervised data. "Our goal is to explore the potential of LLMs to develop reasoning capabilities without any supervised data, focusing on their self-evolution through a pure RL process," said the DeepSeek team. And it's working—DeepSeek-R1-Zero achieved a 71% pass rate on AIME 2024, up from just 15.6%, while DeepSeek-R1 surpassed OpenAI's o1-1217 with a 79.8% Pass@1. The AI community is taking notice. Paras Chopra, founder of Wingify, called it a breakthrough: "I love DeepSeek, so much! o1 level model is now open-source (MIT license)." Bindu Reddy, founder of Abacus AI, added: "DeepSeek R1 is on par with o1 and is open-source!! It blows my mind that Chinese make great, open and transparent tech." While OpenAI is caught up in o3 benchmark controversies, DeepSeek is setting new standards for transparency. "Whale 🐋 folks, respect," said KissanAI founder Pratik Desai, summing up the industry's reaction. Enjoy the full story here.
Nandan Nilakani is 'Wrong' |
Комментариев нет:
Отправить комментарий
Примечание. Отправлять комментарии могут только участники этого блога.