DeepSeek Releases MIT-Licensed 685B-Parameter Model Matching OpenAI o1, Up to 30x Cheaper
Jan 20, 2025, 01:14 PM
DeepSeek, an AI company, has officially released DeepSeek-R1, a fully open-source reasoning model with 685 billion parameters that claims performance on par with OpenAI's o1 model. Licensed under MIT, the model allows for free distillation and commercialization, offering up to 30 times cost reductions compared to OpenAI's o1. DeepSeek-R1 utilizes reinforcement learning techniques to enhance reasoning capabilities, with DeepSeek-R1-Zero being trained purely through RL without initial supervised fine-tuning. The model reportedly achieves comparable results to OpenAI-o1 across math, code, and reasoning tasks and supports up to 128,000 tokens. Both DeepSeek-R1 and DeepSeek-R1-Zero are available on Hugging Face, and the company has released six distilled smaller models, including 70B and 32B parameter versions based on Llama and Qwen architectures. The model is accessible via API and the DeepSeek website and apps, and a technical report detailing the model's development and training processes has been published.
View original story
Markets
No • 50%
Yes • 50%
Official press releases or announcements from DeepSeek
No • 50%
Yes • 50%
API usage reports or announcements from DeepSeek and OpenAI
No • 50%
Yes • 50%
Official benchmark results published by DeepSeek or independent AI benchmark organizations
DeepSeek-R1 • 25%
Other • 25%
Google's AI model • 25%
OpenAI o1 • 25%
Independent cost-effectiveness analyses published by AI research firms or industry analysts
Other • 25%
DeepSeek-R1 • 25%
OpenAI o1 • 25%
Google's AI model • 25%
Award announcements from major AI conferences and organizations
DeepSeek • 25%
OpenAI • 25%
Google • 25%
Other • 25%
Official announcements and publications from AI companies