Loading...
Loading...
Browse all stories on DeepNewz
VisitChina's StepFun Unveils Step-2, Trillion-Parameter MoE Model Developed in Two Months, Ranks Fifth Globally
Nov 21, 2024, 05:00 AM
Chinese artificial intelligence startup StepFun has developed Step-2, a trillion-parameter Mixture of Experts (MoE) language model with a 16k context length that ranks fifth globally on LiveBench evaluations. The model has surpassed GPT-4o and is just behind o1-mini, demonstrating China's rapid progress in AI technology despite GPU trade blockades. Remarkably, Step-2 was developed in just two months and trained at a fraction of the cost of OpenAI's GPT-4, with an estimated expenditure of $3 million compared to GPT-4's $80–$100 million. The emergence of Step-2, China's top-performing large language model, highlights the country's significant investments in AI and its growing competitiveness in the field.
View original story
Markets
Yes • 50%
No • 50%
LiveBench evaluation rankings
No • 50%
Yes • 50%
Press releases from StepFun or the tech company
No • 50%
Yes • 50%
Official announcements from StepFun
StepFun • 25%
Other • 25%
o1-mini • 25%
OpenAI • 25%
LiveBench evaluation rankings
Other • 25%
China • 25%
USA • 25%
European Union • 25%
Industry reports and AI model announcements
New model with over 1 trillion parameters • 25%
Major partnership with international company • 25%
Breakthrough in AI training efficiency • 25%
Other significant development • 25%
Official announcements from StepFun