Loading...
Loading...
Browse all stories on DeepNewz
VisitChina's StepFun Unveils Step-2, Trillion-Parameter MoE Model Developed in Two Months, Ranks Fifth Globally
Nov 21, 2024, 05:00 AM
Chinese artificial intelligence startup StepFun has developed Step-2, a trillion-parameter Mixture of Experts (MoE) language model with a 16k context length that ranks fifth globally on LiveBench evaluations. The model has surpassed GPT-4o and is just behind o1-mini, demonstrating China's rapid progress in AI technology despite GPU trade blockades. Remarkably, Step-2 was developed in just two months and trained at a fraction of the cost of OpenAI's GPT-4, with an estimated expenditure of $3 million compared to GPT-4's $80–$100 million. The emergence of Step-2, China's top-performing large language model, highlights the country's significant investments in AI and its growing competitiveness in the field.
View original story
Yes • 50%
No • 50%
OpenAI o1 • 33%
Anthropic • 33%
Other • 33%
ARC-AGI • 25%
Frontier Math • 25%
SWE-Bench • 25%
ARC-AGI Semi-Private Evaluation • 25%
SWE-Bench Verified test • 25%
AIME • 25%
GPQA-Diamond benchmark • 25%
OpenAI o1-preview • 25%
Anthropic Claude 3.5 Sonnet • 25%
OpenAI o1 mini • 25%
Other • 25%
GPT-4o • 33%
Gemini 1.5 • 33%
Claude 3.5 Sonnet • 34%
Top 1 • 25%
Top 2 to 3 • 25%
Top 4 to 5 • 25%
Below top 5 • 25%
Cosine • 25%
Amazon • 25%
Cognition • 25%
Other • 25%
Yes • 50%
No • 50%
Yes • 50%
No • 50%
Chatbots • 25%
Virtual assistants • 25%
Content generation • 25%
Coding • 25%
Top 1 • 25%
Top 2 • 25%
Top 3 • 25%
Below Top 3 • 25%
StepFun • 25%
Other • 25%
o1-mini • 25%
OpenAI • 25%
Other • 25%
China • 25%
USA • 25%
European Union • 25%