Loading...
Loading...
Browse all stories on DeepNewz
VisitChina's StepFun Unveils Step-2, Trillion-Parameter MoE Model Developed in Two Months, Ranks Fifth Globally
Nov 21, 2024, 05:00 AM
Chinese artificial intelligence startup StepFun has developed Step-2, a trillion-parameter Mixture of Experts (MoE) language model with a 16k context length that ranks fifth globally on LiveBench evaluations. The model has surpassed GPT-4o and is just behind o1-mini, demonstrating China's rapid progress in AI technology despite GPU trade blockades. Remarkably, Step-2 was developed in just two months and trained at a fraction of the cost of OpenAI's GPT-4, with an estimated expenditure of $3 million compared to GPT-4's $80–$100 million. The emergence of Step-2, China's top-performing large language model, highlights the country's significant investments in AI and its growing competitiveness in the field.
View original story
Cosine • 25%
Amazon • 25%
Cognition • 25%
Other • 25%
OpenAI • 25%
Google DeepMind • 25%
Anthropic • 25%
Cohere • 25%
OpenAI • 25%
Google • 25%
Meta • 25%
Other • 25%
DeepSeek • 25%
OpenAI • 25%
Google DeepMind • 25%
Other • 25%
Nvidia • 25%
OpenAI • 25%
Anthropic • 25%
Other • 25%
Tech industry • 25%
Healthcare • 25%
Finance • 25%
Education • 25%
Apple • 25%
Microsoft • 25%
Nvidia • 25%
Other • 25%
Apple • 25%
Tesla • 25%
Microsoft • 25%
Other • 25%
Nvidia • 25%
Apple • 25%
Microsoft • 25%
Other • 25%
Other • 25%
China • 25%
USA • 25%
European Union • 25%