Loading...
Loading...
Browse all stories on DeepNewz
VisitWill Cerebras surpass Nvidia in AI inference speed for Llama 3.1 8B by end of 2024?
Yes • 50%
No • 50%
Publicly available benchmarks and announcements from Cerebras Systems and Nvidia
Cerebras Challenges Nvidia with Fastest AI Inference Service
Aug 27, 2024, 04:05 PM
Cerebras Systems has launched what it claims to be the world's fastest AI inference service, marking a significant challenge to Nvidia's dominance in the AI computing sector. The new service, powered by Cerebras' custom wafer-scale AI accelerator chips, offers AI developers access to high-speed applications at a lower cost. Cerebras has set a new record for AI inference speed, serving Llama 3.1 8B at 1,850 output tokens per second and 70B at 446 output tokens per second. This move is part of a broader trend where several chipmakers are attempting to break Nvidia's stronghold in the AI market.
View original story
Yes • 50%
No • 50%
Cerebras significantly outperforms Nvidia • 25%
Cerebras slightly outperforms Nvidia • 25%
Nvidia slightly outperforms Cerebras • 25%
Nvidia significantly outperforms Cerebras • 25%
Yes • 50%
No • 50%
Cerebras offers significantly better price-performance • 25%
Cerebras offers slightly better price-performance • 25%
Nvidia offers slightly better price-performance • 25%
Nvidia offers significantly better price-performance • 25%
Yes • 50%
No • 50%
Yes • 50%
No • 50%
Yes • 50%
No • 50%
Yes • 50%
No • 50%
GPT-5 • 25%
BERT-3 • 25%
Claude 3.0 • 25%
Other • 25%
Nvidia • 25%
Other • 25%
Google • 25%
Cerebras • 25%
Cerebras • 25%
Other • 25%
Nvidia • 25%
Intel • 25%