Loading...
Loading...
Browse all stories on DeepNewz
VisitCerebras Challenges Nvidia with Fastest AI Inference Service
Aug 27, 2024, 04:05 PM
Cerebras Systems has launched what it claims to be the world's fastest AI inference service, marking a significant challenge to Nvidia's dominance in the AI computing sector. The new service, powered by Cerebras' custom wafer-scale AI accelerator chips, offers AI developers access to high-speed applications at a lower cost. Cerebras has set a new record for AI inference speed, serving Llama 3.1 8B at 1,850 output tokens per second and 70B at 446 output tokens per second. This move is part of a broader trend where several chipmakers are attempting to break Nvidia's stronghold in the AI market.
View original story
Markets
No • 50%
Yes • 50%
Official press releases and announcements from Cerebras Systems and the cloud provider
No • 50%
Yes • 50%
Publicly available benchmarks and announcements from Cerebras Systems and Nvidia
Yes • 50%
No • 50%
Publicly available benchmarks and announcements from Nvidia
Nvidia • 25%
Other • 25%
Google • 25%
Cerebras • 25%
Benchmark reports from reputable sources such as MLPerf or independent testing labs
Cerebras • 25%
Other • 25%
Nvidia • 25%
Intel • 25%
Market analysis reports from reputable firms like Gartner or IDC
Other • 25%
Nvidia • 25%
Cerebras • 25%
AMD • 25%
Financial reports and earnings releases from the companies