Loading...
Loading...
Browse all stories on DeepNewz
VisitHow will Cerebras Inference performance compare to Nvidia's offerings by end of 2024?
Cerebras significantly outperforms Nvidia • 25%
Cerebras slightly outperforms Nvidia • 25%
Nvidia slightly outperforms Cerebras • 25%
Nvidia significantly outperforms Cerebras • 25%
Benchmark reports from independent testing labs or AI research institutions
Cerebras Launches AI Inference Service, 20x Faster with 1,850 Tokens/sec for 8B Model
Aug 27, 2024, 05:22 PM
Cerebras Systems has launched what it claims to be the world's fastest AI inference service, directly challenging Nvidia's dominance in the AI computing sector. The new service, known as Cerebras Inference, utilizes the company's custom wafer-scale AI accelerator chips to achieve significant performance gains. It is capable of processing Llama 3.1 models at impressive speeds: 1,850 tokens per second for the 8B model and 450 tokens per second for the 70B model. Cerebras asserts that its inference tool offers a 20x speed increase over traditional GPU-based systems, and it is priced at 60 cents per million tokens, which is a fifth of the cost compared to hyperscalers. The launch aims to provide a cost-effective and efficient alternative to Nvidia's GPUs, with claims of delivering 100x better price-performance.
View original story
Yes • 50%
No • 50%
Yes • 50%
No • 50%
Significantly better • 25%
Slightly better • 25%
Comparable • 25%
Worse • 25%
Yes • 50%
No • 50%
Yes • 50%
No • 50%
Nvidia • 25%
Cerebras • 25%
Google • 25%
Other • 25%
Cerebras • 25%
Nvidia • 25%
Groq • 25%
Other • 25%
Amazon • 25%
Nvidia • 25%
Microsoft • 25%
Alphabet • 25%
Below 5% • 25%
5% to 10% • 25%
10% to 15% • 25%
Above 15% • 25%
1st • 25%
2nd • 25%
3rd • 25%
4th or lower • 25%
Yes • 50%
No • 50%
Nvidia offers significantly better price-performance • 25%
Cerebras offers significantly better price-performance • 25%
Cerebras offers slightly better price-performance • 25%
Nvidia offers slightly better price-performance • 25%