Loading...
Loading...
Browse all stories on DeepNewz
VisitSambaNova Launches Fastest AI Inference Platform for Llama 3.1 at 570 Tokens/Second
Sep 10, 2024, 04:34 PM
SambaNova has announced the launch of its new cloud inference platform, SambaNova Cloud, which offers unprecedented speeds for AI model inference. Notably, the Llama 3.1 405B model achieves a speed of 132 tokens per second in full precision, while the Llama 3.1 70B model reaches up to 570 tokens per second. This performance is significantly faster than traditional GPUs, with claims of up to 10 times faster inference speeds. The platform operates in real-time and serves the Llama 3.1 405B model in 16-bit precision. It is available for developers starting today, with free access via API and no waitlist. The service has been independently verified and is expected to enable advanced AI applications.
View original story
Markets
No • 50%
Yes • 50%
Official performance benchmarks and announcements from SambaNova
Yes • 50%
No • 50%
Publicly available reports and press releases from top 100 AI companies
No • 50%
Yes • 50%
Official announcements from SambaNova and developer registration data
Less than 10% • 25%
More than 30% • 25%
20% to 30% • 25%
10% to 20% • 25%
Market analysis reports from reputable firms like Gartner or IDC
4th or lower • 25%
1st • 25%
2nd • 25%
3rd • 25%
Rankings from market research firms like Gartner or Forrester
Technology • 25%
Other • 25%
Healthcare • 25%
Finance • 25%
Industry usage reports and case studies published by SambaNova