Loading...
Loading...
Browse all stories on DeepNewz
VisitWill companies report a 4X cost reduction in AI deployments using Meta's Llama 3.1 models by December 31, 2024?
Yes • 50%
No • 50%
Public financial reports or cost analysis from companies using the models
Meta's Llama 3.1 Models Offer 4X Cheaper AI Deployments, Achieve 99% Performance
Aug 14, 2024, 09:16 AM
Meta's AI division has released the Llama 3.1 models, which have been quantized to 4 bits by Neural Magic's research team. This quantization allows for 4X cheaper deployments, reducing the need from two 8x80GB nodes to one 4x80GB node. The models, including the 405B, 70B, and 8B versions, maintain approximately 100% recovery of the original performance. The HQQ Llama-3.1-70B model, in particular, achieves 99% of the base model's performance across various benchmarks, marking a significant advancement in AI model efficiency and cost-effectiveness.
View original story
Less than 50 • 25%
50 to 100 • 25%
101 to 150 • 25%
More than 150 • 25%
Yes • 50%
No • 50%
Yes • 50%
No • 50%
OpenAI • 25%
Anthropic • 25%
Google • 25%
Other • 25%
Yes • 50%
No • 50%
Yes • 50%
No • 50%
Yes • 50%
No • 50%
No • 50%
Yes • 50%
Yes • 50%
No • 50%
Meta's Llama 3.1-70B • 25%
Other • 25%
Google's Bard • 25%
OpenAI's GPT-4 • 25%
OpenAI's GPT-4 models • 25%
Other • 25%
Meta's Llama 3.1 models • 25%
Google's Bard models • 25%