Loading...
Loading...
Browse all stories on DeepNewz
VisitWhich quantized AI model will see the highest adoption by December 31, 2024?
Meta's Llama 3.1-70B • 25%
OpenAI's GPT-4 quantized • 25%
Google's Bard quantized • 25%
Other • 25%
Public announcements or press releases from companies adopting the models
Meta's Llama 3.1 Models Offer 4X Cheaper AI Deployments, Achieve 99% Performance
Aug 14, 2024, 09:16 AM
Meta's AI division has released the Llama 3.1 models, which have been quantized to 4 bits by Neural Magic's research team. This quantization allows for 4X cheaper deployments, reducing the need from two 8x80GB nodes to one 4x80GB node. The models, including the 405B, 70B, and 8B versions, maintain approximately 100% recovery of the original performance. The HQQ Llama-3.1-70B model, in particular, achieves 99% of the base model's performance across various benchmarks, marking a significant advancement in AI model efficiency and cost-effectiveness.
View original story
OpenAI's O1 model • 25%
GPT-4 • 25%
Gemini • 25%
Anthropic's Claude • 25%
Claude 3.5 Sonnet • 33%
GPT-4o • 33%
Google's AI Model • 33%
Claude 3.5 • 25%
GPT-4 • 25%
Gemini Pro • 25%
Llama • 25%
Meta (Llama 3) • 25%
OpenAI (GPT-4o) • 25%
Anthropic (Claude 3.5 Sonnet) • 25%
Other • 25%
Phi-4 • 25%
Gemini Pro • 25%
Llama 3.3 • 25%
Other • 25%
OpenAI's O1 model • 25%
GPT-4 • 25%
Gemini • 25%
Anthropic's Claude • 25%
Claude 3.5 Sonnet • 25%
GPT-4o • 25%
Gemini Pro • 25%
Llama-3 • 25%
Google's Gemini • 25%
OpenAI's GPT • 25%
Microsoft's Azure AI • 25%
Other • 25%
Llama 3.1 • 25%
GPT-4o • 25%
Bard • 25%
Other • 25%
Llama 3.1 405B • 25%
GPT-4o • 25%
Claude Sonnet 3.5 • 25%
Other • 25%
Claude 3.5 Sonnet • 33%
GPT-4o • 33%
Google's AI Model • 33%
ChatGPT-4o • 25%
Google's Gemini • 25%
Another AI model • 25%
No clear winner • 25%
No • 50%
Yes • 50%
No • 50%
Yes • 50%
Yes • 50%
No • 50%
Meta's Llama 3.1-70B • 25%
Other • 25%
Google's Bard • 25%
OpenAI's GPT-4 • 25%