Loading...
Loading...
Browse all stories on DeepNewz
VisitPerformance benchmarks of UC Santa Cruz AI model compared to traditional models by end of 2024
Better • 25%
Equal • 25%
Worse • 25%
Inconclusive • 25%
Performance benchmark reports from AI research institutions
UC Santa Cruz Researchers Develop Efficient AI Models Running at 13 Watts by Eliminating Matrix Multiplication
Jun 26, 2024, 12:58 AM
Researchers from UC Santa Cruz and partners have developed a groundbreaking approach to operating AI language models by eliminating matrix multiplication. This innovative method promises to reduce power consumption and operational costs significantly while maintaining performance. Traditional large language models (LLMs) rely heavily on matrix multiplication, which strains GPUs. The new custom architecture achieves similar performance with smaller models, leading to more efficient and sustainable AI operations. This advancement could revolutionize the AI industry by making AI models more eco-friendly and cost-effective. AI chatbots can now run at a lightbulb-esque 13 watts with no performance loss.
View original story
Llama 3.1 405B • 25%
GPT-4o • 25%
Claude Sonnet 3.5 • 25%
Other • 25%
Claude 3.5 Sonnet • 33%
GPT-4o • 33%
Google's AI Model • 33%
Meta's Llama 3.1-70B • 25%
OpenAI's GPT-4 • 25%
Google's Bard • 25%
Other • 25%
MMLU • 25%
ARC • 25%
GSM8K • 25%
None by June 30, 2024 • 25%
ChatGPT-4o • 25%
Google's Gemini • 25%
Another AI model • 25%
No clear leader • 25%
Claude 3.5 Sonnet • 33%
GPT-4o • 33%
Gemini • 34%
Llama 3.1 405B • 25%
GPT-4o • 25%
Claude Sonnet 3.5 • 25%
Other • 25%
Nvidia • 25%
OpenAI • 25%
Anthropic • 25%
Other • 25%
Google's Gemini • 25%
OpenAI's GPT • 25%
Microsoft's Azure AI • 25%
Other • 25%
DeepSeek-R1-Lite-Preview • 25%
OpenAI's o1-preview • 25%
Google DeepMind's model • 25%
Other • 25%
Chatbots • 25%
Virtual assistants • 25%
Content generation • 25%
Coding • 25%
Better than NVIDIA H100 • 25%
Comparable to NVIDIA H100 • 25%
Worse than NVIDIA H100 • 25%
Other • 25%
No • 50%
Yes • 50%
5 or more products • 25%
None • 25%
1-2 products • 25%
3-4 products • 25%