Loading...
Loading...
Browse all stories on DeepNewz
VisitWill MatMul-free language model surpass 75% market share by end of 2024?
Yes • 50%
No • 50%
Industry reports or surveys conducted by AI research organizations
Researchers Develop Scalable MatMul-Free Language Model with 61% Memory Reduction
Jun 6, 2024, 12:12 PM
Researchers have developed a scalable, MatMul-free language model that eliminates the need for matrix multiplication operations while maintaining strong performance at billion-parameter scales. This new approach, which replaces MatMul operations with addition and negation, has shown to reduce memory usage by up to 61% and improve GPU efficiency. The model processes billion-parameter scale models at 13W beyond human-readable throughput, moving large language models (LLMs) closer to brain-like efficiency. The implementation has been a collaborative effort involving researchers W Guo, J Long, Y Zeng, and Z Liu from Princeton University, Stevens Institute of Technology, and the University of Pennsylvania.
View original story
Falcon 2 • 33%
Meta's Llama 3 • 33%
OpenAI's models • 34%
GPT-4o • 25%
Claude 3 • 25%
Google Bard • 25%
Other • 25%
Arctic AI • 25%
GPT-4 by OpenAI • 25%
Google's Language Model • 25%
Meta's Language Model • 25%
Top 3 • 25%
Top 5 • 25%
Top 10 • 25%
Outside Top 10 • 25%
Qwen2 • 25%
Llama 3 • 25%
GLM 4 • 25%
Other • 25%
Jina AI • 33%
Nomic AI • 33%
OpenAI • 33%
Most impactful • 25%
Equally impactful as others • 25%
Less impactful than Google's LLM • 25%
Less impactful than OpenAI's LLM • 25%
No • 50%
Yes • 50%
Memory reduction • 33%
Performance at scale • 33%
GPU efficiency • 33%
University of Pennsylvania • 33%
Stevens Institute of Technology • 33%
Princeton University • 33%