Loading...
Loading...
Browse all stories on DeepNewz
VisitWhat will be the main advantage cited for adopting MatMul-free language model by end of 2024?
Memory reduction • 33%
GPU efficiency • 33%
Performance at scale • 33%
Public statements or press releases by companies adopting the model
Researchers Develop Scalable MatMul-Free Language Model with 61% Memory Reduction
Jun 6, 2024, 12:12 PM
Researchers have developed a scalable, MatMul-free language model that eliminates the need for matrix multiplication operations while maintaining strong performance at billion-parameter scales. This new approach, which replaces MatMul operations with addition and negation, has shown to reduce memory usage by up to 61% and improve GPU efficiency. The model processes billion-parameter scale models at 13W beyond human-readable throughput, moving large language models (LLMs) closer to brain-like efficiency. The implementation has been a collaborative effort involving researchers W Guo, J Long, Y Zeng, and Z Liu from Princeton University, Stevens Institute of Technology, and the University of Pennsylvania.
View original story
Falcon 2 • 33%
Meta's Llama 3 • 33%
OpenAI's models • 34%
Performance • 25%
Context length • 25%
Language support • 25%
Licensing model • 25%
Yes • 50%
No • 50%
Llama3 • 25%
GPT-4 • 25%
BERT • 25%
NVIDIA's Latest Model • 25%
Less than 20% improvement • 25%
20-40% improvement • 25%
40-60% improvement • 25%
More than 60% improvement • 25%
Intel • 25%
TSMC • 25%
Samsung • 25%
NVIDIA • 25%
Llama8B • 25%
Qwen 2 • 25%
Nemotron • 25%
Other • 25%
No • 50%
Yes • 50%
University of Pennsylvania • 33%
Stevens Institute of Technology • 33%
Princeton University • 33%