Loading...
Loading...
Browse all stories on DeepNewz
VisitWhich institution will lead in research papers on MatMul-free language models by end of 2024?
Princeton University • 33%
Stevens Institute of Technology • 33%
University of Pennsylvania • 33%
Databases of academic publications such as Google Scholar or arXiv
Researchers Develop Scalable MatMul-Free Language Model with 61% Memory Reduction
Jun 6, 2024, 12:12 PM
Researchers have developed a scalable, MatMul-free language model that eliminates the need for matrix multiplication operations while maintaining strong performance at billion-parameter scales. This new approach, which replaces MatMul operations with addition and negation, has shown to reduce memory usage by up to 61% and improve GPU efficiency. The model processes billion-parameter scale models at 13W beyond human-readable throughput, moving large language models (LLMs) closer to brain-like efficiency. The implementation has been a collaborative effort involving researchers W Guo, J Long, Y Zeng, and Z Liu from Princeton University, Stevens Institute of Technology, and the University of Pennsylvania.
View original story
Arctic AI • 25%
GPT-4 by OpenAI • 25%
Google's Language Model • 25%
Meta's Language Model • 25%
MIT • 20%
Stanford • 20%
Carnegie Mellon • 20%
University of California, Berkeley • 20%
DeepMind • 20%
Arizona State University • 25%
Harvard University • 25%
Stanford University • 25%
MIT • 25%
MIT • 20%
Harvard • 20%
Caltech • 20%
Carnegie Mellon • 20%
Oxford • 20%
Llama8B-related • 25%
GPT-5-related • 25%
Gemini-related • 25%
Other • 25%
Falcon 2 • 33%
Meta's Llama 3 • 33%
OpenAI's models • 34%
NeurIPS • 25%
ICML • 25%
CVPR • 25%
None • 25%
Qwen2 • 25%
Llama 3 • 25%
GLM 4 • 25%
Other • 25%
Yes • 50%
No • 50%
Google AI • 25%
OpenAI • 25%
Microsoft MAI-1 • 25%
Anthropic • 25%
No • 50%
Yes • 50%
Memory reduction • 33%
Performance at scale • 33%
GPU efficiency • 33%