Loading...
Loading...
Browse all stories on DeepNewz
VisitMeta's Multi-Token Prediction Makes LLMs Up to 3X Faster
May 6, 2024, 04:34 PM
Recent advancements in large language models (LLMs) have been highlighted by a new study from Meta researchers, which proposes a multi-token prediction approach to enhance the training and performance of these models. This method, which involves training LLMs to predict multiple future tokens simultaneously, has shown to significantly increase both the speed and efficiency of the models, potentially making them up to three times faster. The research, led by Gloeckle et al., has been recognized as a simple yet effective innovation that could improve the functionality of LLMs, particularly in coding, planning, and robotics applications.
View original story
Llama-2-70B-Chat • 50%
Meta-Llama-3-8B • 50%
Google • 25%
Microsoft • 25%
IBM • 25%
Amazon • 25%
GSM8K • 33%
MATH • 33%
ARC-Challenge • 34%
Automotive • 20%
Healthcare • 20%
Education • 20%
Finance • 20%
Retail • 20%
IBM • 25%
Google • 25%
Amazon • 25%
Microsoft • 25%