Loading...
Loading...
Browse all stories on DeepNewz
VisitMeta's 2024 Multi-token Prediction Makes LLMs 3x Faster
May 2, 2024, 03:23 PM
Meta has introduced a new method for training large language models (LLMs) such as GPT and Llama, titled 'Better & Faster Large Language Models via Multi-token Prediction'. This approach, detailed in a 2024 paper by F Gloeckle, B Y Idrissi, B Rozière, D Lopez-Paz, and G Synnaeve from FAIR at Meta, involves training models to predict four future tokens simultaneously. This method has shown to improve sample efficiency and increase inference speeds by up to three times.
View original story
Healthcare • 20%
Finance • 20%
Automotive • 20%
Retail • 20%
Education • 20%
Content Creation • 25%
Video Game Development • 25%
Film Production • 25%
Other • 25%
OpenAI • 25%
Google • 25%
Microsoft • 25%
Other • 25%
AI for news writing • 25%
AI for content curation • 25%
AI for audience engagement • 25%
AI for ad targeting • 25%
Entertainment • 20%
Personal Assistants • 20%
Customer Service • 20%
Healthcare • 20%
Other • 20%
Personalized learning • 33%
Workplace productivity • 33%
Daily tasks assistance • 34%
Personal Assistant • 25%
Customer Support • 25%
Language Translation • 25%
Content Generation • 25%
WhatsApp • 33%
Messenger • 33%
Instagram • 34%
Jina AI • 33%
Nomic AI • 33%
OpenAI • 33%
Customer Support • 33%
Sales Facilitation • 33%
Both • 33%
Chatbots • 25%
Autonomous vehicles • 25%
Healthcare AI • 25%
Other AI applications • 25%
Microsoft • 25%
Amazon • 25%
Google • 25%
IBM • 25%