Loading...
Loading...
Browse all stories on DeepNewz
VisitMeta's Llama 3 AI Model Shows Advanced Capabilities, Struggles with Quantization
May 12, 2024, 04:32 PM
Meta's recent launch of Llama 3, an advanced language model, has sparked significant interest in the AI community. Llama 3, notable for being trained on a record 15 trillion tokens, demonstrates a nuanced understanding of data relationships, which becomes evident in its performance degradation when quantized compared to its predecessor, Llama 2. The model's sensitivity is attributed to its use of BF16 precision, which captures even the minutest decimals. Specifically, the LLAMA3-8B model under LoRA-FT quantization shows that low-rank finetuning on the Alpaca dataset cannot compensate for the errors introduced by quantization. Additionally, Llama 3, which offers models like the 8B and 70B, is being integrated into platforms such as Promptitude for enhanced language processing and task management. The model's development was highlighted during a 24-hour hackathon involving over 500 AI engineers, showcasing its capabilities and potential applications in real-world scenarios.
View original story
Markets
Yes • 50%
No • 50%
Official announcements from Meta or recognized AI industry news outlets.
Yes • 50%
No • 50%
Public announcements or reports from major tech companies.
Yes • 50%
No • 50%
Press releases or official announcements from Meta.
Automotive • 25%
Retail • 25%
Healthcare • 25%
Finance • 25%
Industry reports or announcements by companies adopting Llama 3.
Llama 3-8B • 33%
Llama 3-70B • 33%
Other size • 34%
Usage data from Meta or developer surveys.
Language Processing • 25%
Image Recognition • 25%
Data Analysis • 25%
Automated Decision Making • 25%
Reports on performance enhancements in various applications due to Llama 3.