Loading...
Loading...
Browse all stories on DeepNewz
VisitNVIDIA's Mistral-NeMo-Minitron 8B Model Achieves High Accuracy
Aug 21, 2024, 04:08 PM
NVIDIA has announced the release of the Mistral-NeMo-Minitron 8B, a small language model derived from the Mistral NeMo 12B model through pruning and distillation techniques. This new model is designed to offer state-of-the-art accuracy while being more computationally efficient. The Mistral-NeMo-Minitron 8B achieves high accuracy in nine popular benchmarks, including those for chatbots, virtual assistants, content generation, and coding. The model was developed using 400 billion tokens for distillation, significantly reducing the computational cost compared to training from scratch. NVIDIA's innovation in pruning and distillation has enabled the creation of a commercially permissive model that maintains high performance levels. The model achieved an MMLU score of 69.5.
View original story
Markets
No • 50%
Yes • 50%
Official announcements from major tech companies or press releases
Yes • 50%
No • 50%
Official announcements from NVIDIA or press releases
No • 50%
Yes • 50%
Benchmark results published by NVIDIA or independent benchmarking organizations
Below 69.0 • 25%
70.0 or higher • 25%
69.5 to 69.9 • 25%
69.0 to 69.4 • 25%
Benchmark results published by NVIDIA or independent benchmarking organizations
Coding • 25%
Chatbots • 25%
Virtual assistants • 25%
Content generation • 25%
Benchmark results published by NVIDIA or independent benchmarking organizations
Amazon • 25%
Meta • 25%
Google • 25%
Microsoft • 25%
Official announcements from the companies or press releases