Loading...
Loading...
Browse all stories on DeepNewz
VisitWhat will be the rank of NVLM-1.0-D-72B in math tasks by March 31, 2025?
Top 1 • 25%
Top 2 • 25%
Top 3 • 25%
Below Top 3 • 25%
Evaluation results published by credible sources or Nvidia's official announcements
Nvidia Releases NVLM-1.0-D-72B Multimodal LLM with Decoder-Only Architecture Achieving SOTA Results on Vision-Language Tasks
Oct 1, 2024, 05:58 AM
Nvidia has released NVLM-1.0-D-72B, a frontier-class multimodal large language model (LLM) with a decoder-only architecture. The model achieves state-of-the-art (SOTA) results on both vision-language and text-only tasks. It is reported to rival other advanced models such as GPT-4o, Llama 3-V 405B, and InternVL 2 in various evaluations, including math and coding. Nvidia has also made the checkpoint and inference scripts available on Hugging Face, with training code and additional versions like NVLM-1.0-X and NVLM-1.0-H expected to follow.
View original story
Top 5 • 25%
Top 10 • 25%
Top 20 • 25%
Outside Top 20 • 25%
Top 1 • 25%
Top 3 • 25%
Top 5 • 25%
Below Top 5 • 25%
DeepSeek-R1-Lite-Preview • 25%
OpenAI's o1-preview • 25%
Google DeepMind's model • 25%
Other • 25%
Less than 80% • 25%
80% to 85% • 25%
85% to 90% • 25%
Over 90% • 25%
Top 10% • 25%
Top 25% • 25%
Top 50% • 25%
Below 50% • 25%
Top-1 • 25%
Top-3 • 25%
Top-5 • 25%
Not in Top-5 • 25%
Rank 1 • 25%
Rank 2 • 25%
Rank 3 • 25%
Rank 4 or lower • 25%
Yes • 50%
No • 50%
Less than 5,000 • 25%
5,000 to 10,000 • 25%
10,000 to 20,000 • 25%
More than 20,000 • 25%
Phi-4 • 25%
Gemini Pro • 25%
Llama 3.3 • 25%
Other • 25%
Top 3 • 25%
Top 5 • 25%
Top 10 • 25%
Outside Top 10 • 25%
Yes • 50%
No • 50%
No • 50%
Yes • 50%
20-30% • 25%
Below 20% • 25%
Above 40% • 25%
30-40% • 25%