Loading...
Loading...
Browse all stories on DeepNewz
VisitVILA 1.5 Most Adopted in Academia by Year-End?
Yes • 50%
No • 50%
surveys or adoption reports from major academic institutions
NVIDIA and MIT Launch VILA 1.5: Top OSS Vision Model with State-of-the-Art Accuracy
May 4, 2024, 07:16 PM
NVIDIA, in collaboration with MIT, has introduced a new vision language model named VILA 1.5, which can reason among multiple images, learn in context, and understand videos. This model, described as the best open-source vision language model currently available, has been fully open-sourced, including training code and data. VILA 1.5 has achieved state-of-the-art accuracy on the MMMU dataset and supports multi-image processing. It is optimized for performance on NVIDIA GPUs, including the Jetson Orin Nano, and is capable of running on multiple GPUs. The model also features AWQ quantized models and is touted as the fastest on NVIDIA's Jetson Orin Nano. The advancements of VILA 1.5 are detailed in the CVPR'24 paper.
View original story
Widely adopted • 33%
Moderately adopted • 33%
Poorly adopted • 33%
Low (0-10 institutions) • 25%
Moderate (11-30 institutions) • 25%
High (31-50 institutions) • 25%
Very High (over 50 institutions) • 25%
Phi-3-mini • 25%
Phi-3 14B • 25%
Llama-3 8B • 25%
GPT-3.5 • 25%
Below 25% • 25%
25% to 50% • 25%
50% to 75% • 25%
Above 75% • 25%
Yes • 50%
No • 50%
0-50 • 25%
51-100 • 25%
101-200 • 25%
201+ • 25%
Healthcare • 25%
Security • 25%
Automotive • 25%
Retail • 25%
Asia • 25%
Europe • 25%
North America • 25%
Rest of the World • 25%