Loading...
Loading...
Browse all stories on DeepNewz
VisitWhich proprietary model will Molmo outperform next in a publicly available benchmark test by June 30, 2025?
GPT-4V • 25%
Claude 3.5 Sonnet • 25%
Flash • 25%
Other • 25%
Results from publicly available AI benchmark tests
AI2 Releases Molmo, Open-Source AI Model in 1B, 7B, and 72B Sizes Outperforming Proprietary Systems
Sep 25, 2024, 01:50 PM
The Allen Institute for AI (AI2) has released the Multimodal Open Language Model (Molmo), a state-of-the-art open-source AI model. Molmo is available in 1B, 7B, and 72B-parameter sizes and has been shown to outperform proprietary models such as GPT-4V, Claude 3.5 Sonnet, and Flash. The model's performance is attributed to its focus on data quality over quantity, utilizing a meticulously curated dataset called PixMo. Molmo's capabilities include understanding and acting on multimodal data, enabling rich interactions in both physical and virtual worlds. The release includes four model checkpoints: MolmoE-1B, Molmo-7B-O, and others, making it the most capable open-source AI model to date. Human preference for the 72B model is on par with top API models. The dataset PixMo was curated over 9 months.
View original story
Ranked 1st • 25%
Ranked 2nd • 25%
Ranked 3rd to 5th • 25%
Ranked below 5th • 25%
MMLU • 25%
ARC • 25%
GSM8K • 25%
None by June 30, 2024 • 25%
OpenAI o1 • 33%
Anthropic • 33%
Other • 33%
GPQA Diamond (CoT) • 25%
Math (CoT) • 25%
Other AI Benchmark • 25%
None • 25%
Yes • 50%
No • 50%
Less than 10% • 25%
10% to 20% • 25%
20% to 30% • 25%
More than 30% • 25%
Yes • 50%
No • 50%
Top 1 • 25%
Top 3 • 25%
Top 5 • 25%
Below Top 5 • 25%
Google • 25%
Microsoft • 25%
Meta • 25%
Amazon • 25%
No • 50%
Yes • 50%
No • 50%
Yes • 50%
Yes • 50%
No • 50%
72B • 25%
Other • 25%
1B • 25%
7B • 25%