Loading...
Loading...
Browse all stories on DeepNewz
VisitWhich benchmark will Google DeepMind's new LLM approach top first by end of 2024?
GLUE • 25%
SuperGLUE • 25%
SQuAD • 25%
Other • 25%
Published benchmark results in AI research papers or official AI benchmark sites
Google DeepMind's New LLM Approach Outperforms Larger Models
Aug 7, 2024, 01:54 AM
Google DeepMind and UC Berkeley researchers have released a paper demonstrating that optimizing test-time computation for large language models (LLMs) can be more effective than merely increasing model parameters. This approach allows LLMs to outperform models that are 14 times larger by utilizing more computation during inference. The research highlights a significant advancement in the development of self-improving LLMs, marking a critical step towards more efficient AI systems.
View original story
25% gain in MATH • 25%
15% gain in other domains • 25%
Integration into multiple commercial products • 25%
Other • 25%
Yes • 50%
No • 50%
MMLU • 25%
ARC • 25%
GSM8K • 25%
None by June 30, 2024 • 25%
Llama 3.1 405B • 25%
GPT-4o • 25%
Claude Sonnet 3.5 • 25%
Other • 25%
Meta's Llama 3.1-70B • 25%
OpenAI's GPT-4 • 25%
Google's Bard • 25%
Other • 25%
Claude 3.5 Sonnet • 33%
GPT-4o • 33%
Google's AI Model • 33%
Gold medal at 2025 IMO • 25%
Winning a Kaggle competition • 25%
Breakthrough in protein folding • 25%
Other • 25%
25% gain in self-correction for MATH dataset • 25%
15% gain in self-correction for other datasets • 25%
Adoption by three major tech companies • 25%
Other • 25%
Beginner • 25%
Intermediate • 25%
Advanced • 25%
Professional • 25%
OpenAI o1 • 25%
GPT-4 • 25%
Gemini • 25%
Claude • 25%
OpenAI o1-preview • 25%
Anthropic Claude 3.5 Sonnet • 25%
OpenAI o1 mini • 25%
Other • 25%
No • 50%
Yes • 50%
Yes • 50%
No • 50%
Google Cloud AI • 25%
Google Search • 25%
Other • 25%
Google Assistant • 25%