Loading...
Loading...
Browse all stories on DeepNewz
VisitLlama 3.1 8B model outperforms GPT-4 on CodeContests by end of 2024?
Yes • 50%
No • 50%
Results from an independent evaluation on CodeContests published by credible AI research organizations or journals
Meta AI Advances Coding LLMs with RLEF, Llama 3.1 Outperforms GPT-4 on CodeContests
Oct 4, 2024, 01:57 PM
Meta AI has introduced a significant advancement in coding large language models (LLMs) with the development of Reinforcement Learning with Execution Feedback (RLEF). This technique integrates execution feedback at training time to enhance performance at inference time. The approach has been successfully applied to fine-tune Llama 3.1 models, with the 8B model surpassing GPT-4 on DeepMind’s CodeContests and the 70B model achieving state-of-the-art results. Additionally, the method has been validated through extensive evaluations, including on SWE-bench, demonstrating its effectiveness in improving LLMs for code generation tasks. The evaluations were conducted using a cloud-based infrastructure that speeds up evaluations by 30x.
View original story
HumanEval • 25%
MMLU_social_sciences • 25%
Both • 25%
Neither • 25%
Yes • 50%
No • 50%
Yes • 50%
No • 50%
More than GPT-4o • 25%
Less than GPT-4o • 25%
Equal to GPT-4o • 25%
No significant usage • 25%
Meta AI • 25%
OpenAI • 25%
DeepMind • 25%
Other • 25%
Other • 25%
Llama 3.1 8B • 25%
GPT-4 • 25%
Llama 3.1 70B • 25%