Loading...
Loading...
Browse all stories on DeepNewz
VisitTop Performing Model Trained on FineWeb by 2025
LLaMA-3 • 25%
BERT-Enhanced • 25%
GPT-4 • 25%
Transformer-XL • 25%
Benchmarking results published in peer-reviewed journals or major AI conferences.
HuggingFace Launches FineWeb, a 15T Token Dataset Outperforming Others
Apr 21, 2024, 08:02 AM
HuggingFace has released a new dataset named FineWeb, available at HuggingFace, which comprises 15 trillion tokens of high-quality, deduplicated web data sourced from CommonCrawl spanning from 2013 to 2024. This 275GB dataset, available under an Open Data Commons license, has shown to outperform existing datasets such as RefinedWeb, C4, DolmaV1.6, The Pile, and SlimPajama in various benchmarks, including 350B token ablations. FineWeb has been utilized to train models like LLaMA-3, demonstrating significant improvements due to its scale and quality. The dataset is expected to support extensive training runs given its size and has been made open-source for broad accessibility.
View original story
Phi-3 14B • 25%
Mixtral 8x7B • 25%
GPT-3.5 • 25%
Llama-3 8B • 25%
Falcon 2 • 33%
Meta's Llama 3 • 33%
OpenAI's latest model • 34%
Phi-3-mini • 25%
Phi-3 14B • 25%
Llama-3 8B • 25%
GPT-3.5 • 25%
Top performer • 25%
Top 3 • 25%
Top 5 • 25%
Outside Top 5 • 25%
Model A • 33%
Model B • 33%
Model C • 33%
Llama-2-70B-Chat • 50%
Meta-Llama-3-8B • 50%
Phi-3 series • 25%
Mixtral series • 25%
GPT series • 25%
Llama series • 25%
Social Media • 25%
Advertising • 25%
E-commerce • 25%
Customer Service • 25%
Language Processing • 25%
Image Recognition • 25%
Data Analysis • 25%
Automated Decision Making • 25%
FineWeb • 20%
The Pile • 20%
DolmaV1.6 • 20%
C4 • 20%
RefinedWeb • 20%