Loading...
Loading...
Browse all stories on DeepNewz
VisitNVIDIA Releases Open-Source Nemotron-4 340B with 340B Parameters for Synthetic Data Generation
Jun 14, 2024, 07:11 PM
NVIDIA has announced the release of Nemotron-4 340B, a family of open-source models designed for generating synthetic data to train large language models (LLMs) for commercial applications. The models, which include Base, Instruct, and Reward variants, are optimized for NVIDIA NeMo and TensorRT-LLM platforms. Nemotron-4 340B boasts 340 billion parameters and was trained on 9 trillion tokens, making it one of the largest models to date. It is designed to help developers in various industries, such as healthcare, finance, manufacturing, and retail, to create synthetic data for training LLMs. The models have received a permissive license, allowing developers to own any derivative and model output. Additionally, Nemotron-4 340B ranks first on the Hugging Face RewardBench leaderboard and surpasses Llama-3-70B in several benchmarks. The model was tested under the codename "June-chatbot" and the Reward models were trained with HelpSteer2 preference data, excelling in Arena-Hard-Auto benchmarks.
View original story
Markets
No • 50%
Yes • 50%
Public announcements and press releases from major financial institutions
No • 50%
Yes • 50%
Public announcements and press releases from major healthcare organizations
No • 50%
Yes • 50%
Benchmark results published on platforms like Hugging Face and other relevant AI research journals
Nemotron-5 • 33%
Nemotron-7 • 33%
Nemotron-6 • 33%
Official announcements from NVIDIA
Finance • 25%
Retail • 25%
Manufacturing • 25%
Healthcare • 25%
Public announcements and press releases from companies in various industries
Reward • 33%
Base • 33%
Instruct • 33%
Benchmark results published on platforms like Hugging Face and other relevant AI research journals