Loading...
Loading...
Browse all stories on DeepNewz
VisitWill nGPT achieve 10x training speed improvement in a major LLM project by end of 2024?
Yes • 50%
No • 50%
Research papers and public project reports from AI research organizations
NVIDIA Unveils Normalized Transformer (nGPT) for 4-20x Faster Training and Improved Stability of Large Language Models
Oct 19, 2024, 02:20 PM
NVIDIA researchers have introduced a new architecture called the Normalized Transformer (nGPT), which utilizes a hypersphere-based design to enhance the training speed and stability of large language models (LLMs). The nGPT architecture reportedly achieves training speeds that are 4 to 20 times faster than previous models. This advancement is expected to significantly improve convergence speed in transformer models, as noted by various experts in the field. The introduction of nGPT marks a notable development in the ongoing evolution of AI and machine learning technologies.
View original story
Yes • 50%
No • 50%
Yes • 50%
No • 50%
Yes • 50%
No • 50%
HumanEval • 25%
MMLU_social_sciences • 25%
Both • 25%
Neither • 25%
Energy Efficiency • 25%
Other • 25%
Training Speed • 25%
Stability • 25%
CVPR 2024 • 25%
NeurIPS 2024 • 25%
Other • 25%
ICML 2024 • 25%