Loading...
Loading...
Browse all stories on DeepNewz
VisitWhich AI conference will highlight nGPT as a key innovation by end of 2024?
NeurIPS 2024 • 25%
ICML 2024 • 25%
CVPR 2024 • 25%
Other • 25%
Conference agendas and keynote announcements
NVIDIA Unveils Normalized Transformer (nGPT) for 4-20x Faster Training and Improved Stability of Large Language Models
Oct 19, 2024, 02:20 PM
NVIDIA researchers have introduced a new architecture called the Normalized Transformer (nGPT), which utilizes a hypersphere-based design to enhance the training speed and stability of large language models (LLMs). The nGPT architecture reportedly achieves training speeds that are 4 to 20 times faster than previous models. This advancement is expected to significantly improve convergence speed in transformer models, as noted by various experts in the field. The introduction of nGPT marks a notable development in the ongoing evolution of AI and machine learning technologies.
View original story
Grok 2 • 25%
Imagen 3 • 25%
GPT-5 • 25%
Other • 25%
NeurIPS • 25%
ICML • 25%
CVPR • 25%
Other or None • 25%
NeurIPS • 25%
ICML • 25%
CVPR • 25%
Other • 25%
NeurIPS • 25%
ICML • 25%
AAAI • 25%
Other • 25%
NeurIPS • 25%
ICML • 25%
CVPR • 25%
Other • 25%
Behavioral Simulation • 25%
Customer Service • 25%
Content Creation • 25%
Data Analysis • 25%
NeurIPS • 25%
ICML • 25%
AAAI • 25%
CVPR • 25%
OpenAI • 25%
Microsoft • 25%
Both • 25%
Neither • 25%
OpenAI • 25%
DeepMind • 25%
Anthropic • 25%
Other • 25%
CVPR • 25%
NeurIPS • 25%
ICML • 25%
Other • 25%
Natural Language Processing • 25%
Computer Vision • 25%
Autonomous Vehicles • 25%
Robotics • 25%
No • 50%
Yes • 50%
Energy Efficiency • 25%
Other • 25%
Training Speed • 25%
Stability • 25%