Loading...
Loading...
Browse all stories on DeepNewz
VisitModernBERT Released: State-of-the-Art Encoder Model with 8,192 Token Support, 2 Trillion Tokens, and 2-3x Faster Processing
Dec 19, 2024, 04:51 PM
ModernBERT, a new family of state-of-the-art encoder-only models, has been released by AnswerAI and LightOn. This updated version of BERT features a sequence length of 8,192 tokens and is designed for improved performance in tasks such as classification and retrieval. Trained on 2 trillion tokens, ModernBERT is reported to be twice as fast as DeBERTaV3 on short contexts and three times faster than NomicBERT and GTE on long contexts. The model is available in two sizes: ModernBERT-base with 149 million parameters and ModernBERT-large with 395 million parameters. This release marks a significant upgrade to traditional BERT models, which have not seen a major overhaul in over six years.
View original story
Markets
Yes • 50%
No • 50%
Official announcements or press releases from top 5 tech companies
Yes • 50%
No • 50%
Leaderboard results from major NLP benchmarks like GLUE or SuperGLUE
No • 50%
Yes • 50%
Google Scholar citation count comparison between ModernBERT and BERT
Text Classification • 25%
Other • 25%
Sentiment Analysis • 25%
Information Retrieval • 25%
Review of published research papers and abstracts in NLP journals
Technology • 25%
Healthcare • 25%
Finance • 25%
Education • 25%
Industry reports and case studies on ModernBERT adoption
GTE • 25%
Other • 25%
DeBERTaV3 • 25%
NomicBERT • 25%
Analysis of academic papers and citations in databases like Google Scholar