Loading...
Loading...
Browse all stories on DeepNewz
VisitMicrosoft Open-Sources bitnet.cpp, Achieves 6x Speed Improvements and 82% Energy Reduction
Oct 20, 2024, 10:00 AM
Microsoft has open-sourced bitnet.cpp, a highly efficient 1-bit large language model (LLM) inference framework that runs directly on CPUs. This groundbreaking technology allows for the deployment of large 100-billion parameter models on local devices with significant performance improvements. Bitnet.cpp offers up to 6x speed improvements on x86 CPUs and up to 5x on ARM CPUs, along with energy reductions between 71.9% to 82.2%. The framework, which quantizes models to 1.58 bits, supports running these models at 5-7 tokens per second. BitNet b1.58 was 4.1 times faster and 8.9 times higher throughput capable. This development marks a significant advancement in AI technology, making high-performance LLMs more accessible and energy-efficient.
View original story
Markets
Yes • 50%
No • 50%
Public announcements or press releases from major tech companies
No • 50%
Yes • 50%
Product launch announcements from major AI companies
No • 50%
Yes • 50%
Official Microsoft announcements or updates on GitHub
RISC-V • 25%
x86 • 25%
Other • 25%
ARM • 25%
Benchmark reports or studies comparing CPU performance with bitnet.cpp
Healthcare • 25%
Automotive • 25%
Consumer Electronics • 25%
Enterprise Software • 25%
Industry reports or announcements from companies using bitnet.cpp