Facefam ArticlesFacefam Articles
  • webmaster
    • How to
    • Developers
    • Hosting
    • monetization
    • Reports
  • Technology
    • Software
  • Downloads
    • Windows
    • android
    • PHP Scripts
    • CMS
  • REVIEWS
  • Donate
  • Join Facefam
Search

Archives

  • May 2025
  • April 2025
  • March 2025
  • January 2025
  • December 2024
  • November 2024

Categories

  • Advertiser
  • AI
  • android
  • betting
  • Bongo
  • Business
  • CMS
  • cryptocurrency
  • Developers
  • Development
  • Downloads
  • Entertainment
  • Entrepreneur
  • Finacial
  • General
  • Hosting
  • How to
  • insuarance
  • Internet
  • Kenya
  • monetization
  • Music
  • News
  • Phones
  • PHP Scripts
  • Reports
  • REVIEWS
  • RUSSIA
  • Software
  • Technology
  • Tips
  • Tragic
  • Ukraine
  • Uncategorized
  • USA
  • webmaster
  • webmaster
  • Windows
  • Women Empowerment
  • Wordpress
  • Wp Plugins
  • Wp themes
Facefam 2025
Notification Show More
Font ResizerAa
Facefam ArticlesFacefam Articles
Font ResizerAa
  • Submit a Post
  • Donate
  • Join Facefam social
Search
  • webmaster
    • How to
    • Developers
    • Hosting
    • monetization
    • Reports
  • Technology
    • Software
  • Downloads
    • Windows
    • android
    • PHP Scripts
    • CMS
  • REVIEWS
  • Donate
  • Join Facefam
Have an existing account? Sign In
Follow US
Technologywebmaster

Microsoft Releases Largest 1-Bit LLM, Letting Powerful AI Run on Some Older Hardware

Ronald Kenyatta
Last updated: April 18, 2025 9:13 pm
By
Ronald Kenyatta
ByRonald Kenyatta
Follow:
Share
4 Min Read
SHARE

Contents
What makes a bitnet model different?BitNet b1.58 2B4T model performance, purpose, and limitationsPerformance compared to other AI modelsResearchers’ goal for this bitnetMore must-read AI coverageWhat’s next?

Microsoft researchers claim to have developed the first 1-bit large language model with 2 billion parameters. The model, BitNet b1.58 2B4T, can run on commercial CPUs such as Apple’s M2.

“Trained on a corpus of 4 trillion tokens, this model demonstrates how native 1-bit LLMs can achieve performance comparable to leading open-weight, full-precision models of similar size, while offering substantial advantages in computational efficiency (memory, energy, latency),” Microsoft wrote in the project’s Hugging Face depository.

What makes a bitnet model different?

Bitnets, or 1-bit LLMs, are compressed versions of large language models. The original 2-billion parameter scale model trained on a corpus of 4 billion tokens was shrunken down into a version with drastically reduced memory requirements. All weights are expressed as one of three values: -1, 0, and 1. Other LLMs might use 32-bit or 16-bit floating-point formats.

SEE: Threat actors can inject malicious packages into AI models that resurface during “vibe coding.”

In the research paper, which was posted on Arxiv as a work in progress, the researchers detail how they created the bitnet. Other groups have created bitnets before, but, the researchers say, most of their efforts are either post-training quantization (PTQ) methods applied to pre-trained full-precision models or native 1-bit models trained from scratch that were developed at a smaller scale in the first place. BitNet b1.58 2B4T is a native 1-bit LLM trained at scale; it only takes up 400MB, compared to other “small models” that can reach up to 4.8 GB.

BitNet b1.58 2B4T model performance, purpose, and limitations

Performance compared to other AI models

BitNet b1.58 2B4T outperforms other 1-bit models, according to Microsoft. BitNet b1.58 2B4T has a maximum sequence length of 4096 tokens; Microsoft claims it outperforms small models like Meta’s Llama 3.2 1B or Google’s Gemma 3 1B.

Researchers’ goal for this bitnet

Microsoft’s goal is to make LLMs accessible to more people by creating versions that run on edge devices, in resource-constrained environments, or in real-time applications.

However, BitNet b1.58 2B4T still isn’t simple to run; it requires hardware compatible with Microsoft’s bitnet.cpp framework. Running it on a standard transformers library won’t produce any of the benefits in terms of speed, latency, or energy consumption. BitNet b1.58 2B4T doesn’t run on GPUs, as the majority of AI models do.

More must-read AI coverage

What’s next?

Microsoft’s researchers plan to explore training larger, native 1-bit models (7B, 13B parameters and more).They note that most of today’s AI infrastructure lacks suitable hardware for 1-bit models, so they plan to explore “co-designing future hardware accelerators” specifically designed for compressed AI. The researchers also aim to:

  • Increase context length.
  • Improve performance on long-context chain-of-thought reasoning tasks.
  • Add support for multiple languages other than English.
  • Integrate 1-bit models into multimodal architectures.
  • Better understand the theory behind why 1-bit training at scale produced efficiencies.
TAGGED:1Bitartificial intelligencebitnetgooglegoogle geminiHardwarelarge language modelsLargestLettingllmmetameta llamaMicrosoftOlderPowerfulReleasesRun
Share This Article
Facebook Whatsapp Whatsapp Email Copy Link Print
What do you think?
Love0
Sad0
Happy0
Sleepy0
Angry0
Dead0
Wink0
Previous Article OpenAI Agents Now Support Rival Anthropic’s Protocol OpenAI’s New AI Models o3 and o4-mini Can Now ‘Think With Images’
Next Article Flat vector illustration of the automation concept. The End of Fragmented Automation
Leave a review

Leave a Review Cancel reply

Your email address will not be published. Required fields are marked *

Please select a rating!

Feature-by-Feature Comparison: ShaunSocial vs. ColibriPlus – Which Social Network Script Comes Out on Top?
How Enterprise IT Can Achieve Water Sustainability Despite the Demands of AI
AI Benchmark Discrepancy Reveals Gaps in Performance Claims
Huawei Readies Ascend 920 Chip to Replace Restricted NVIDIA H20
‘AI Is Fundamentally Incompatible With Environmental Sustainability’

Recent Posts

  • Feature-by-Feature Comparison: ShaunSocial vs. ColibriPlus – Which Social Network Script Comes Out on Top?
  • How Enterprise IT Can Achieve Water Sustainability Despite the Demands of AI
  • AI Benchmark Discrepancy Reveals Gaps in Performance Claims
  • Huawei Readies Ascend 920 Chip to Replace Restricted NVIDIA H20
  • ‘AI Is Fundamentally Incompatible With Environmental Sustainability’

Recent Comments

  1. https://tubemp4.ru on Best Features of PHPFox Social Network Script
  2. Вулкан Платинум on Best Features of PHPFox Social Network Script
  3. Вулкан Платинум официальный on Best Features of PHPFox Social Network Script
  4. Best Quality SEO Backlinks on DDoS Attacks Now Key Weapons in Geopolitical Conflicts, NETSCOUT Warns
  5. http://boyarka-inform.com on Comparing Wowonder and ShaunSocial

You Might Also Like

Photo of Google
Technologywebmaster

Google is Betting Big on Nuclear Energy – Here’s Why

April 19, 2025
Screenshot from Microsoft
Technologywebmaster

Microsoft’s New Copilot Studio Feature Offers More User-Friendly Automation

April 19, 2025
iot-spy.jpg
Technologywebmaster

US Officials Claim DeepSeek AI App Is ‘Designed To Spy on Americans’

April 19, 2025
Flat vector illustration of the automation concept.
Technologywebmaster

The End of Fragmented Automation

April 18, 2025
OpenAI Agents Now Support Rival Anthropic’s Protocol
Technologywebmaster

OpenAI’s New AI Models o3 and o4-mini Can Now ‘Think With Images’

April 18, 2025
Previous Next
Facefam ArticlesFacefam Articles
Facefam Articles 2025
  • Submit a Post
  • Donate
  • Join Facefam social
Welcome Back!

Sign in to your account

Username or Email Address
Password

Lost your password?

Not a member? Sign Up