{"id":822,"date":"2025-04-18T17:30:03","date_gmt":"2025-04-18T17:30:03","guid":{"rendered":"https:\/\/www.techrepublic.com\/?p=4302447"},"modified":"2025-04-18T17:30:03","modified_gmt":"2025-04-18T17:30:03","slug":"microsoft-releases-largest-1-bit-llm-letting-powerful-ai-run-on-some-older-hardware","status":"publish","type":"post","link":"https:\/\/blog.tecnoartesanos.com\/index.php\/2025\/04\/18\/microsoft-releases-largest-1-bit-llm-letting-powerful-ai-run-on-some-older-hardware\/","title":{"rendered":"Microsoft Releases Largest 1-Bit LLM, Letting Powerful AI Run on Some Older Hardware"},"content":{"rendered":"<div><img decoding=\"async\" src=\"https:\/\/assets.techrepublic.com\/uploads\/2025\/04\/chip-small-ai-apr-25.jpg\" class=\"ff-og-image-inserted\"><\/div>\n<p>Microsoft researchers claim to have developed the first 1-bit large language model with 2 billion parameters. The model, BitNet b1.58 2B4T, can run on commercial CPUs such as Apple\u2019s M2.<\/p>\n<p>\u201cTrained on a corpus of 4 trillion tokens, this model demonstrates how native 1-bit LLMs can achieve performance comparable to leading open-weight, full-precision models of similar size, while offering substantial advantages in computational efficiency (memory, energy, latency),\u201d Microsoft wrote in <a href=\"https:\/\/huggingface.co\/microsoft\/bitnet-b1.58-2B-4T\" target=\"_blank\" rel=\"noopener\">the project\u2019s Hugging Face depository<\/a>.<\/p>\n<h2><span class=\"ez-toc-section\" id=\"What_makes_a_bitnet_model_different\"><\/span>What makes a bitnet model different?<span class=\"ez-toc-section-end\"><\/span><\/h2>\n<p>Bitnets, or 1-bit LLMs, are compressed versions of large language models. The original 2-billion parameter scale model trained on a corpus of 4 billion tokens was shrunken down into a version with drastically reduced memory requirements. All weights are expressed as one of three values: -1, 0, and 1. Other LLMs might use 32-bit or 16-bit floating-point formats.<\/p>\n<p><strong>SEE: Threat actors can <a href=\"https:\/\/www.techrepublic.com\/article\/news-slopsquatting-vide-coding-ai-cybersecurity-risk\/\">inject malicious packages into AI models<\/a> that resurface during \u201cvibe coding.\u201d<\/strong><\/p>\n<p>In <a href=\"https:\/\/arxiv.org\/abs\/2504.12285\" target=\"_blank\" rel=\"noopener\">the research paper<\/a>, which was posted on Arxiv as a work in progress, the researchers detail how they created the bitnet. Other groups have created bitnets before, but, the researchers say, most of their efforts are either post-training quantization (PTQ) methods applied to pre-trained full-precision models or native 1-bit models trained from scratch that were developed at a smaller scale in the first place. BitNet b1.58 2B4T is a native 1-bit LLM trained at scale; it only takes up 400MB, compared to other \u201csmall models\u201d that can reach up to 4.8 GB.<\/p>\n<h2><span class=\"ez-toc-section\" id=\"BitNet_b158_2B4T_model_performance_purpose_and_limitations\"><\/span>BitNet b1.58 2B4T model performance, purpose, and limitations<span class=\"ez-toc-section-end\"><\/span><\/h2>\n<h3>Performance compared to other AI models<\/h3>\n<p>BitNet b1.58 2B4T outperforms other 1-bit models, according to Microsoft. BitNet b1.58 2B4T has a maximum sequence length of 4096 tokens; Microsoft claims it outperforms small models like Meta\u2019s Llama 3.2 1B or Google\u2019s Gemma 3 1B.<\/p>\n<h3>Researchers\u2019 goal for this bitnet<\/h3>\n<p>Microsoft\u2019s goal is to make LLMs accessible to more people by creating versions that run on edge devices, in resource-constrained environments, or in real-time applications.<\/p>\n<p>However, BitNet b1.58 2B4T still isn\u2019t simple to run; it requires hardware compatible with Microsoft\u2019s bitnet.cpp framework. Running it on a standard transformers library won\u2019t produce any of the benefits in terms of speed, latency, or energy consumption. BitNet b1.58 2B4T doesn\u2019t run on GPUs, as the majority of AI models do.<\/p>\n<aside class=\"pinbox right\">\n<h3 class=\"heading\">More must-read AI coverage<\/h3>\n<\/aside>\n<h3>What\u2019s next?<\/h3>\n<p>Microsoft\u2019s researchers plan to explore training larger, native 1-bit models (7B, 13B parameters and more).They note that most of today\u2019s AI infrastructure lacks suitable hardware for 1-bit models, so they plan to explore \u201cco-designing future hardware accelerators\u201d specifically designed for compressed AI. The researchers also aim to:<\/p>\n<ul>\n<li aria-level=\"1\">Increase context length.<\/li>\n<li aria-level=\"1\">Improve performance on long-context chain-of-thought reasoning tasks.<\/li>\n<li aria-level=\"1\">Add support for multiple languages other than English.<\/li>\n<li aria-level=\"1\">Integrate 1-bit models into multimodal architectures.<\/li>\n<li aria-level=\"1\">Better understand the theory behind why 1-bit training at scale produced efficiencies.<\/li>\n<\/ul>\n","protected":false},"excerpt":{"rendered":"<p>Microsoft researchers claim to have developed the first 1-bit large language model with 2 billion parameters. The model, BitNet b1.58 2B4T, can run on commercial CPUs such as Apple\u2019s M2. \u201cTrained on a corpus of 4 trillion tokens, this model demonstrates how native 1-bit LLMs can achieve performance comparable to leading open-weight, full-precision models of [&hellip;]<\/p>\n","protected":false},"author":2,"featured_media":823,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[1],"tags":[],"class_list":["post-822","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-technology"],"yoast_head":"<!-- This site is optimized with the Yoast SEO plugin v24.5 - https:\/\/yoast.com\/wordpress\/plugins\/seo\/ -->\n<title>Microsoft Releases Largest 1-Bit LLM, Letting Powerful AI Run on Some Older Hardware - TecnoArtesanos Tech Blog<\/title>\n<meta name=\"robots\" content=\"index, follow, max-snippet:-1, max-image-preview:large, max-video-preview:-1\" \/>\n<link rel=\"canonical\" href=\"https:\/\/blog.tecnoartesanos.com\/index.php\/2025\/04\/18\/microsoft-releases-largest-1-bit-llm-letting-powerful-ai-run-on-some-older-hardware\/\" \/>\n<meta property=\"og:locale\" content=\"en_US\" \/>\n<meta property=\"og:type\" content=\"article\" \/>\n<meta property=\"og:title\" content=\"Microsoft Releases Largest 1-Bit LLM, Letting Powerful AI Run on Some Older Hardware - TecnoArtesanos Tech Blog\" \/>\n<meta property=\"og:description\" content=\"Microsoft researchers claim to have developed the first 1-bit large language model with 2 billion parameters. The model, BitNet b1.58 2B4T, can run on commercial CPUs such as Apple\u2019s M2. \u201cTrained on a corpus of 4 trillion tokens, this model demonstrates how native 1-bit LLMs can achieve performance comparable to leading open-weight, full-precision models of [&hellip;]\" \/>\n<meta property=\"og:url\" content=\"https:\/\/blog.tecnoartesanos.com\/index.php\/2025\/04\/18\/microsoft-releases-largest-1-bit-llm-letting-powerful-ai-run-on-some-older-hardware\/\" \/>\n<meta property=\"og:site_name\" content=\"TecnoArtesanos Tech Blog\" \/>\n<meta property=\"article:published_time\" content=\"2025-04-18T17:30:03+00:00\" \/>\n<meta property=\"og:image\" content=\"https:\/\/assets.techrepublic.com\/uploads\/2025\/04\/chip-small-ai-apr-25.jpg\" \/>\n<meta name=\"author\" content=\"Sergio Morales\" \/>\n<meta name=\"twitter:card\" content=\"summary_large_image\" \/>\n<meta name=\"twitter:label1\" content=\"Written by\" \/>\n\t<meta name=\"twitter:data1\" content=\"Sergio Morales\" \/>\n\t<meta name=\"twitter:label2\" content=\"Est. reading time\" \/>\n\t<meta name=\"twitter:data2\" content=\"2 minutes\" \/>\n<script type=\"application\/ld+json\" class=\"yoast-schema-graph\">{\"@context\":\"https:\/\/schema.org\",\"@graph\":[{\"@type\":\"WebPage\",\"@id\":\"https:\/\/blog.tecnoartesanos.com\/index.php\/2025\/04\/18\/microsoft-releases-largest-1-bit-llm-letting-powerful-ai-run-on-some-older-hardware\/\",\"url\":\"https:\/\/blog.tecnoartesanos.com\/index.php\/2025\/04\/18\/microsoft-releases-largest-1-bit-llm-letting-powerful-ai-run-on-some-older-hardware\/\",\"name\":\"Microsoft Releases Largest 1-Bit LLM, Letting Powerful AI Run on Some Older Hardware - TecnoArtesanos Tech Blog\",\"isPartOf\":{\"@id\":\"https:\/\/blog.tecnoartesanos.com\/#website\"},\"primaryImageOfPage\":{\"@id\":\"https:\/\/blog.tecnoartesanos.com\/index.php\/2025\/04\/18\/microsoft-releases-largest-1-bit-llm-letting-powerful-ai-run-on-some-older-hardware\/#primaryimage\"},\"image\":{\"@id\":\"https:\/\/blog.tecnoartesanos.com\/index.php\/2025\/04\/18\/microsoft-releases-largest-1-bit-llm-letting-powerful-ai-run-on-some-older-hardware\/#primaryimage\"},\"thumbnailUrl\":\"https:\/\/blog.tecnoartesanos.com\/wp-content\/uploads\/2025\/04\/microsoft-releases-largest-1-bit-llm-letting-powerful-ai-run-on-some-older-hardware.jpg\",\"datePublished\":\"2025-04-18T17:30:03+00:00\",\"author\":{\"@id\":\"https:\/\/blog.tecnoartesanos.com\/#\/schema\/person\/ec88bc1410fd158963717c4216f04807\"},\"breadcrumb\":{\"@id\":\"https:\/\/blog.tecnoartesanos.com\/index.php\/2025\/04\/18\/microsoft-releases-largest-1-bit-llm-letting-powerful-ai-run-on-some-older-hardware\/#breadcrumb\"},\"inLanguage\":\"en-US\",\"potentialAction\":[{\"@type\":\"ReadAction\",\"target\":[\"https:\/\/blog.tecnoartesanos.com\/index.php\/2025\/04\/18\/microsoft-releases-largest-1-bit-llm-letting-powerful-ai-run-on-some-older-hardware\/\"]}]},{\"@type\":\"ImageObject\",\"inLanguage\":\"en-US\",\"@id\":\"https:\/\/blog.tecnoartesanos.com\/index.php\/2025\/04\/18\/microsoft-releases-largest-1-bit-llm-letting-powerful-ai-run-on-some-older-hardware\/#primaryimage\",\"url\":\"https:\/\/blog.tecnoartesanos.com\/wp-content\/uploads\/2025\/04\/microsoft-releases-largest-1-bit-llm-letting-powerful-ai-run-on-some-older-hardware.jpg\",\"contentUrl\":\"https:\/\/blog.tecnoartesanos.com\/wp-content\/uploads\/2025\/04\/microsoft-releases-largest-1-bit-llm-letting-powerful-ai-run-on-some-older-hardware.jpg\",\"width\":1400,\"height\":933},{\"@type\":\"BreadcrumbList\",\"@id\":\"https:\/\/blog.tecnoartesanos.com\/index.php\/2025\/04\/18\/microsoft-releases-largest-1-bit-llm-letting-powerful-ai-run-on-some-older-hardware\/#breadcrumb\",\"itemListElement\":[{\"@type\":\"ListItem\",\"position\":1,\"name\":\"Home\",\"item\":\"https:\/\/blog.tecnoartesanos.com\/\"},{\"@type\":\"ListItem\",\"position\":2,\"name\":\"Microsoft Releases Largest 1-Bit LLM, Letting Powerful AI Run on Some Older Hardware\"}]},{\"@type\":\"WebSite\",\"@id\":\"https:\/\/blog.tecnoartesanos.com\/#website\",\"url\":\"https:\/\/blog.tecnoartesanos.com\/\",\"name\":\"TecnoArtesanos Tech Blog\",\"description\":\"\",\"potentialAction\":[{\"@type\":\"SearchAction\",\"target\":{\"@type\":\"EntryPoint\",\"urlTemplate\":\"https:\/\/blog.tecnoartesanos.com\/?s={search_term_string}\"},\"query-input\":{\"@type\":\"PropertyValueSpecification\",\"valueRequired\":true,\"valueName\":\"search_term_string\"}}],\"inLanguage\":\"en-US\"},{\"@type\":\"Person\",\"@id\":\"https:\/\/blog.tecnoartesanos.com\/#\/schema\/person\/ec88bc1410fd158963717c4216f04807\",\"name\":\"Sergio Morales\",\"image\":{\"@type\":\"ImageObject\",\"inLanguage\":\"en-US\",\"@id\":\"https:\/\/blog.tecnoartesanos.com\/#\/schema\/person\/image\/\",\"url\":\"https:\/\/secure.gravatar.com\/avatar\/3d45178fc8fbbe32d39278bc504fa9093f947f406ff4f1ddcfa27505ab772184?s=96&d=mm&r=g\",\"contentUrl\":\"https:\/\/secure.gravatar.com\/avatar\/3d45178fc8fbbe32d39278bc504fa9093f947f406ff4f1ddcfa27505ab772184?s=96&d=mm&r=g\",\"caption\":\"Sergio Morales\"},\"sameAs\":[\"https:\/\/sergiomorales.space\"],\"url\":\"https:\/\/blog.tecnoartesanos.com\/index.php\/author\/sergiomorales\/\"}]}<\/script>\n<!-- \/ Yoast SEO plugin. -->","yoast_head_json":{"title":"Microsoft Releases Largest 1-Bit LLM, Letting Powerful AI Run on Some Older Hardware - TecnoArtesanos Tech Blog","robots":{"index":"index","follow":"follow","max-snippet":"max-snippet:-1","max-image-preview":"max-image-preview:large","max-video-preview":"max-video-preview:-1"},"canonical":"https:\/\/blog.tecnoartesanos.com\/index.php\/2025\/04\/18\/microsoft-releases-largest-1-bit-llm-letting-powerful-ai-run-on-some-older-hardware\/","og_locale":"en_US","og_type":"article","og_title":"Microsoft Releases Largest 1-Bit LLM, Letting Powerful AI Run on Some Older Hardware - TecnoArtesanos Tech Blog","og_description":"Microsoft researchers claim to have developed the first 1-bit large language model with 2 billion parameters. The model, BitNet b1.58 2B4T, can run on commercial CPUs such as Apple\u2019s M2. \u201cTrained on a corpus of 4 trillion tokens, this model demonstrates how native 1-bit LLMs can achieve performance comparable to leading open-weight, full-precision models of [&hellip;]","og_url":"https:\/\/blog.tecnoartesanos.com\/index.php\/2025\/04\/18\/microsoft-releases-largest-1-bit-llm-letting-powerful-ai-run-on-some-older-hardware\/","og_site_name":"TecnoArtesanos Tech Blog","article_published_time":"2025-04-18T17:30:03+00:00","og_image":[{"url":"https:\/\/assets.techrepublic.com\/uploads\/2025\/04\/chip-small-ai-apr-25.jpg","type":"","width":"","height":""}],"author":"Sergio Morales","twitter_card":"summary_large_image","twitter_misc":{"Written by":"Sergio Morales","Est. reading time":"2 minutes"},"schema":{"@context":"https:\/\/schema.org","@graph":[{"@type":"WebPage","@id":"https:\/\/blog.tecnoartesanos.com\/index.php\/2025\/04\/18\/microsoft-releases-largest-1-bit-llm-letting-powerful-ai-run-on-some-older-hardware\/","url":"https:\/\/blog.tecnoartesanos.com\/index.php\/2025\/04\/18\/microsoft-releases-largest-1-bit-llm-letting-powerful-ai-run-on-some-older-hardware\/","name":"Microsoft Releases Largest 1-Bit LLM, Letting Powerful AI Run on Some Older Hardware - TecnoArtesanos Tech Blog","isPartOf":{"@id":"https:\/\/blog.tecnoartesanos.com\/#website"},"primaryImageOfPage":{"@id":"https:\/\/blog.tecnoartesanos.com\/index.php\/2025\/04\/18\/microsoft-releases-largest-1-bit-llm-letting-powerful-ai-run-on-some-older-hardware\/#primaryimage"},"image":{"@id":"https:\/\/blog.tecnoartesanos.com\/index.php\/2025\/04\/18\/microsoft-releases-largest-1-bit-llm-letting-powerful-ai-run-on-some-older-hardware\/#primaryimage"},"thumbnailUrl":"https:\/\/blog.tecnoartesanos.com\/wp-content\/uploads\/2025\/04\/microsoft-releases-largest-1-bit-llm-letting-powerful-ai-run-on-some-older-hardware.jpg","datePublished":"2025-04-18T17:30:03+00:00","author":{"@id":"https:\/\/blog.tecnoartesanos.com\/#\/schema\/person\/ec88bc1410fd158963717c4216f04807"},"breadcrumb":{"@id":"https:\/\/blog.tecnoartesanos.com\/index.php\/2025\/04\/18\/microsoft-releases-largest-1-bit-llm-letting-powerful-ai-run-on-some-older-hardware\/#breadcrumb"},"inLanguage":"en-US","potentialAction":[{"@type":"ReadAction","target":["https:\/\/blog.tecnoartesanos.com\/index.php\/2025\/04\/18\/microsoft-releases-largest-1-bit-llm-letting-powerful-ai-run-on-some-older-hardware\/"]}]},{"@type":"ImageObject","inLanguage":"en-US","@id":"https:\/\/blog.tecnoartesanos.com\/index.php\/2025\/04\/18\/microsoft-releases-largest-1-bit-llm-letting-powerful-ai-run-on-some-older-hardware\/#primaryimage","url":"https:\/\/blog.tecnoartesanos.com\/wp-content\/uploads\/2025\/04\/microsoft-releases-largest-1-bit-llm-letting-powerful-ai-run-on-some-older-hardware.jpg","contentUrl":"https:\/\/blog.tecnoartesanos.com\/wp-content\/uploads\/2025\/04\/microsoft-releases-largest-1-bit-llm-letting-powerful-ai-run-on-some-older-hardware.jpg","width":1400,"height":933},{"@type":"BreadcrumbList","@id":"https:\/\/blog.tecnoartesanos.com\/index.php\/2025\/04\/18\/microsoft-releases-largest-1-bit-llm-letting-powerful-ai-run-on-some-older-hardware\/#breadcrumb","itemListElement":[{"@type":"ListItem","position":1,"name":"Home","item":"https:\/\/blog.tecnoartesanos.com\/"},{"@type":"ListItem","position":2,"name":"Microsoft Releases Largest 1-Bit LLM, Letting Powerful AI Run on Some Older Hardware"}]},{"@type":"WebSite","@id":"https:\/\/blog.tecnoartesanos.com\/#website","url":"https:\/\/blog.tecnoartesanos.com\/","name":"TecnoArtesanos Tech Blog","description":"","potentialAction":[{"@type":"SearchAction","target":{"@type":"EntryPoint","urlTemplate":"https:\/\/blog.tecnoartesanos.com\/?s={search_term_string}"},"query-input":{"@type":"PropertyValueSpecification","valueRequired":true,"valueName":"search_term_string"}}],"inLanguage":"en-US"},{"@type":"Person","@id":"https:\/\/blog.tecnoartesanos.com\/#\/schema\/person\/ec88bc1410fd158963717c4216f04807","name":"Sergio Morales","image":{"@type":"ImageObject","inLanguage":"en-US","@id":"https:\/\/blog.tecnoartesanos.com\/#\/schema\/person\/image\/","url":"https:\/\/secure.gravatar.com\/avatar\/3d45178fc8fbbe32d39278bc504fa9093f947f406ff4f1ddcfa27505ab772184?s=96&d=mm&r=g","contentUrl":"https:\/\/secure.gravatar.com\/avatar\/3d45178fc8fbbe32d39278bc504fa9093f947f406ff4f1ddcfa27505ab772184?s=96&d=mm&r=g","caption":"Sergio Morales"},"sameAs":["https:\/\/sergiomorales.space"],"url":"https:\/\/blog.tecnoartesanos.com\/index.php\/author\/sergiomorales\/"}]}},"_links":{"self":[{"href":"https:\/\/blog.tecnoartesanos.com\/index.php\/wp-json\/wp\/v2\/posts\/822","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/blog.tecnoartesanos.com\/index.php\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/blog.tecnoartesanos.com\/index.php\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/blog.tecnoartesanos.com\/index.php\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/blog.tecnoartesanos.com\/index.php\/wp-json\/wp\/v2\/comments?post=822"}],"version-history":[{"count":0,"href":"https:\/\/blog.tecnoartesanos.com\/index.php\/wp-json\/wp\/v2\/posts\/822\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/blog.tecnoartesanos.com\/index.php\/wp-json\/wp\/v2\/media\/823"}],"wp:attachment":[{"href":"https:\/\/blog.tecnoartesanos.com\/index.php\/wp-json\/wp\/v2\/media?parent=822"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/blog.tecnoartesanos.com\/index.php\/wp-json\/wp\/v2\/categories?post=822"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/blog.tecnoartesanos.com\/index.php\/wp-json\/wp\/v2\/tags?post=822"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}