TechCrunch News 04月16日 23:51
Microsoft researchers say they’ve developed a hyper-efficient AI model that can run on CPUs
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

微软研究人员发布了迄今为止最大的1-bit AI模型BitNet b1.58 2B4T,该模型采用MIT许可证开源。Bitnet是一种压缩模型,设计用于在轻量级硬件上运行。它将权重量化为-1、0和1三个值,使其在内存和计算方面更高效。BitNet b1.58 2B4T在GSM8K和PIQA等基准测试中,表现优于Meta的Llama 3 1B、Google的Gemma 3 1B和阿里巴巴的Qwen 2.5 1.5B。该模型速度更快,内存使用量更少。不过,Bitnet的兼容性是一个挑战,目前仅支持特定硬件,不支持在AI基础设施中占主导地位的GPU。

💡BitNet b1.58 2B4T是微软开发的1-bit AI模型,是迄今为止最大的同类模型,采用了MIT许可证开源,可以在CPU上运行,包括苹果的M2芯片。

💾Bitnet是一种压缩模型,通过将权重量化为-1、0和1三个值,从而实现更低的内存占用和更高的计算效率,使其更适合轻量级硬件。

🚀研究表明,BitNet b1.58 2B4T在GSM8K和PIQA等基准测试中,性能优于Meta的Llama 3 1B、Google的Gemma 3 1B和阿里巴巴的Qwen 2.5 1.5B。同时,该模型在速度上也有优势,某些情况下速度是其他模型的两倍,且内存占用更少。

⚠️BitNet b1.58 2B4T需要使用微软的定制框架bitnet.cpp才能发挥最佳性能,但该框架目前仅支持特定硬件,不支持GPU,这限制了其应用范围。

Microsoft researchers claim they’ve developed the largest-scale 1-bit AI model, also known as a “bitnet,” to date. Called BitNet b1.58 2B4T, it’s openly available under an MIT license and can run on CPUs, including Apple’s M2.

Bitnets are essentially compressed models designed to run on lightweight hardware. In standard models, weights, the values that define the internal structure of a model, are often quantized so the models perform well on a wide range of machines. Quantizing the weights lowers the number of bits — the smallest units a computer can process — needed to represent those weights, enabling models to run on chips with less memory, faster.

Bitnets quantize weights into just three values: -1, 0, and 1. In theory, that makes them far more memory- and computing-efficient than most models today.

The Microsoft researchers say that BitNet b1.58 2B4T is the first bitnet with 2 billion parameters, “parameters” being largely synonymous with “weights.” Trained on a data set of 4 trillion tokens — equivalent to about 33 million books, by one estimate — BitNet b1.58 2B4T outperforms traditional models of similar sizes, the researchers claim.

BitNet b1.58 2B4T doesn’t sweep the floor with rival 2 billion-parameter models, to be clear, but it seemingly holds its own. According to the researchers’ testing, the model surpasses Meta’s Llama 3.2 1B, Google’s Gemma 3 1B, and Alibaba’s Qwen 2.5 1.5B on benchmarks including GSM8K (a collection of grade-school-level math problems) and PIQA (which tests physical commonsense reasoning skills).

Perhaps more impressively, BitNet b1.58 2B4T is speedier than other models of its size — in some cases, twice the speed — while using a fraction of the memory.

There is a catch, however.

Achieving that performance requires using Microsoft’s custom framework, bitnet.cpp, which only works with certain hardware at the moment. Absent from the list of supported chips are GPUs, which dominate the AI infrastructure landscape.

That’s all to say that bitnets may hold promise, particularly for resource-constrained devices. But compatibility is — and will likely remain — a big sticking point.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

BitNet 1-bit AI 微软 AI模型
相关文章