#BITNET

AIagent.at 🤖 AI Newsai@defcon.social
2026-03-11

Learn how to install bitnet.cpp, download the BitNet b1.58 model, and run a fully local AI chat and inference server on your machine. This beginner guide shows you how to run tiny AI models without cloud dependencies. kdnuggets.com/run-tiny-ai-mode #AIagent #AI #GenAI #AIInfrastructure #BitNet

Hacker Newsh4ckernews
2026-03-11
AI Daily Postaidailypost
2026-03-10

🚀 Want to run BitNet-b1.58-2B-4T locally? The new setup_env.py script automates a CMake build of the C++ backend, turning Python-driven setup into a fast inference engine. Perfect for hobbyists and researchers eager to experiment with large AI models offline. Dive into the details and see how easy open-source deployment can be!

🔗 aidailypost.com/news/python-se

2026-01-25

Tạo động cơ LLM 1.58-bit chạy 117 token/giây trên 1 nhân CPU với Rust và AVX-512, nhưng bị lỗi ở lớp Activation khiến đầu ra luôn là <unk>. Cần hỗ trợ về: (1) Weight tying trong BitNet – thiếu hệ số tỉ lệ? (2) Cách scale tích lũy nguyên từ VPOPCNTDQ trước khi đưa vào RMSNorm/SiLU. Dự án mã nguồn mở, zero-copy, không heap allocation. #Rust #AVX512 #LLM #MachineLearning #AI #R3Engine #BitNet #LocalAI #HPC #Inference #trítuệnhân tạo #môhìnhtonngẫu #xửlýsongsong #tinhoccao

https://www.reddit.

2026-01-22

Chúng ta đang chuyển từ thời đại MatMul sang “AI cộng dồn” với BitNet (trọng số ternary), L‑Mul (thêm thay cho nhân) và mHC (đảm bảo ổn định quy mô). Nếu chạy mô hình 70B+ chỉ dùng 1/100 năng lượng, GPU hiện tại sẽ trở thành lạc hậu, cần ASIC chuyên cộng. Các bạn có nghĩ nên dừng mua GPU và tập trung vào kiến trúc cộng không? #AI #AdditiveAI #BitNet #L_Mul #mHC #CôngNghệ #TríTuệNhânTạo

reddit.com/r/LocalLLaMA/commen

Simonesimone_z
2025-06-22

on a with 128MB of ram? Yes. Up to 15M parameters, using architecture, which uses ternary weights (-1, 0, 1) to reduce computational complexity.

From: mastodon.social/@mindsConnecte

오래된 미래: Mälure_eul
2025-05-31

Their public availability allows for widespread experimentation and adaptation. However, a significant barrier hinders their broader adoption: the substantial computational resources required for deployment and inference. State-of-the-art open LLMs typically require large memory footprints, consume considerable energy, and exhibit notable inference latency, rendering them impractical for many edge devices, resource-constrained environments, and real-time applications.

2025-05-23

Weird thing is that I am on a martial arts mailing list (originally created to mock the newbie rec.martial-arts poseurs) that I have been on since 1987 and I am by far the youngest member of the group. I have no idea why they invited me, weird old croaks probably just wanted a youngster perspective. Everyone on there is still alive and kicking though - not very high kicking, but still!

#usenet #bitnet #history #martialarts #wiseguys #meikdo

szymonskszymon
2025-04-27

🔬🤯 Modele 1-bitowe to rewolucja w AI! Wagi sieci neuronowej zapisujemy tylko 1 bitem – zamiast 32 czy 16. To nawet 16x mniejszy rozmiar i ogromne oszczędności energii, przy zachowaniu jakości klasycznych LLM. Przyszłość AI jest lekka! 🚀#AI

Kyesos - The Big Gameovskikyesos.bsky.social@bsky.brid.gy
2025-04-20

Rassurez-vous : les auteurs ont tous été payés. ^^' #OhWait! #AI #BitNet #SpywareWithASmile ^^'

Anything that helps reduce the environmental impacts of LLMs is a good thing.
bitnet.cpp is the official inference framework for 1-bit LLMs (e.g., BitNet b1.58). It offers a suite of optimized kernels, that support fast and lossless inference of 1.58-bit models on CPU (with NPU and GPU support coming next).

The first release of bitnet.cpp is to support inference on CPUs. bitnet.cpp achieves speedups of 1.37x to 5.07x on ARM CPUs, with larger models experiencing greater performance gains. Additionally, it reduces energy consumption by 55.4% to 70.0%, further boosting overall efficiency. On x86 CPUs, speedups range from 2.37x to 6.17x with energy reductions between 71.9% to 82.2%. Furthermore, bitnet.cpp can run a 100B BitNet b1.58 model on a single CPU, achieving speeds comparable to human reading (5-7 tokens per second), significantly enhancing the potential for running LLMs on local devices.
https://github.com/microsoft/BitNet #BitNet
Andreas BeckerCaramba1
2025-04-19

1-Bit statt Milliarden Parameter: Microsofts BitNet b1.58 zeigt, dass KI auch ohne High-End-Hardware leistungsfähig sein kann. Ein radikaler Ansatz mit Potenzial für mehr Nachhaltigkeit und Zugänglichkeit. Ist das der Anfang vom Ende der GPU-Abhängigkeit? 👉 all-ai.de/news/top-news24/bitn

st1nger :unverified: 🏴‍☠️ :linux: :freebsd:st1nger@infosec.exchange
2025-04-18

#Microsoft has introduced #BitNet b1.58 2B4T, the largest-scale 1-bit #AI model to date with 2 billion parameters and the ability to run efficiently on #CPU - It's openly available under an MIT license huggingface.co/microsoft/bitne

N-gated Hacker Newsngate
2025-04-17

🎉🎊 Behold, the groundbreaking 587th iteration of , where meet their ultimate destiny: "Technical Report"! A riveting tale of acronyms, citations, and a plea for , all while you desperately try to figure out if those numbers actually mean anything 📊🤯. Remember, it's not a real tech report without a job ad and some grateful acknowledgments for funding! 💰👏
arxiv.org/abs/2504.12285

Mr Tech Kingmrtechking
2025-04-17

Microsoft open-sourced BitNet b1.58, a big 2B param 1-bit AI. Super efficient on CPUs like M2, it rivals similar models using less memory/speed. Needs a custom framework for now, GPU support pending.

New Microsoft AI Runs Super Efficiently, Even on CPUs.

Was looking at the source to a very early arXiv paper (arxiv.org/abs/hep-ph/9210243). The PDF is unavailable, for reasons that are obscure ("pre-1996 submission which cannot be processed"). But there's a lot of history in the source code: it looks like it was submitted, as a single file, emailed from BITNET to the arXiv via a gateway. It also uses a now-obscure TeX package phyzzx (ctan.org/tex-archive/obsolete/).

I know I'll sound like a young person when I say this but I'd love to know how that worked in practice and what it was like to be in academia before everyone had access to a TCP/IP internet connection but after internetworked computers were ubiquitous. Sort of like the TV series Halt and Catch Fire but with physicists.

#arXiv #BITNET #TeX

Screenshot of source code to arXiv submission. The visible lines read:

%Paper: hep-ph/9210243
%From: <GRADWOHL%UCLAHEP.BITNET@CORNELLC.cit.cornell.edu>
%Date: Fri, 16 Oct 92 15:19 PDT

%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
%  1.) The file has to be "TeXed", using the macropackage `phyzzx'.
%  2.) The 5 figures are available on request. They can be received
%      through e-mail in form of PS files, or by regular mail.
%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%

%\input phyzzx

\hsize 6.5truein
\vsize 9.0truein
\hoffset 0.2truein
\voffset 0.06truein

Client Info

Server: https://mastodon.social
Version: 2025.07
Repository: https://github.com/cyevgeniy/lmst