#SmallAI

2025-04-18

TechCrunch: Microsoft researchers say they’ve developed a hyper-efficient AI model that can run on CPUs. “Microsoft researchers claim they’ve developed the largest-scale 1-bit AI model, also known as a ‘bitnet,’ to date. Called BitNet b1.58 2B4T, it’s openly available under an MIT license and can run on CPUs, including Apple’s M2.”

https://rbfirehose.com/2025/04/18/techcrunch-microsoft-researchers-say-theyve-developed-a-hyper-efficient-ai-model-that-can-run-on-cpus/

Alasdair Allanaallan
2024-06-25

Okay. Now I've seen literally everything. An LLM and an inference engine, embedded in a font. fuglede.github.io/llama.ttf/ youtube.com/watch?v=Q4bOyYctgFI

Alasdair Allanaallan
2024-05-31
Alasdair Allanaallan
2024-05-30
Alasdair Allanaallan
2024-05-21

It’s sort of interesting to see the financial markets paying attention to . That means that anything that affects “big” AI is important enough to move the markets. Makes sense, if it’s going to affect the $NVDA stock price, it’ll affect the market. finimize.com/content/small-talk

Matija Šukljehook@toot.si
2023-09-26

#Firefox 118 has local, in-browser machine translations. That’s pretty freaking cool!

> Automated translation of web content is now available to Firefox users! Unlike cloud-based alternatives, translation is done locally in Firefox, so that the text being translated does not leave your machine.

mozilla.org/en-US/firefox/118.

#MachineTranslation #SmallAI

Aurelie Herbelot is movingminimalparts@fosstodon.org
2023-07-10

On My Disk is a personal cloud storage solution that keeps your files secure and private. We are integrating PeARS search into the service to make your files incredibly easy to find, regardless of their physical location.

One ideal shared by PeARS and On My Disk is energy efficiency. So we are dedicated to create 'small AI' solutions that combine old and new machine learning techniques, to create the most climate-friendly NLP tools 🍃 🍐 #smallai #NLP

2022-09-16

#AI #SmallAI #SelfTrain

One of the arguments in favour of surveillance capitalism is the great usefulness of cloud-based ML predictions.

After all, who can deny the usefulness of photo apps that automatically recognize faces, detect your speech or help you making sense of the deluge of information in a social feed?

The argument usually goes like this: these features require large neural networks, which in turn require a lot of computational power to train the models, and a lot of memory and disk storage in order to load and save those models.

You can't do such things on small devices that run on batteries. Therefore your phone *HAS* to send your data to #BigTech servers if you want to get those features. Otherwise, you just won't get those features.

Except that... What if this whole argument is bollocks?

#POET (Private Optimal Energy Training) proves that you can run both the training and the predictions locally, without compromising neither on precision, nor on performance.

After all, the really expensive part of training is back-propagation. POET breaks down the back-propagation performance issue by quantizing the layers (so real-number large tensor multiplications get reduced to smaller multiplications of integer tensors, without sacrificing precision too much), and a clever way of caching the layers that are most likely to be needed, so we don't have to recalculate them, without caching everything though (which would be prohibitive in terms of storage).

The arguments in the paper sound very convincing to me. The code is publicly available on Github. I haven't yet had time to test it myself, but I will quite soon - and try to finally build an alternative voice assistant that can completely run on my phone.

proceedings.mlr.press/v162/pat

Client Info

Server: https://mastodon.social
Version: 2025.04
Repository: https://github.com/cyevgeniy/lmst