#SmallLanguageModels

2025-06-09

It is about time we stop the use of big speak tables and instead try small speak tables.

I think the ten hundred words used for the Up Goer Five flying space car blue picture will work nice.

Up Goer Five blue picture:
xkcd.com/1133/

@theosanderson's nice Up Goer Five word look up thing:
splasho.com/upgoer5/

#LargeLanguageModels #LLM #SmallLanguageModels

Bigger isn’t always better, #SmallLanguageModels are gaining traction as cheaper, faster options that excel in targeted, domain-specific tasks. It’s about choosing the right-sized tool for the job. #AI #SLM #NLP #MachineLearning #ResearchSky

Small Language Models Are the ...

2025-04-07

🤖 🗣️ Small language models are more reliable and secure than their large counterparts, primarily because they draw information from a circumscribed dataset. Expect to see more chatbots running on these slimmed-down alternatives in the coming months.

#SmallLanguageModels #Robots #Innovation

Read more: go.epfl.ch/b7s-en

Doug Ortizdougortiz
2025-03-28

📱 Small Language Models (SLMs) are gaining traction! 🧠

These compact AI models run efficiently on local devices while offering impressive capabilities:
- Enhanced privacy (your data stays on your device)
- Lower computational costs
- Minimal latency
- Works offline

Perfect for specific tasks where you don't need the full power of massive models.

Have you tried any SLMs yet?

2025-01-14

Key Points:
➡️ SLMs with 1-8B parameters can perform as well or better than LLMs.
➡️ SLMs are task-agnostic or task-specific.
➡️ SLMs balance performance, efficiency, scalability, and cost.
➡️ SLMs are effective in resource-constrained environments.
➡️ SLMs can be trained on consumer-grade GPUs.
➡️ SLMs include models like #Llama2, #Mistral, #Phi, and #Gemini.

arxiv.org/abs/2501.05465

#SLM #LLM #AI #MachineLearning #ArtificialIntelligence #Scalability #Performance #GPU #SmallLanguageModels

2024-12-23

Hugging Face shows how test-time scaling helps small language models punch above their weight venturebeat.com/ai/hugging-fac #AI #SmallLanguageModels

Victoria Stuart 🇨🇦 🏳️‍⚧️persagen
2024-12-16

Phi-4: Microsoft’s Newest Small Language Model Specializing in Complex Reasoning
techcommunity.microsoft.com/bl
arxiv.org/abs/2412.08905
news.ycombinator.com/item?id=4

* most language models" pre-training based primarily on organic data sources such as web content or code
* phi-4 strategically incorp. synthetic data throughout training
* strong performance rel. its size, esp. on reasoning-focused benchmarks


Phi-4: Microsoft’s Newest Small Language Model Specializing in Complex Reasoning
https://techcommunity.microsoft.com/blog/aiplatformblog/introducing-phi-4-microsoft%E2%80%99s-newest-small-language-model-specializing-in-comple/4357090
https://arxiv.org/abs/2412.08905
https://news.ycombinator.com/item?id=42405323

* most language models" pre-training based primarily on organic data sources such as web content or code
* phi-4 strategically incorp. synthetic data throughout training
* strong performance rel. its size, esp. on reasoning-focused benchmarks

#LLM #SLM #SmallLanguageModels #LanguageModels #NLP #ML #AI
#Microsoft #Phi3 #Phi4 #SyntheticData
Victoria Stuart 🇨🇦 🏳️‍⚧️persagen
2024-12-16

[thread] Small language models
see also: en.wikipedia.org/wiki/Large_la
ibm.com/think/topics/small-lan

* machine learning models
* processing, understanding, generating natural language content
* SLM more compact/efficient than LLM: large language models
* few million to few billion parameters vs LLM: 100B's - trillions
* parameters: internal variables that a model learns during training
* influence how model behaves/performs


2024-11-11

#EdgeComputing is like the cool kid on the tech block, handling data right where it’s created instead of sending everything back to the cloud.

The result? Faster decision-making, reduced bandwidth usage, and enhanced privacy.

But here’s the catch: edge devices often operate under strict constraints regarding processing power, memory, and energy consumption.

💡 Enter … #SmallLanguageModels (SLMs) - the efficient sidekick to save the day.

In this #InfoQ article, Suruchi Shah explores how SLMs can work their magic by learning and adapting to patterns in real-time, reducing the computational burden, and making edge devices smarter without asking for much in return: bit.ly/4fIQsDI

#AI #GenerativeAI #LLMs

2024-11-05

Dives into the world of #GenerativeAI & #SmallLanguageModels - #SLMs!

"So latest trend in the language model evolution are the small language models or SLMs that offer many of the same benefits as LLMs, but they're smaller in size, they're trained using smaller data sets and they don't require a lot of computing resources." - Namee Oberst, Co-Founder of LLMWare

🎧 Listen now: bit.ly/4f9M5Sz

#AI #RAG #LLMs #EdgeComputing

Boston Managed ITbmit
2024-07-25

What's the newest trend in generative AI? It's going from big to small. Learn how small language models (SLMs) are starting to take over from their larger counterparts.

venturebeat.com/ai/why-small-l

While large language AI models continue to make headlines, small language models are where the action is. At least, that’s what Meta appears to be betting on, according to a team of its research scientists. #AI #SmallLanguageModels #SLM #Meta
tinyurl.com/mr3ja28s

Ferdinando Simonetti rimmon1971@misskey.social
2024-02-17

Client Info

Server: https://mastodon.social
Version: 2025.04
Repository: https://github.com/cyevgeniy/lmst