#SLMs

Dr. Thompsonrogt_x1997
2025-06-06

๐Ÿš€ Why pay more for cloud AI when smarter AI fits in your watch?
Discover how Small Language Models are quietly outperforming LLMs โ€”
โ€ข 8X faster
โ€ข 90% cheaper
โ€ข 100% offline ๐Ÿคฏ

From Tesla to smart clinics, this is the AI story no one's telling โ€” yet.
Read the full piece ๐Ÿ‘‡
๐Ÿ”— medium.com/@rogt.x1997/8x-fast


medium.com/@rogt.x1997/8x-fast

Miguel Afonso Caetanoremixtures@tldr.nettime.org
2025-06-02

"Ai2 tested DataDecide across a wide range of datasets and model sizes, using 10 benchmarks to evaluate how well small models predict large-scale performance. The findings arenโ€™t earth-shattering, but they present useful takeaways for AI developers and researchers.

For one, Ai2 found that small models (around 150 million parameters) can predict large-scale outcomes with surprising accuracy. Some benchmarks reached over 80% decision accuracy using just 0.01% of the compute compared to billion-parameter models.

Since small-model experiments use less compute than other methods, developers donโ€™t need to run full-scale tests just to predict outcomes. โ€œThe promise of this work is lower compute costs during training,โ€ said Pijanowski.

Ai2 found that scaling laws didnโ€™t outperform the simpler method of ranking datasets by small-model results. Scaling laws, a more sophisticated and more costly testing method, aim to predict how accuracy improves with model size. For now, โ€œjust stick with ablating things at one scale,โ€ advised Magnusson.

The findings should give LLM devs pause for thought, Hunt said: โ€œThere are scaling laws that have been derived from empirical studies between data volume, compute resources and performance. Ai2โ€™s research points out that we may want to revisit some of those assumptions.โ€"

thenewstack.io/new-tools-help-

#AI #GenerativeAI #LLMs #AITraining #SLMs

Alex JimenezAlexJimenez@mas.to
2025-04-13

Small Language Models Are the New Rage, Researchers Say

Larger models can pull off a wider variety of feats, but the reduced footprint of smaller models makes them attractive tools.

wired.com/story/why-researcher

#SLMs #LLMs #AI

2025-03-19

Are you passionate about the latest in #AI? Here's your chance to shine!

โœ๏ธ Join the #InfoQ Annual Article Writing Competition!

๐Ÿ† Win a #FreeTicket to #QCon or #InfoQDevSummit!

๐Ÿ”— Submit by March 30, 2025: bit.ly/417KPtk

Which AI topic are you most excited to explore?

Explore topics like #LLMs, #SLMs, #vLLMs, #GenAI, #VectorDatabases, #ExplainableAI, #RAG, and more!

QCon Software Conferencesqcon@techhub.social
2025-03-19

Are you passionate about the latest in #AI? Here's your chance to shine!

โœ๏ธ Join the #InfoQ Annual Article Writing Competition!

๐Ÿ† Win a #FreeTicket to #QCon or #InfoQDevSummit!

๐Ÿ”— Submit by March 30, 2025: bit.ly/4gKC51N

Which AI topic are you most excited to explore?

Explore topics like #LLMs, #SLMs, #vLLMs, #GenAI, #VectorDatabases, #ExplainableAI, #RAG, and more!

Alex JimenezAlexJimenez@mas.to
2025-02-18

The Big Power of Small #AI in 2025

An IBM VP explains how small language models are a boon for companies of all sizes, enabling them to overcome resource and budget constraints to tap into the business value of AI.

techrepublic.com/article/ibm-s

#SLMs

Alex JimenezAlexJimenez@mas.to
2025-02-01
2025-01-16

This #InfoQ #eMag brings together our most popular InfoQ Trends Reports from 2024, offering a deep dive into:
๐Ÿ’ก Cell-based architectures
๐Ÿ’ก Socio-technical systems
๐Ÿ’ก Large and small language models (LLMs & SLMs)
๐Ÿ’ก State-of-the-art innovations in the Java ecosystem

Whether you're a developer, architect, technology leader, or simply a tech enthusiast, these reports provide actionable insights and valuable perspectives to help you:
๐Ÿš€ Plan your future roadmaps
๐Ÿš€ Explore emerging technologies & practices

๐Ÿ”— Download it for free: bit.ly/3PEiyoG

#TrendsReport #SoftwareTrends #FreeDownload

#SoftwareArchitecture #SoftwareDevelopment #LLMs #SLMs #Java

Miguel Afonso Caetanoremixtures@tldr.nettime.org
2025-01-15

"To prevent AI models from memorizing their input, we know exactly one robust method: differential privacy (DP). But crucially, DP requires you to precisely define what you want to protect. For example, to protect individual people, you must know which piece of data comes from which person in your dataset. If you have a dataset with identifiers, that's easy. If you want to use a humongous pile of data crawled from the open Web, that's not just hard: that's fundamentally impossible.

In practice, this means that for massive AI models, you can't really protect the massive pile of training data. This probably doesn't matter to you: chances are, you can't afford to train one from scratch anyway. But you may want to use sensitive data to fine-tune them, so they can perform better on some task. There, you may be able to use DP to mitigate the memorization risks on your sensitive data.

This still requires you to be OK with the inherent risk of the off-the-shelf LLMs, whose privacy and compliance story boils down to "everyone else is doing it, so it's probably fine?".

To avoid this last problem, and get robust protection, and probably get better resultsโ€ฆ Why not train a reasonably-sized model entirely on data that you fully understand instead?"

desfontain.es/blog/privacy-in-

#AI #GenerativeAI #LLMs #SLMs #Privacy #DifferentialPrivacy #Memorization

2025-01-15

In this edition of #tech on ice, I talk about #github models and how great it is to kick the tires on different #llms
#LLMs #SLMs #ai #developer #programming #coldplunge #technology

tiktok.com/@isaacrlevin/video/

2025-01-13

In this edition of #tech on ice, I discuss the difference between #LLMs vs #SLMs in the #ai space.
#developer #programming #coldplunge #gpt #LLM

tiktok.com/@isaacrlevin/video/

Winbuzzerwinbuzzer
2025-01-10
Alex JimenezAlexJimenez@mas.to
2025-01-04

Small language models: 10 Breakthrough Technologies 2025

Large language models unleashed the power of AI. Now itโ€™s time for more efficient AIs to take over.

technologyreview.com/2025/01/0

#LLMs #SLMs #AI

Miguel Afonso Caetanoremixtures@tldr.nettime.org
2025-01-01

"Rather than building massive, complex large language models, many organizations choose smaller language models that focus on niche applications such as supply chain management or inventory control. This is the โ€œlean AIโ€ concept, and it entails purpose-built models able to deliver value without the high costs and complexity associated with larger systems, according to Linthicum.

โ€œWe have Agentic AI and certainly using things like small language models where weโ€™re leveraging generative AI and AI in general for more tactical implementation,โ€ he said. โ€œItโ€™s dealing with supply chain integration, dealing with inventory control. Weโ€™re not building LLMs, and I donโ€™t think the businesses out there are going to get the value from building huge LLMs that they think theyโ€™re going to get.โ€"

siliconangle.com/2024/12/31/ge

#AI #GenerativeAI #LLMs #SLMs #LeanAI #CIOs #AgenticAI

EduLabsedulabs
2024-12-19

๐Ÿ” ืžื—ืคืฉื™ื ืœืฉื“ืจื’ ืืช ื”ืขืกืง ืฉืœื›ื ืขื ื‘ื™ื ื” ืžืœืื›ื•ืชื™ืช ื—ืกื›ื•ื ื™ืช ื•ื™ืขื™ืœื”? ืžื•ื“ืœื™ื ืงื˜ื ื™ื (SLMs) ื”ื ื”ืคืชืจื•ืŸ ื”ืžื•ืฉืœื! ื”ื ื—ื•ืกื›ื™ื ื‘ืขืœื•ื™ื•ืช, ืžืฉืคืจื™ื ื‘ื™ืฆื•ืขื™ื ื•ืฉื•ืžืจื™ื ืขืœ ื”ื“ืื˜ื” ืฉืœื›ื โ€“ ื›ืœ ื–ื” ืขื ื“ืจื™ืฉื•ืช ืžื—ืฉื•ื‘ ืฆื ื•ืขื•ืช.
ืงืจืื• ืืช ื”ืžืืžืจ ื”ืžืœื:

edulabs.co.il/he/blog/post-202

2024-11-05

Dives into the world of #GenerativeAI & #SmallLanguageModels - #SLMs!

"So latest trend in the language model evolution are the small language models or SLMs that offer many of the same benefits as LLMs, but they're smaller in size, they're trained using smaller data sets and they don't require a lot of computing resources." - Namee Oberst, Co-Founder of LLMWare

๐ŸŽง Listen now: bit.ly/4f9M5Sz

#AI #RAG #LLMs #EdgeComputing

Miguel Afonso Caetanoremixtures@tldr.nettime.org
2024-09-26

#LLMs #SLMs #AI #GenerativeAI #Chatbots: "With the growing attention and investment in recent AI approaches such as large language models, the narrative that the larger the AI system the more valuable, powerful and interesting it is is increasingly seen as common sense. But what is this assumption based on, and how are we measuring value, power, and performance? And what are the collateral consequences of this race to ever-increasing scale? Here, we scrutinize the current scaling trends and trade-offs across multiple axes and refute two common assumptions underlying the โ€˜bigger-is-betterโ€™ AI paradigm: 1) that improved performance is a product of increased scale, and 2) that all interesting problems addressed by AI require large-scale models. Rather, we argue that this approach is not only fragile scientifically, but comes with undesirable consequences. First, it is not sustainable, as its compute demands increase faster than model performance, leading to unreasonable economic requirements and a disproportionate environmental footprint. Second, it implies focusing on certain problems at the expense of others, leaving aside important applications, e.g. health, education, or the climate. Finally, it exacerbates a concentration of power, which centralizes decision-making in the hands of a few actors while threatening to disempower others in the context of shaping both AI research and its applications throughout society."

arxiv.org/html/2409.14160v1

Client Info

Server: https://mastodon.social
Version: 2025.04
Repository: https://github.com/cyevgeniy/lmst