#LocalLLaMA

Introducing Mistral Small 4 | Mistral AI

aussie.zone/post/30679863

LLM Architecture Gallery

lemdro.id/post/37634842

Short Doco: How LLMs Took Over The World - Everything is a Pattern

aussie.zone/post/30630219

2026-03-14

I am wondering, is the path big AI corps are going with providing models via huge server farms quite opposing capitalism?
Normally costs run down over time (see solar or microchips). LLMs get smaller and suddenly they fit on your device.

I checked OVH cloud for their offerings of cloud models. They all fit on a 64gb strix halo, probably even 32gb ram. The SOTA models still have an edge, but honestly not much.

#localllm #localllama

CanIRun.ai — Can your machine run AI models?

discuss.tchncs.de/post/56526579

llama.cpp + mcp - docker and more

lemmy.zip/post/60682562

2026-03-12

How to... (Maybe I am missing something)

downonthestreet.eu/post/560800

Guide to run Qwen3.5 locally

communick.news/post/5598650

Anyone interested in AI radio?

lemmy.world/post/43936980

A possible hardware solution for ultra speed (73x faster than H200) self hosted small models that is not dependent on RAM

lemmy.ca/post/61063772

PewDiePie trains his own AI

lemdro.id/post/36894302

llmfit - find best model that runs on your computer

lemmy.ml/post/43700780

2026-02-24

Smaller qwen3.5 models released

lemmy.ml/post/43653865

2026-02-21

ollama 0.17 Released With Improved OpenClaw Onboarding

lemy.lol/post/61478766

ggml.ai (the founding team of llama.cpp) joins Hugging Face to ensure the long-term progress of Local AI

aussie.zone/post/29869105

Artificial Analysis Intelligence Index and cost benchmarks are useful decision/guidance determinants for which models to use. Analysis for top models.

lemmy.ca/post/60678292

I Sold Out for $20 a Month and All I Got Was This Perfectly Generated Terraform

aussie.zone/post/29731374

Qwen3.5: Towards Native Multimodal Agents

aussie.zone/post/29725602

gpt-oss:20b running in your browser thanks to Transformers.js v4 and ONNX Runtime Web

lemdro.id/post/36295188

cedric (@cedric_chee)

MiniMax M2.5는 총 230B 파라미터에 활성 파라미터가 단 10B로 작아 홈랩에서 유력한 후보라는 평가입니다. M2 시리즈는 LocalLLaMA 커뮤니티에서 인기를 끌고 있으며, 작성자는 추론 제공자들이 TPS를 극한까지 끌어올려 성능을 보여주길 바란다고 언급했습니다.

x.com/cedric_chee/status/20221

#minimax #m25 #homelab #localllama

Client Info

Server: https://mastodon.social
Version: 2025.07
Repository: https://github.com/cyevgeniy/lmst