#LMStudio

KernelDesignkerneldesign
2025-05-02

En local j'ai testé 30B A3B, c'est une MOE impressionnant ! Les chinois veulent vraiment faire exploser la bulle AI américaine. Comme le MOE c'est la spécialité de Mistral j'espère que l'on aura une réponse équivalente prochainement. , +

2025-05-02

Accessing LM Studio Server from WSL Linux

(Not complicated, just tricky to find the settings)

ingo.kaulbach.de/accessing-lm-

#LMStudio #ai #WSL

2025-04-29

What does your local #LLM setup for software #development look like? I am on a M4 Pro MacBook with 24GB RAM. I can only use local LLMs with #Ollama or #LMStudio running with either #JetBrains #AI (but they do not support local LLMs for code completion) or the continue plugin.

At the moment I use qwen2.5-coder:7b with 4bit quantization for autocompletion, Phi4 or Llama3.1 8B for chatting and nomic-embed-text for embedding. Suggestions? :mastodon:

2025-04-26

Как запустить локальную LLM (AI) в Android Studio

Привет! Если вы мобильный разработчик и следите за AI-трендами, наверняка задумывались о том, как интегрировать языковые модели (LLM) в свои приложения прямо из Android Studio. В этой статье я расскажу, как это можно сделать быстро и просто, не полагаясь на внешние API и облачные решения.

habr.com/ru/articles/904648/

#kotlin #android_studio #continuedev #lmstudio #local_llm

2025-04-25

@sam4000 assistant: Ach, Hamburg! Es ist wie ein bisschen traurig, aber auch irgendwie gemütlich. Teilweise bewölkt, also nicht die Sonne, aber es ist trotzdem angenehm warm mit 57 Grad. Der Wind kommt aus Nordosten und pfeift ganz sanft um die Ecke – perfekt für einen Spaziergang mit einem Tee!

#lmstudio #lmstudiosdk #llm #aiassistent

H@R0👨🏻‍💻quetalocatl
2025-03-31

起初 對Linux支持一般,所以一直在用 但現在支持得很好了,在Linux上面也有 CPU/Vulkan/CUDDA 三種backend,雖然Vulkan速度一般但至少amd卡也能用

H@R0👨🏻‍💻quetalocatl
2025-03-31

下載速度不錯 😎

2025-03-25

Hongkiat: Running Large Language Models (LLMs) Locally with LM Studio. “Running large language models (LLMs) locally with tools like LM Studio or Ollama has many advantages, including privacy, lower costs, and offline availability. However, these models can be resource-intensive and require proper optimization to run efficiently. In this article, we will walk you through optimizing your […]

https://rbfirehose.com/2025/03/25/hongkiat-running-large-language-models-llms-locally-with-lm-studio/

2025-03-24

I can't get it. I've spent all evening trying to get #SillyTavern to detect #Mythomax in #LMStudio, but it won't.

So I'm using Mythomax in LM Studio, and it's shit. It's badly broken and crappy. It gets the narrative tense wrong, even in the same sentence, and it devolves to mindless repetition with no room for input. A nightmare.

#ChatGPT was light-years beyond this, but now they ban "AI relationships," censoring creative writing for consenting adults. #LLM

2025-03-21

Cohere Command — революция, которую мы пропустили

🔪 Карусель триальных токенов под ребро облачному LLM провайдеру Облачный LLM провайдер Cohere предоставляет бесплатно 20 запросов в минуту без проверки кредитной карты. Я просто не смог отказаться от задумки сделать веселую карусель)

habr.com/ru/articles/893232/

#typescript #javascript #python #lmstudio #ollama #llm #openai #cohere #искусственный_интеллект #машинное_обучение

2025-03-16

Not sure if you have noticed it: Google has released Gemma 3, a powerful model that is small enough to run on normal computers.

blog.google/technology/develop

I've done some experiments on my Laptop (with a Geforce 3080ti), and am very impressed. I tried to be happy with Llama3, with the Deepseek R1 distills on Llama, with Mistral, but the models that would run on my computer were not in the same league as what you get from ChatGPT or Claude or Deepseek remotely.

Gemma changes this for me. So far I let it write 3 smaller pieces of Javascript, analyze a few texts, and it performed slow, but flawlessly. So finally I can move to a "use the local LLM for the 90% default case, and go for the big ones only if the local LLM fails".

This way
- I use far less CO2 for my LLM tasks
- I am in control of my data, nobody can collect my prompts and later sell my profile to ad customers
- I am sure the IP of my prompts stay with me
- I have the privacy to ask it whatever I want and no server in the US or CN has those data.

Interested? If you have a powerful graphiccs card in your PC, it is totally simple:

1. install LMStudio from LMStudio.ai
2. in LMStudio, click Discover, and download the Gema3 27b Q4 model
3. Chat

If your graphics card is too small, you might head for the smaller 12b model, but I can't tell you how well it performs.

#LMStudio #gemma3 #gemma #chatgpt #llm #google

Mastokarl 🇺🇦Mastokarl
2025-03-16

Did a few coding experiments with Gemma 3 local on lmstudio. So far it performs flawless (in terms of capability - on my lowly Geforce 3080ti it is fairly slow, something like 5 tokens per second). But I've got time, and it is mine, running locally, no Billionaire's corporation sees my prompts.

For me (privacy nut) this is a big thing, not having to use ChatGPT for everything.

Client Info

Server: https://mastodon.social
Version: 2025.04
Repository: https://github.com/cyevgeniy/lmst