Run local LLMs like gpt-oss, Qwen3, Gemma3, DeepSeek and many more on your computer, privately and for free.


LM Studio is the app I now use for running local LLMs.[^1] In LM Studio I mainly use gemma‑3‑12b, mostly for alt‑text generation, and the gpt‑oss‑20b model. Both run fine on my Mac equipped with 32 GB of RAM.

Screenshot Homepage LM Studio (2025-11-09)

lmstudio.ai/

[^1]: I switched from Ollama because LM Studio is much more versatile, and I don’t get their “sign‑up, log‑in, and run remote LLMs” options for a software that was designed for offline usage.