Run local LLMs like gpt-oss, Qwen3, Gemma3, DeepSeek and many more on your computer, privately and for free.

LM Studio is the app I now use for running local LLMs.1 In LM Studio I mainly use gemma‑3‑12b, mostly for alt‑text generation, and the gpt‑oss‑20b model. Both run fine on my Mac equipped with 32 GB of RAM.

Screenshot Homepage LM Studio (2025-11-09)

lmstudio.ai/


  1. I switched from Ollama because LM Studio is much more versatile, and I don’t get their “sign‑up, log‑in, and run remote LLMs” options for a software that was designed for offline usage. ↩︎