Chris Messina

Run OpenAI's gpt-oss locally in LM Studio

Want to run @OpenAI's open models locally? Now you can with LM Studio!

The model comes in 2 sizes: 20B and 120B.

Around ~13GB of RAM is needed for the smaller model.

  • gpt-oss-20b — for lower latency, and local or specialized use-cases (21B parameters with 3.6B active parameters)

  • gpt-oss-120b — for production, general purpose, high reasoning use-cases that fits into a single H100 GPU (117B parameters with 5.1B active parameters)

149 views

Add a comment

Replies

Best
Iacop Hafiane

This is huge. Running serious models locally without relying on cloud APIs opens up a ton of possibilities for speed, privacy, and control.

What kind of workflows are you most excited to unlock with this setup?

Jeffrey Chiang

I like how @Ollama also shows these new models by default on the new update.

Rohan Gayen

Around ~13GB of RAM is needed for the smaller model. = RAM or GPU?