Gemma 4 Guides
Does LM Studio Support Gemma 4? Compatibility, Model List, and Requirements

If you are searching for LM Studio Gemma 4 support, the short answer is yes.
As of April 7, 2026, LM Studio's own model catalog has public pages for all four Gemma 4 sizes:
- Gemma 4 E2B
- Gemma 4 E4B
- Gemma 4 26B A4B
- Gemma 4 31B
So the question is no longer "Does LM Studio support Gemma 4?" The better question is what kind of support LM Studio is giving you, and which Gemma 4 model your machine can actually handle.
Does LM Studio support Gemma 4? Short answer
Yes. Google's own Gemma docs link to a Run Gemma with LM Studio integration page, and LM Studio itself has live model pages for the Gemma 4 lineup.
That means LM Studio Gemma 4 support is real, current, and documented by both sides.
Which Gemma 4 models are in LM Studio?
LM Studio currently publishes model pages for:
| Model | LM Studio minimum system memory |
|---|---|
| Gemma 4 E2B | 4 GB |
| Gemma 4 E4B | 6 GB |
| Gemma 4 26B A4B | 17 GB |
| Gemma 4 31B | 19 GB |
For most people, that is the fastest way to sanity-check fit before they download anything.
What LM Studio is actually supporting
This is the part that matters.
LM Studio docs describe the app as a local runtime environment that supports:
- llama.cpp on Mac, Windows, and Linux
- MLX on Apple Silicon
The Gemma 4 model pages in LM Studio point to GGUF-based community builds. So in practice, LM Studio Gemma 4 support means an easy local path for supported GGUF builds, not "download Google's raw safetensors and run them unchanged in the UI."
That distinction matters because it changes what you should expect:
- good for local inference
- good for chat and quick evaluation
- good for people who want a GUI
- not the right path if you specifically want raw training workflows
Which Gemma 4 model should you try first in LM Studio?
For most users:
- start with E4B if you want the best small-model default
- start with E2B only if your hardware is very tight
- start with 26B A4B if you have a serious local box and want the best speed-quality tradeoff
- use 31B only if you already know you can afford the memory
In other words, LM Studio Gemma 4 support exists across the lineup, but your hardware should decide what you actually open first.
Is LM Studio a good fit for Gemma 4?
LM Studio is a good fit if you want:
- the easiest GUI-first local experience
- a clean search-and-download workflow
- quick comparisons between small and large Gemma 4 variants
- local inference without building your own stack first
It is especially attractive if you are still in the "Which Gemma 4 model do I even like?" phase.
When LM Studio is not the best path
LM Studio is not the best answer if you want:
- the most configurable command-line local server
- fine-tuning or adapter training
- raw Hugging Face training workflows
- the smallest possible deployment stack for automation
In those cases, llama.cpp or Unsloth may be the better next step.
FAQ
Does LM Studio support Gemma 4 today?
Yes. As of April 7, 2026, LM Studio has public model pages for all four Gemma 4 sizes.
What does LM Studio support for Gemma 4 actually mean?
It means LM Studio can run supported local Gemma 4 builds through its local runtimes, especially GGUF-based paths.
Which Gemma 4 model should I open first in LM Studio?
Usually E4B first, then 26B A4B if your hardware is strong enough.
Can LM Studio run Gemma 4 31B?
Yes, but LM Studio lists 19 GB minimum system memory, so this is not the model to start with unless you know your machine can handle it.
Official references
- Google: Run Gemma with LM Studio
- LM Studio docs
- LM Studio: Gemma 4 model family
- LM Studio: Gemma 4 E2B
- LM Studio: Gemma 4 E4B
- LM Studio: Gemma 4 26B A4B
- LM Studio: Gemma 4 31B
Related guides
Related guides
Continue through the Gemma 4 cluster with the next guide that matches your current decision.

How to Run Gemma 4 in LM Studio
A practical LM Studio guide for Gemma 4, focused on model choice, hardware fit, first-run workflow, and what to check before you blame the model.

Does llama.cpp Support Gemma 4? GGUF Status, Fixes, and What Works
A practical answer to whether llama.cpp supports Gemma 4, with the official GGUF links, current support status, and what 'supported' really means.

Does Unsloth Support Gemma 4? Local Run and Fine-Tuning Status
A practical answer to whether Unsloth supports Gemma 4, covering local run support, fine-tuning support, and the model-specific caveats that matter.
Still deciding what to read next?
Go back to the guide hub to browse model comparisons, setup walkthroughs, and hardware planning pages.
