LMStudio: Failed to load model error loading model: error loading model architecture: unknown model architecture: 'dots1'

#1
by ljupco - opened

Is anyone getting an error in LMStudio:

Failed to load model error loading model: error loading model architecture: unknown model architecture: 'dots1'

? This is trying to load model quants dots.llm1.inst-UD-IQ2_M.

I assume I need an engine and or LMStudio merge of support for it. And I should just wait and re-try from time to time whether support was merged. Posting here jic it's something else.

My versions on MacOS are

LMStuidio: You are all up to date! The current version is 0.3.16
llama.cpp: Metal llama.cpp v1.33.0
MLX: LM Studio MLX v0.17.0

LMStudio is not working for me either. I don't even see an error log. I just click load and nothing happens. I assume LMStudio needs to be updated before we'll be able to run dots. I'm on windows 11.

Unsloth AI org
โ€ข
edited 10 days ago

It works in llama.cpp so yes, you will need to update LM Studio

Thereโ€™s no update to support this yet in LM studio.

It's working now after updating llama.cpp engine

Thanks! Updated LMStudio now:

You are all up to date! The current version is 0.3.17

But still, the llama.cpp runtime bundled with todays LMStudio on my MacOS is from a month ago:

Metal llama.cpp
v1.33.0

Engine
Apple Metal accelerated llama.cpp engine
Latest Version Installed

  • Llama 4 vision support
  • Enable any LLM to be used as an embedding model (requires LM Studio 0.3.16-b6)
  • Fixed prompt processing bugs when chats exceed context length
  • llama.cpp updated to b5459 (commit 8a1d206)

I see the 8a1d206 commit in llama.cpp be from 2025-05-22 :

8a1d206f [email protected] 2025-05-22 22:21:07 +0300 : tts : fix n_ubatch + make WavTokenizer cache-less (#13713)

I see latter commit from 2025-06-15 that looks like important to have:

9ae4143b [email protected] 2025-06-15 00:52:06 -0700 : model : add dots.llm1 architecture support (#14044) (#14118)

I guess will just wait some more, from LMStudio to move to news llama.cpp. Atm on trying to load dots.llm1 I'm still getting:

error loading model: error loading model architecture: unknown model architecture: 'dots1'

and that kind of makes sense. I can confirm that llama.cpp built from source today works (mbp m2 96gb ram) - this:

sudo sysctl iogpu.wired_limit_mb=80000
build/bin/llama-server --model models/dots.llm1.inst-UD-TQ1_0.gguf --temp 0 --top_p 0.95 --min_p 0 --ctx-size 4096 &

(Edit) Can use larger context with flash attention on - this takes 75gb (v)ram and delivers 16 tps on M2:

build/bin/llama-server --model models/dots.llm1.inst-UD-TQ1_0.gguf --temp 0 --top_p 0.95 --min_p 0 --ctx-size 32758 --flash-attn --cache-type-k q8_0 --cache-type-v q8_0 &

LMStudio updated just now (for Gemma), and got the update for dots.llm1 in too.

Metal llama.cpp
v1.37.1
Apple Metal accelerated llama.cpp engine

  • Google Gemma3n support (text-only)
  • dots.llm1 model support
  • NeoBERT embedding model support
  • llama.cpp updated to b5757 (commit bf5bcd0)

Nice! - all should works now. :-)

ljupco changed discussion status to closed

Sign up or log in to comment