Broken model?

#4
by th1nhhdk - opened

I tried using lm studio but the output is "<pad>" repeated over and over

MLX Community org
edited 17 days ago

Here's what I see...
image.png

yes, something is broken
model needs an image in context window to work properly

Screenshot 2025-03-13 at 23.18.33.png

Here's what I see...
image.png

ah, I misremembered, it's actually<pad>

I am using lm studio as well and the output is "<pad>" repeated again and again like the previous images. The max context size is also like 4k tokens instead of 128k I think.

MLX Community org

Taking a look.

@pcuenq
Could you also confirm which base model was used: google/gemma-3-12b-pt or google/gemma-3-12b-it?
I am asking because the model card lists google/gemma-3-12b-it as the base model, but the link in the right menu points to google/gemma-3-12b-pt.

MLX Community org

Can confirm this report on gemma-3-27b-it-4bit

MLX Community org
edited 16 days ago

For me too, it is indeed mlx-community/gemma-3-12b-it-4bit
Also, I can confirm that adding an image to context does make it work, here's a screenshot of the 12B model describing it... impressive.

image.png

Can confirm that gemma-3-4b-it-4bit have the same problem

Same issue here for gemma-3-12b-it-4-bit

I can confirm that the model now works properly with this commit.

th1nhhdk changed discussion status to closed
Your need to confirm your account before you can post a new comment.

Sign up or log in to comment