We Have Gemini At Home

#1
by MarinaraSpaghetti - opened

a1iuy4.jpg

All jokes aside, this model was blatantly trained on Gemini's outputs. It reads the same, makes the same mistakes, and has the same writing style. I had Gemini Flash set as a fallback model on OpenRouter, and I couldn't tell the difference between the reply from GLM and the previously mentioned one.

If you love locally run Gemini, this model is for you. Otherwise, don't bother, and go for the actual Gemini, since that one is smarter and has a better context (this one is barely usable on 64k). Hybrid thinking is never a good idea, as we've seen in Qwen3's example. Keep in mind, I tested the model in role-playing/creative writing scenarios. It might do better at coding.

To devs, don't mind my harsh review, I have very high expectations. The gooner crowd is very tough to please. Keep up the good work and cheers.

Lol, tru. GLM-4.5 performs better at coding and agentic capabilixties. But with such a cost of API, what can I say? Not bad.

tru,lol

Trying to use this for RP/writing on OpenRouter. It's available on NanoGPT, but seems to still be set to placeholder pricing (ie. $200.00 input rate, lol).

So I do like how it's very cheap pricing. Even at filled 24k ctx it only uses $0.015 per inference. Gemini-2.5 is free tho, and free beats cheap, so I still give the nod to Gemini-2.5-Pro.

I noticed Chat Completion caused group chat chars to speak as each other in other char messages. Text Completion seems to fix this.

It's also slower than Gemini-2.5-Pro, and I know it's unfair to compare GLM-4.5 via OpenRouter to Google's main servers, but it is what it is. It's also possible OpenRouter is getting slammed as I randomly get "too many request" error messages, so hopefully the situation improves.

I'm going to hold off reviewing the writing until I see more as OpenRouter is just way too slow or unresponsive with it atm.

Sign up or log in to comment