Image
Promise I will never go blonde like Kanye

Overview

Didn't really have any cool README ideas for this so we're just going with just whatever song i'm listening to rn and it happened to be Baby i'm bleeding

Nevertheless, This is a finetune from the 32K context extended (or fixed?) Arcee GLM4 base - Trained shrimply with just the Tulu-SFT-Mixture but I removed Safety alignment examples. Came out pretty well, It uses chatML due to the GLM4 Format giving me a headache. It's a decently competant assistant although I haven't done any testing on how well the model performs at longer-contexts, nor have i done any RL afterwards to fix up it's edges.

Think it should be a decent base for any future finetunes, I felt that GLM4 really wasn't given the proper time of day and it's a way better base then any Qwen3 model.

Quants

GGUF: https://huggingface.co/mradermacher/GLM-Tulu-ChatML-GGUF

Imatrix GGUF: https://huggingface.co/mradermacher/GLM-Tulu-ChatML-i1-GGUF

Prompting

The model was trained with ChatML formatting

"""<|im_start|>system
system prompt<|im_end|>
<|im_start|>user
Hi there!<|im_end|>
<|im_start|>assistant
Nice to meet you!<|im_end|>
<|im_start|>user
Can I ask a question?<|im_end|>
<|im_start|>assistant
"""

Configs

WandB : https://wandb.ai/new-eden/Training-A100/runs/05kktve8?nw=nwuserdeltavector

This train took 15 hours on 8xB200s provided by Deepinfra and Cognitive Computations, Config is linked in the WandB

Credits

Thank you to Lucy, Auri, NyxKrage, Creators of the Tulu-SFT-Mix and everyone at Anthracite & Allura

Downloads last month
1,805
Safetensors
Model size
32.6B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for Delta-Vector/GLM-4-32B-Tulu-Instruct

Finetuned
(2)
this model
Adapters
2 models
Quantizations
6 models

Dataset used to train Delta-Vector/GLM-4-32B-Tulu-Instruct