Input type (torch.cuda.FloatTensor) and weight type (torch.cuda.HalfTensor) should be the same

#3
by treehugg3 - opened

I get this error when I load the model using the latest transformers: Input type (torch.cuda.FloatTensor) and weight type (torch.cuda.HalfTensor) should be the same

I am just loading it using

model = LlavaForConditionalGeneration.from_pretrained(vlm_path,
    use_safetensors=True)

I get this error on this quantized version but not the full pixtral-12b repo. I did however copy the tokenizer_config.json from that repo into this one in order to use the tokenizer.

edit: It fails even with the example in #1.
edit2: I got it to work on this ANCIENT transformers fork branch mentioned here: https://huggingface.co/DewEfresh/pixtral-12b-8bit/discussions/1#66f1a38916c5478fa68c05d6

Unfortunately it is broken on the modern transformers package :(

Sign up or log in to comment