This is just my IQ4_XS quant of https://huggingface.co/failspy/Meta-Llama-3-70B-Instruct-abliterated-v3.5
Also included is my generated imatrix data file, generated on 200 x 2048 tokens of Wiki.Train.Raw against Q8_0 (i dont have enough RAM + VRAM to do the full fp16).
- Downloads last month
- 12
Hardware compatibility
Log In
to view the estimation
4-bit
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
๐
Ask for provider support