This is the 8-bit quantized version of Alibaba-NLP/gte-Qwen2-1.5B-instruct by following the example from the AutoGPTQ repository.
- Downloads last month
- 93
Inference Providers
NEW
This model is not currently available via any of the supported third-party Inference Providers, and
the HF Inference API does not support transformers models with pipeline type sentence-similarity
Model tree for ktoprakucar/gte-Qwen2-1.5B-instruct-Q8-GPTQ
Base model
Alibaba-NLP/gte-Qwen2-1.5B-instruct