Model Summary

This repository hosts quantized versions of the bge-reranker-v2-m3 embedding model.

Format: GGUF
Converter: llama.cpp 8733e0cf6eefc7c7752297cc22d0836706f4222c
Quantizer: LM-Kit.NET 2025.5.2

For more detailed information on the base model, please visit the following links

Downloads last month
43
GGUF
Model size
568M params
Architecture
bert
Hardware compatibility
Log In to view the estimation

4-bit

8-bit

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support