Update README.md
Browse files
README.md
CHANGED
@@ -12,3 +12,12 @@ base_model:
|
|
12 |
|
13 |
This model was converted to GGUF format from [`manu/bge-m3-custom-fr`](https://huggingface.co/manu/bge-m3-custom-fr) using llama.cpp.
|
14 |
Refer to the [original model card](https://huggingface.co/manu/bge-m3-custom-fr) for more details on the model.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
12 |
|
13 |
This model was converted to GGUF format from [`manu/bge-m3-custom-fr`](https://huggingface.co/manu/bge-m3-custom-fr) using llama.cpp.
|
14 |
Refer to the [original model card](https://huggingface.co/manu/bge-m3-custom-fr) for more details on the model.
|
15 |
+
|
16 |
+
|
17 |
+
You can run the model as an embedding model using llama-server.
|
18 |
+
|
19 |
+
For installation, you can follow the instructions from the [repository](https://github.com/ggml-org/llama.cpp/blob/master/examples/server/README.md#build) !
|
20 |
+
|
21 |
+
```
|
22 |
+
./build/bin/llama-server -m bge-m3-custom-fr_q8_0.gguf --embedding --pooling mean -ub 8192 --port 8001 --batch-size 4096
|
23 |
+
```
|