hassenhamdi commited on
Commit
73d505f
·
verified ·
1 Parent(s): 8b09e49

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +4 -7
README.md CHANGED
@@ -19,9 +19,6 @@ tags:
19
  This model was converted to GGUF format from [`suayptalha/DeepSeek-R1-Distill-Llama-3B`](https://huggingface.co/suayptalha/DeepSeek-R1-Distill-Llama-3B) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
20
  Refer to the [original model card](https://huggingface.co/suayptalha/DeepSeek-R1-Distill-Llama-3B) for more details on the model.
21
 
22
- This a Q4_K_M quantization.
23
- You can use the model in lmstudio or any other text generatio ui of choice able of running gguf format.
24
-
25
  ## Use with llama.cpp
26
  Install llama.cpp through brew (works on Mac and Linux)
27
 
@@ -33,12 +30,12 @@ Invoke the llama.cpp server or the CLI.
33
 
34
  ### CLI:
35
  ```bash
36
- llama-cli --hf-repo hassenhamdi/DeepSeek-R1-Distill-Llama-3B-GGUF --hf-file deepseek-r1-distill-llama-3b-q4_k_m.gguf -p "The meaning to life and the universe is"
37
  ```
38
 
39
  ### Server:
40
  ```bash
41
- llama-server --hf-repo hassenhamdi/DeepSeek-R1-Distill-Llama-3B-GGUF --hf-file deepseek-r1-distill-llama-3b-q4_k_m.gguf -c 2048
42
  ```
43
 
44
  Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well.
@@ -55,9 +52,9 @@ cd llama.cpp && LLAMA_CURL=1 make
55
 
56
  Step 3: Run inference through the main binary.
57
  ```
58
- ./llama-cli --hf-repo hassenhamdi/DeepSeek-R1-Distill-Llama-3B-GGUF --hf-file deepseek-r1-distill-llama-3b-q4_k_m.gguf -p "The meaning to life and the universe is"
59
  ```
60
  or
61
  ```
62
- ./llama-server --hf-repo hassenhamdi/DeepSeek-R1-Distill-Llama-3B-GGUF --hf-file deepseek-r1-distill-llama-3b-q4_k_m.gguf -c 2048
63
  ```
 
19
  This model was converted to GGUF format from [`suayptalha/DeepSeek-R1-Distill-Llama-3B`](https://huggingface.co/suayptalha/DeepSeek-R1-Distill-Llama-3B) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
20
  Refer to the [original model card](https://huggingface.co/suayptalha/DeepSeek-R1-Distill-Llama-3B) for more details on the model.
21
 
 
 
 
22
  ## Use with llama.cpp
23
  Install llama.cpp through brew (works on Mac and Linux)
24
 
 
30
 
31
  ### CLI:
32
  ```bash
33
+ llama-cli --hf-repo hassenhamdi/DeepSeek-R1-Distill-Llama-3B-GGUF --hf-file deepseek-r1-distill-llama-3b-q8_0.gguf -p "The meaning to life and the universe is"
34
  ```
35
 
36
  ### Server:
37
  ```bash
38
+ llama-server --hf-repo hassenhamdi/DeepSeek-R1-Distill-Llama-3B-GGUF --hf-file deepseek-r1-distill-llama-3b-q8_0.gguf -c 2048
39
  ```
40
 
41
  Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well.
 
52
 
53
  Step 3: Run inference through the main binary.
54
  ```
55
+ ./llama-cli --hf-repo hassenhamdi/DeepSeek-R1-Distill-Llama-3B-GGUF --hf-file deepseek-r1-distill-llama-3b-q8_0.gguf -p "The meaning to life and the universe is"
56
  ```
57
  or
58
  ```
59
+ ./llama-server --hf-repo hassenhamdi/DeepSeek-R1-Distill-Llama-3B-GGUF --hf-file deepseek-r1-distill-llama-3b-q8_0.gguf -c 2048
60
  ```