lordjimen commited on
Commit
54d9d10
·
verified ·
1 Parent(s): 916e3e3

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +67 -0
README.md ADDED
@@ -0,0 +1,67 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: gemma
3
+ language:
4
+ - uk
5
+ - en
6
+ base_model:
7
+ - google/gemma-2-9b-it
8
+ - google/gemma-2-9b
9
+ ---
10
+ # MamayLM-Gemma-2-9B-IT-v0.1-GGUF
11
+
12
+ MamayLM is distributed under [Gemma Terms of Use](https://ai.google.dev/gemma/terms).
13
+
14
+ This repo contains the GGUF format model files for [INSAIT-Institute/MamayLM-Gemma-2-9B-IT-v0.1](https://huggingface.co/INSAIT-Institute/MamayLM-Gemma-2-9B-IT-v0.1).
15
+
16
+ ## Quick Start using Python
17
+
18
+ Install the required package:
19
+ ```bash
20
+ pip install llama-cpp-python
21
+ ```
22
+
23
+ Example chat completion:
24
+
25
+ ```python
26
+ from llama_cpp import Llama
27
+
28
+ llm = Llama(
29
+ model_path="path/to/your/model.gguf",
30
+ n_ctx=8192,
31
+ penalize_nl=False
32
+ )
33
+
34
+ messages = [{"role": "user", "content": "Хто такий Козак Мамай??"}]
35
+ response = llm.create_chat_completion(
36
+ messages=messages,
37
+ max_tokens=2048, # Choose maximum generated tokens
38
+ temperature=0.1,
39
+ top_p=0.9,
40
+ repeat_penalty=1.0,
41
+ stop=["<eos>", "<end_of_turn>"]
42
+ )
43
+
44
+
45
+ ```
46
+
47
+ Example normal completion:
48
+
49
+ ```python
50
+ from llama_cpp import Llama
51
+
52
+ llm = Llama(
53
+ model_path="path/to/your/model.gguf",
54
+ n_ctx=8192,
55
+ penalize_nl=False
56
+ )
57
+
58
+ prompt = "<start_of_turn>user\nХто такий Козак Мамай?<end_of_turn>\n<start_of_turn>model\n"
59
+ response = llm(
60
+ prompt,
61
+ max_tokens=2048, # Choose maximum generated tokens
62
+ temperature=0.1,
63
+ top_p=0.9,
64
+ repeat_penalty=1.0,
65
+ stop=["<eos>","<end_of_turn>"]
66
+ )
67
+ ```