Spaces:
Sleeping
Sleeping
remove logits
Browse files- rag_app/rag_2.py +2 -2
rag_app/rag_2.py
CHANGED
@@ -38,12 +38,12 @@ llm = LlamaCPP(
|
|
38 |
temperature=0.1,
|
39 |
max_new_tokens=256,
|
40 |
context_window=16384,
|
41 |
-
model_kwargs={"n_gpu_layers":-1, 'logits_all':
|
42 |
messages_to_prompt=messages_to_prompt,
|
43 |
completion_to_prompt=completion_to_prompt,)
|
44 |
|
45 |
llm2 = Llama(model_path="models/Llama-3.2-1B-Instruct-Q4_K_M.gguf",
|
46 |
-
n_gpu_layers=-1, n_ctx=8000
|
47 |
|
48 |
|
49 |
embedding_model = HuggingFaceEmbedding(
|
|
|
38 |
temperature=0.1,
|
39 |
max_new_tokens=256,
|
40 |
context_window=16384,
|
41 |
+
model_kwargs={"n_gpu_layers":-1, 'logits_all': False},
|
42 |
messages_to_prompt=messages_to_prompt,
|
43 |
completion_to_prompt=completion_to_prompt,)
|
44 |
|
45 |
llm2 = Llama(model_path="models/Llama-3.2-1B-Instruct-Q4_K_M.gguf",
|
46 |
+
n_gpu_layers=-1, n_ctx=8000)
|
47 |
|
48 |
|
49 |
embedding_model = HuggingFaceEmbedding(
|