Spaces:
Sleeping
Sleeping
remove logits
Browse files- rag_app/rag_2.py +2 -2
rag_app/rag_2.py
CHANGED
|
@@ -38,12 +38,12 @@ llm = LlamaCPP(
|
|
| 38 |
temperature=0.1,
|
| 39 |
max_new_tokens=256,
|
| 40 |
context_window=16384,
|
| 41 |
-
model_kwargs={"n_gpu_layers":-1, 'logits_all':
|
| 42 |
messages_to_prompt=messages_to_prompt,
|
| 43 |
completion_to_prompt=completion_to_prompt,)
|
| 44 |
|
| 45 |
llm2 = Llama(model_path="models/Llama-3.2-1B-Instruct-Q4_K_M.gguf",
|
| 46 |
-
n_gpu_layers=-1, n_ctx=8000
|
| 47 |
|
| 48 |
|
| 49 |
embedding_model = HuggingFaceEmbedding(
|
|
|
|
| 38 |
temperature=0.1,
|
| 39 |
max_new_tokens=256,
|
| 40 |
context_window=16384,
|
| 41 |
+
model_kwargs={"n_gpu_layers":-1, 'logits_all': False},
|
| 42 |
messages_to_prompt=messages_to_prompt,
|
| 43 |
completion_to_prompt=completion_to_prompt,)
|
| 44 |
|
| 45 |
llm2 = Llama(model_path="models/Llama-3.2-1B-Instruct-Q4_K_M.gguf",
|
| 46 |
+
n_gpu_layers=-1, n_ctx=8000)
|
| 47 |
|
| 48 |
|
| 49 |
embedding_model = HuggingFaceEmbedding(
|