Update app.py
Browse files
app.py
CHANGED
@@ -35,8 +35,8 @@ import subprocess
|
|
35 |
|
36 |
subprocess.run([
|
37 |
"huggingface-cli", "download",
|
38 |
-
"
|
39 |
-
"
|
40 |
"--local-dir", "./models",
|
41 |
"--local-dir-use-symlinks", "False"
|
42 |
], check=True)
|
@@ -58,7 +58,7 @@ embeddings = HuggingFaceEmbeddings(
|
|
58 |
# Load Mistral GGUF via llama.cpp (CPU optimized)
|
59 |
# ------------------------------
|
60 |
llm_cpp = Llama(
|
61 |
-
model_path="./models/
|
62 |
n_ctx=2048,
|
63 |
n_threads=4, # Adjust based on your CPU cores
|
64 |
n_gpu_layers=0, # Force CPU-only
|
|
|
35 |
|
36 |
subprocess.run([
|
37 |
"huggingface-cli", "download",
|
38 |
+
"microsoft/Phi-3-mini-4k-instruct-gguf",
|
39 |
+
"Phi-3-mini-4k-instruct-gguf",
|
40 |
"--local-dir", "./models",
|
41 |
"--local-dir-use-symlinks", "False"
|
42 |
], check=True)
|
|
|
58 |
# Load Mistral GGUF via llama.cpp (CPU optimized)
|
59 |
# ------------------------------
|
60 |
llm_cpp = Llama(
|
61 |
+
model_path="./models/Phi-3-mini-4k-instruct-gguf",
|
62 |
n_ctx=2048,
|
63 |
n_threads=4, # Adjust based on your CPU cores
|
64 |
n_gpu_layers=0, # Force CPU-only
|