Kal1510 commited on
Commit
d460de4
·
verified ·
1 Parent(s): 9d83d90

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +3 -5
app.py CHANGED
@@ -33,13 +33,11 @@ warnings.filterwarnings("ignore")
33
  print("Start")
34
  import subprocess
35
 
36
- import subprocess
37
-
38
  try:
39
  subprocess.run([
40
  "huggingface-cli", "download",
41
- "TheBloke/Phi-3-mini-4k-instruct-GGUF",
42
- "phi-3-mini-4k-instruct.Q4_K_M.gguf",
43
  "--local-dir", "./models",
44
  "--local-dir-use-symlinks", "False"
45
  ], check=True)
@@ -65,7 +63,7 @@ embeddings = HuggingFaceEmbeddings(
65
  # Load Mistral GGUF via llama.cpp (CPU optimized)
66
  # ------------------------------
67
  llm_cpp = Llama(
68
- model_path="./models/Phi-3-mini-4k-instruct-gguf",
69
  n_ctx=2048,
70
  n_threads=4, # Adjust based on your CPU cores
71
  n_gpu_layers=0, # Force CPU-only
 
33
  print("Start")
34
  import subprocess
35
 
 
 
36
  try:
37
  subprocess.run([
38
  "huggingface-cli", "download",
39
+ "microsoft/Phi-3-mini-4k-instruct-gguf",
40
+ "Phi-3-mini-4k-instruct-q4.gguf",
41
  "--local-dir", "./models",
42
  "--local-dir-use-symlinks", "False"
43
  ], check=True)
 
63
  # Load Mistral GGUF via llama.cpp (CPU optimized)
64
  # ------------------------------
65
  llm_cpp = Llama(
66
+ model_path="./models/Phi-3-mini-4k-instruct-q4.gguf",
67
  n_ctx=2048,
68
  n_threads=4, # Adjust based on your CPU cores
69
  n_gpu_layers=0, # Force CPU-only