Spaces:
Running
Running
Update app.py
Browse files
app.py
CHANGED
@@ -33,28 +33,8 @@ filename_list = [
|
|
33 |
"Wizard-Vicuna-7B-Uncensored.ggmlv3.q8_0.bin",
|
34 |
]
|
35 |
|
36 |
-
URL = "https://huggingface.co/TheBloke/Wizard-Vicuna-7B-Uncensored-GGML/raw/main/Wizard-Vicuna-7B-Uncensored.ggmlv3.q4_K_M.bin" # 4.05G
|
37 |
-
|
38 |
-
url = "https://huggingface.co/savvamadar/ggml-gpt4all-j-v1.3-groovy/blob/main/ggml-gpt4all-j-v1.3-groovy.bin"
|
39 |
-
url = "https://huggingface.co/TheBloke/Llama-2-13B-GGML/blob/main/llama-2-13b.ggmlv3.q4_K_S.bin" # 7.37G
|
40 |
-
# url = "https://huggingface.co/TheBloke/Llama-2-13B-chat-GGML/blob/main/llama-2-13b-chat.ggmlv3.q3_K_L.bin"
|
41 |
-
url = "https://huggingface.co/TheBloke/Llama-2-13B-chat-GGML/blob/main/llama-2-13b-chat.ggmlv3.q3_K_L.bin" # 6.93G
|
42 |
-
# url = "https://huggingface.co/TheBloke/Llama-2-13B-chat-GGML/blob/main/llama-2-13b-chat.ggmlv3.q3_K_L.binhttps://huggingface.co/TheBloke/Llama-2-13B-chat-GGML/blob/main/llama-2-13b-chat.ggmlv3.q4_K_M.bin" # 7.87G
|
43 |
-
|
44 |
-
url = "https://huggingface.co/localmodels/Llama-2-13B-Chat-ggml/blob/main/llama-2-13b-chat.ggmlv3.q4_K_S.bin" # 7.37G
|
45 |
-
|
46 |
-
_ = (
|
47 |
-
"golay" in platform.node()
|
48 |
-
or "okteto" in platform.node()
|
49 |
-
or Path("/kaggle").exists()
|
50 |
-
# or psutil.cpu_count(logical=False) < 4
|
51 |
-
or 1 # run 7b in hf
|
52 |
-
)
|
53 |
|
54 |
-
|
55 |
-
# url = "https://huggingface.co/TheBloke/Llama-2-13B-chat-GGML/blob/main/llama-2-13b-chat.ggmlv3.q2_K.bin"
|
56 |
-
url = "https://huggingface.co/TheBloke/Llama-2-7B-Chat-GGML/blob/main/llama-2-7b-chat.ggmlv3.q2_K.bin" # 2.87G
|
57 |
-
url = "https://huggingface.co/TheBloke/Llama-2-7B-Chat-GGML/blob/main/llama-2-7b-chat.ggmlv3.q4_K_M.bin" # 2.87G
|
58 |
|
59 |
prompt_template = """Below is an instruction that describes a task. Write a response that appropriately completes the request.
|
60 |
|
@@ -402,17 +382,6 @@ with gr.Blocks(
|
|
402 |
)
|
403 |
clear.click(lambda: None, None, chatbot, queue=False)
|
404 |
|
405 |
-
with gr.Accordion("For Chat/Translation API", open=False, visible=False):
|
406 |
-
input_text = gr.Text()
|
407 |
-
api_btn = gr.Button("Go", variant="primary")
|
408 |
-
out_text = gr.Text()
|
409 |
-
|
410 |
-
api_btn.click(
|
411 |
-
predict_api,
|
412 |
-
input_text,
|
413 |
-
out_text,
|
414 |
-
api_name="api",
|
415 |
-
)
|
416 |
|
417 |
# block.load(update_buff, [], buff, every=1)
|
418 |
# block.load(update_buff, [buff_var], [buff_var, buff], every=1)
|
|
|
33 |
"Wizard-Vicuna-7B-Uncensored.ggmlv3.q8_0.bin",
|
34 |
]
|
35 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
36 |
|
37 |
+
url = "https://huggingface.co/TheBloke/Llama-2-7B-Chat-GGML/blob/main/llama-2-7b-chat.ggmlv3.q4_K_M.bin" # 2.87G
|
|
|
|
|
|
|
38 |
|
39 |
prompt_template = """Below is an instruction that describes a task. Write a response that appropriately completes the request.
|
40 |
|
|
|
382 |
)
|
383 |
clear.click(lambda: None, None, chatbot, queue=False)
|
384 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
385 |
|
386 |
# block.load(update_buff, [], buff, every=1)
|
387 |
# block.load(update_buff, [buff_var], [buff_var, buff], every=1)
|