Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -10,11 +10,7 @@ model_name = "meta-llama/Meta-Llama-3-8B-Instruct"
|
|
10 |
hf_token = os.environ.get("HF_TOKEN")
|
11 |
|
12 |
tokenizer = AutoTokenizer.from_pretrained(model_name, use_auth_token=hf_token)
|
13 |
-
model = AutoModelForCausalLM.from_pretrained(
|
14 |
-
model_name,
|
15 |
-
quantization_config=bnb_config,
|
16 |
-
use_auth_token=hf_token
|
17 |
-
)
|
18 |
model.to("cpu") # move model to CPU (if not using GPU Space)
|
19 |
|
20 |
# Prompt generator
|
|
|
10 |
hf_token = os.environ.get("HF_TOKEN")
|
11 |
|
12 |
tokenizer = AutoTokenizer.from_pretrained(model_name, use_auth_token=hf_token)
|
13 |
+
model = AutoModelForCausalLM.from_pretrained(model_name, use_auth_token=hf_token)
|
|
|
|
|
|
|
|
|
14 |
model.to("cpu") # move model to CPU (if not using GPU Space)
|
15 |
|
16 |
# Prompt generator
|