Spaces:
Sleeping
Sleeping
Donald Winkelman
commited on
Commit
·
3264d58
1
Parent(s):
0f1df4e
Updating Side-By-Side Space
Browse files
app.py
CHANGED
@@ -142,8 +142,11 @@ def load_models(progress=None):
|
|
142 |
print(f"Loading base model from {BASE_MODEL_PATH}")
|
143 |
base_model = Llama(
|
144 |
model_path=BASE_MODEL_PATH,
|
145 |
-
n_ctx=
|
146 |
-
n_threads=4 # Number of CPU threads to use
|
|
|
|
|
|
|
147 |
)
|
148 |
|
149 |
# Load novel model
|
@@ -153,8 +156,11 @@ def load_models(progress=None):
|
|
153 |
print(f"Loading novel model from {NOVEL_MODEL_PATH}")
|
154 |
novel_model = Llama(
|
155 |
model_path=NOVEL_MODEL_PATH,
|
156 |
-
n_ctx=
|
157 |
-
n_threads=4 # Number of CPU threads to use
|
|
|
|
|
|
|
158 |
)
|
159 |
|
160 |
if progress is not None:
|
|
|
142 |
print(f"Loading base model from {BASE_MODEL_PATH}")
|
143 |
base_model = Llama(
|
144 |
model_path=BASE_MODEL_PATH,
|
145 |
+
n_ctx=8092, # Context window size
|
146 |
+
n_threads=4, # Number of CPU threads to use
|
147 |
+
temperature=0.3, # Temperature for sampling
|
148 |
+
top_p=0.9, # Top-p sampling
|
149 |
+
top_k=20, # Top-k sampling
|
150 |
)
|
151 |
|
152 |
# Load novel model
|
|
|
156 |
print(f"Loading novel model from {NOVEL_MODEL_PATH}")
|
157 |
novel_model = Llama(
|
158 |
model_path=NOVEL_MODEL_PATH,
|
159 |
+
n_ctx=8092, # Context window size
|
160 |
+
n_threads=4, # Number of CPU threads to use
|
161 |
+
temperature=0.3, # Temperature for sampling
|
162 |
+
top_p=0.9, # Top-p sampling
|
163 |
+
top_k=20, # Top-k sampling
|
164 |
)
|
165 |
|
166 |
if progress is not None:
|