Donald Winkelman commited on
Commit
3264d58
·
1 Parent(s): 0f1df4e

Updating Side-By-Side Space

Browse files
Files changed (1) hide show
  1. app.py +10 -4
app.py CHANGED
@@ -142,8 +142,11 @@ def load_models(progress=None):
142
  print(f"Loading base model from {BASE_MODEL_PATH}")
143
  base_model = Llama(
144
  model_path=BASE_MODEL_PATH,
145
- n_ctx=2048, # Context window size
146
- n_threads=4 # Number of CPU threads to use
 
 
 
147
  )
148
 
149
  # Load novel model
@@ -153,8 +156,11 @@ def load_models(progress=None):
153
  print(f"Loading novel model from {NOVEL_MODEL_PATH}")
154
  novel_model = Llama(
155
  model_path=NOVEL_MODEL_PATH,
156
- n_ctx=2048, # Context window size
157
- n_threads=4 # Number of CPU threads to use
 
 
 
158
  )
159
 
160
  if progress is not None:
 
142
  print(f"Loading base model from {BASE_MODEL_PATH}")
143
  base_model = Llama(
144
  model_path=BASE_MODEL_PATH,
145
+ n_ctx=8092, # Context window size
146
+ n_threads=4, # Number of CPU threads to use
147
+ temperature=0.3, # Temperature for sampling
148
+ top_p=0.9, # Top-p sampling
149
+ top_k=20, # Top-k sampling
150
  )
151
 
152
  # Load novel model
 
156
  print(f"Loading novel model from {NOVEL_MODEL_PATH}")
157
  novel_model = Llama(
158
  model_path=NOVEL_MODEL_PATH,
159
+ n_ctx=8092, # Context window size
160
+ n_threads=4, # Number of CPU threads to use
161
+ temperature=0.3, # Temperature for sampling
162
+ top_p=0.9, # Top-p sampling
163
+ top_k=20, # Top-k sampling
164
  )
165
 
166
  if progress is not None: