zhangjf commited on
Commit
be256e8
·
1 Parent(s): 262234a

limit context length to 1500

Browse files
Files changed (1) hide show
  1. app.py +4 -2
app.py CHANGED
@@ -25,10 +25,12 @@ def ask(question, history, behavior):
25
  for i,content in enumerate(history + [question])
26
  ]
27
  length_messages = num_tokens_from_messages(messages)
 
28
  time_penalty = (length_messages-1000)//10
29
  if time_penalty>0:
30
  print(f"sleep for {time_penalty:.2f}s for too long a quest: {length_messages}")
31
  time.sleep(time_penalty)
 
32
  response = openai.ChatCompletion.create(
33
  model="gpt-3.5-turbo",
34
  messages=forget_long_term(messages)
@@ -62,7 +64,7 @@ def num_tokens_from_messages(messages, model="gpt-3.5-turbo"):
62
  raise NotImplementedError(f"""num_tokens_from_messages() is not presently implemented for model {model}.
63
  See https://github.com/openai/openai-python/blob/main/chatml.md for information on how messages are converted to tokens.""")
64
 
65
- def forget_long_term(messages, max_num_tokens=4000):
66
  while num_tokens_from_messages(messages)>max_num_tokens:
67
  if messages[0]["role"]=="system" and not len(messages[0]["content"]>=max_num_tokens):
68
  messages = messages[:1] + messages[2:]
@@ -194,5 +196,5 @@ with gr.Blocks() as demo:
194
  button_rtr.click(fn=retry, inputs=[txt, state, behavior], outputs=[txt, state, chatbot, downloadfile])
195
  button_clr.click(fn=lambda :([],[]), inputs=None, outputs=[chatbot, state])
196
 
197
- demo.queue(concurrency_count=3, max_size=10)
198
  demo.launch()
 
25
  for i,content in enumerate(history + [question])
26
  ]
27
  length_messages = num_tokens_from_messages(messages)
28
+ """
29
  time_penalty = (length_messages-1000)//10
30
  if time_penalty>0:
31
  print(f"sleep for {time_penalty:.2f}s for too long a quest: {length_messages}")
32
  time.sleep(time_penalty)
33
+ """
34
  response = openai.ChatCompletion.create(
35
  model="gpt-3.5-turbo",
36
  messages=forget_long_term(messages)
 
64
  raise NotImplementedError(f"""num_tokens_from_messages() is not presently implemented for model {model}.
65
  See https://github.com/openai/openai-python/blob/main/chatml.md for information on how messages are converted to tokens.""")
66
 
67
+ def forget_long_term(messages, max_num_tokens=1500):
68
  while num_tokens_from_messages(messages)>max_num_tokens:
69
  if messages[0]["role"]=="system" and not len(messages[0]["content"]>=max_num_tokens):
70
  messages = messages[:1] + messages[2:]
 
196
  button_rtr.click(fn=retry, inputs=[txt, state, behavior], outputs=[txt, state, chatbot, downloadfile])
197
  button_clr.click(fn=lambda :([],[]), inputs=None, outputs=[chatbot, state])
198
 
199
+ #demo.queue(concurrency_count=3, max_size=10)
200
  demo.launch()