Spaces:
Sleeping
Sleeping
Tuchuanhuhuhu
commited on
Commit
·
25b27e3
1
Parent(s):
dfcf091
为各个模型添加了个性化软上限
Browse files- ChuanhuChatbot.py +1 -1
- modules/chat_func.py +23 -23
- modules/presets.py +27 -2
ChuanhuChatbot.py
CHANGED
@@ -331,7 +331,7 @@ with gr.Blocks(css=customCSS, theme=small_and_beautiful_theme) as demo:
|
|
331 |
token_count,
|
332 |
top_p,
|
333 |
temperature,
|
334 |
-
gr.State(
|
335 |
model_select_dropdown,
|
336 |
language_select_dropdown,
|
337 |
],
|
|
|
331 |
token_count,
|
332 |
top_p,
|
333 |
temperature,
|
334 |
+
gr.State(sum(token_count.value[-4:])),
|
335 |
model_select_dropdown,
|
336 |
language_select_dropdown,
|
337 |
],
|
modules/chat_func.py
CHANGED
@@ -350,29 +350,29 @@ def predict(
|
|
350 |
+ colorama.Style.RESET_ALL
|
351 |
)
|
352 |
|
353 |
-
|
354 |
-
|
355 |
-
|
356 |
-
|
357 |
-
|
358 |
-
|
359 |
-
|
360 |
-
|
361 |
-
|
362 |
-
|
363 |
-
|
364 |
-
|
365 |
-
|
366 |
-
|
367 |
-
|
368 |
-
|
369 |
-
|
370 |
-
|
371 |
-
|
372 |
-
|
373 |
-
|
374 |
-
|
375 |
-
|
376 |
|
377 |
|
378 |
def retry(
|
|
|
350 |
+ colorama.Style.RESET_ALL
|
351 |
)
|
352 |
|
353 |
+
if stream:
|
354 |
+
max_token = MODEL_SOFT_TOKEN_LIMIT[selected_model]["streaming"]
|
355 |
+
else:
|
356 |
+
max_token = MODEL_SOFT_TOKEN_LIMIT[selected_model]["all"]
|
357 |
+
|
358 |
+
if sum(all_token_counts) > max_token and should_check_token_count:
|
359 |
+
status_text = f"精简token中{all_token_counts}/{max_token}"
|
360 |
+
logging.info(status_text)
|
361 |
+
yield chatbot, history, status_text, all_token_counts
|
362 |
+
iter = reduce_token_size(
|
363 |
+
openai_api_key,
|
364 |
+
system_prompt,
|
365 |
+
history,
|
366 |
+
chatbot,
|
367 |
+
all_token_counts,
|
368 |
+
top_p,
|
369 |
+
temperature,
|
370 |
+
max_token//2,
|
371 |
+
selected_model=selected_model,
|
372 |
+
)
|
373 |
+
for chatbot, history, status_text, all_token_counts in iter:
|
374 |
+
status_text = f"Token 达到上限,已自动降低Token计数至 {status_text}"
|
375 |
+
yield chatbot, history, status_text, all_token_counts
|
376 |
|
377 |
|
378 |
def retry(
|
modules/presets.py
CHANGED
@@ -19,9 +19,7 @@ ssl_error_prompt = "SSL错误,无法获取对话。" # SSL 错误
|
|
19 |
no_apikey_msg = "API key长度不是51位,请检查是否输入正确。" # API key 长度不足 51 位
|
20 |
no_input_msg = "请输入对话内容。" # 未输入对话内容
|
21 |
|
22 |
-
max_token_streaming = 3500 # 流式对话时的最大 token 数
|
23 |
timeout_streaming = 10 # 流式对话时的超时时间
|
24 |
-
max_token_all = 3500 # 非流式对话时的最大 token 数
|
25 |
timeout_all = 200 # 非流式对话时的超时时间
|
26 |
enable_streaming_option = True # 是否启用选择选择是否实时显示回答的勾选框
|
27 |
HIDE_MY_KEY = False # 如果你想在UI中隐藏你的 API 密钥,将此值设置为 True
|
@@ -57,6 +55,33 @@ MODELS = [
|
|
57 |
"gpt-4-32k-0314",
|
58 |
] # 可选的模型
|
59 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
60 |
REPLY_LANGUAGES = [
|
61 |
"中文",
|
62 |
"English",
|
|
|
19 |
no_apikey_msg = "API key长度不是51位,请检查是否输入正确。" # API key 长度不足 51 位
|
20 |
no_input_msg = "请输入对话内容。" # 未输入对话内容
|
21 |
|
|
|
22 |
timeout_streaming = 10 # 流式对话时的超时时间
|
|
|
23 |
timeout_all = 200 # 非流式对话时的超时时间
|
24 |
enable_streaming_option = True # 是否启用选择选择是否实时显示回答的勾选框
|
25 |
HIDE_MY_KEY = False # 如果你想在UI中隐藏你的 API 密钥,将此值设置为 True
|
|
|
55 |
"gpt-4-32k-0314",
|
56 |
] # 可选的模型
|
57 |
|
58 |
+
MODEL_SOFT_TOKEN_LIMIT = {
|
59 |
+
"gpt-3.5-turbo": {
|
60 |
+
"streaming": 3500,
|
61 |
+
"all": 3500
|
62 |
+
},
|
63 |
+
"gpt-3.5-turbo-0301": {
|
64 |
+
"streaming": 3500,
|
65 |
+
"all": 3500
|
66 |
+
},
|
67 |
+
"gpt-4": {
|
68 |
+
"streaming": 7500,
|
69 |
+
"all": 7500
|
70 |
+
},
|
71 |
+
"gpt-4-0314": {
|
72 |
+
"streaming": 7500,
|
73 |
+
"all": 7500
|
74 |
+
},
|
75 |
+
"gpt-4-32k": {
|
76 |
+
"streaming": 31000,
|
77 |
+
"all": 31000
|
78 |
+
},
|
79 |
+
"gpt-4-32k-0314": {
|
80 |
+
"streaming": 31000,
|
81 |
+
"all": 31000
|
82 |
+
}
|
83 |
+
}
|
84 |
+
|
85 |
REPLY_LANGUAGES = [
|
86 |
"中文",
|
87 |
"English",
|