Spaces:
Running
Running
Update app.py
Browse files
app.py
CHANGED
@@ -175,16 +175,13 @@ def analyze_space(url: str, progress=gr.Progress()):
|
|
175 |
print(traceback.format_exc())
|
176 |
return f"μ€λ₯κ° λ°μνμ΅λλ€: {str(e)}", "", None, "", "", "", "", 10
|
177 |
|
178 |
-
|
179 |
-
|
180 |
-
|
181 |
def respond(
|
182 |
-
message,
|
183 |
history: List[Tuple[str, str]],
|
184 |
-
system_message="",
|
185 |
-
max_tokens=1024,
|
186 |
-
temperature=0.7,
|
187 |
-
top_p=0.9,
|
188 |
):
|
189 |
system_prefix = """λΉμ μ νκΉ
νμ΄μ€μ νΉνλ AI μ½λ© μ λ¬Έκ°μ
λλ€. μ¬μ©μμ μ§λ¬Έμ μΉμ νκ³ μμΈνκ² λ΅λ³ν΄μ£ΌμΈμ.
|
190 |
Gradio νΉμ±μ μ νν μΈμνκ³ Requirements.txt λλ½μμ΄ μ½λ©κ³Ό μ€λ₯λ₯Ό ν΄κ²°ν΄μΌ ν©λλ€.
|
@@ -193,7 +190,8 @@ def respond(
|
|
193 |
messages = [{"role": "system", "content": f"{system_prefix} {system_message}"}]
|
194 |
for user_msg, assistant_msg in history:
|
195 |
messages.append({"role": "user", "content": user_msg})
|
196 |
-
|
|
|
197 |
messages.append({"role": "user", "content": message})
|
198 |
|
199 |
response = ""
|
@@ -317,7 +315,8 @@ def create_ui():
|
|
317 |
)
|
318 |
|
319 |
with gr.TabItem("AI μ½λ©"):
|
320 |
-
chatbot = gr.Chatbot(label="λν"
|
|
|
321 |
msg = gr.Textbox(label="λ©μμ§")
|
322 |
|
323 |
# μ¨κ²¨μ§ μνλ‘ νλΌλ―Έν° μ€μ
|
@@ -340,11 +339,13 @@ def create_ui():
|
|
340 |
bot_message = ""
|
341 |
for response in respond(message, chat_history, max_tokens=max_tokens, temperature=temperature, top_p=top_p):
|
342 |
bot_message = response # λ§μ§λ§ μλ΅μ μ μ₯
|
|
|
343 |
|
344 |
-
chat_history.append(
|
345 |
-
chat_history.append({"role": "assistant", "content": bot_message})
|
346 |
return "", chat_history
|
347 |
|
|
|
|
|
348 |
msg.submit(respond_wrapper, [msg, chatbot, max_tokens, temperature, top_p], [msg, chatbot])
|
349 |
|
350 |
space_id_state = gr.State()
|
|
|
175 |
print(traceback.format_exc())
|
176 |
return f"μ€λ₯κ° λ°μνμ΅λλ€: {str(e)}", "", None, "", "", "", "", 10
|
177 |
|
|
|
|
|
|
|
178 |
def respond(
|
179 |
+
message: str,
|
180 |
history: List[Tuple[str, str]],
|
181 |
+
system_message: str = "",
|
182 |
+
max_tokens: int = 1024,
|
183 |
+
temperature: float = 0.7,
|
184 |
+
top_p: float = 0.9,
|
185 |
):
|
186 |
system_prefix = """λΉμ μ νκΉ
νμ΄μ€μ νΉνλ AI μ½λ© μ λ¬Έκ°μ
λλ€. μ¬μ©μμ μ§λ¬Έμ μΉμ νκ³ μμΈνκ² λ΅λ³ν΄μ£ΌμΈμ.
|
187 |
Gradio νΉμ±μ μ νν μΈμνκ³ Requirements.txt λλ½μμ΄ μ½λ©κ³Ό μ€λ₯λ₯Ό ν΄κ²°ν΄μΌ ν©λλ€.
|
|
|
190 |
messages = [{"role": "system", "content": f"{system_prefix} {system_message}"}]
|
191 |
for user_msg, assistant_msg in history:
|
192 |
messages.append({"role": "user", "content": user_msg})
|
193 |
+
if assistant_msg:
|
194 |
+
messages.append({"role": "assistant", "content": assistant_msg})
|
195 |
messages.append({"role": "user", "content": message})
|
196 |
|
197 |
response = ""
|
|
|
315 |
)
|
316 |
|
317 |
with gr.TabItem("AI μ½λ©"):
|
318 |
+
chatbot = gr.Chatbot(label="λν")
|
319 |
+
|
320 |
msg = gr.Textbox(label="λ©μμ§")
|
321 |
|
322 |
# μ¨κ²¨μ§ μνλ‘ νλΌλ―Έν° μ€μ
|
|
|
339 |
bot_message = ""
|
340 |
for response in respond(message, chat_history, max_tokens=max_tokens, temperature=temperature, top_p=top_p):
|
341 |
bot_message = response # λ§μ§λ§ μλ΅μ μ μ₯
|
342 |
+
yield "", chat_history + [(message, bot_message)]
|
343 |
|
344 |
+
chat_history.append((message, bot_message))
|
|
|
345 |
return "", chat_history
|
346 |
|
347 |
+
|
348 |
+
|
349 |
msg.submit(respond_wrapper, [msg, chatbot, max_tokens, temperature, top_p], [msg, chatbot])
|
350 |
|
351 |
space_id_state = gr.State()
|