Update app.py
Browse files
app.py
CHANGED
@@ -133,13 +133,14 @@ def chat_with_model(prompt, document_section, model_choice='gpt-3.5-turbo'):
|
|
133 |
result_textarea = st.empty()
|
134 |
results=[]
|
135 |
for responses in openai.ChatCompletion.create(model=model, messages=conversation, stream=True):
|
136 |
-
results.append(responses
|
137 |
result = "".join(results).strip()
|
138 |
result = result.replace('\n','')
|
139 |
result_textarea.markdown(f'*{result}*')
|
140 |
|
141 |
#return response
|
142 |
-
return response['choices'][0]['message']['content']
|
|
|
143 |
|
144 |
|
145 |
def chat_with_file_contents(prompt, file_content, model_choice='gpt-3.5-turbo'):
|
|
|
133 |
result_textarea = st.empty()
|
134 |
results=[]
|
135 |
for responses in openai.ChatCompletion.create(model=model, messages=conversation, stream=True):
|
136 |
+
results.append(responses['choices'][0].text)
|
137 |
result = "".join(results).strip()
|
138 |
result = result.replace('\n','')
|
139 |
result_textarea.markdown(f'*{result}*')
|
140 |
|
141 |
#return response
|
142 |
+
#return response['choices'][0]['message']['content']
|
143 |
+
return results
|
144 |
|
145 |
|
146 |
def chat_with_file_contents(prompt, file_content, model_choice='gpt-3.5-turbo'):
|