Spaces:
Sleeping
Sleeping
Hjgugugjhuhjggg
commited on
Update app.py
Browse files
app.py
CHANGED
@@ -16,6 +16,7 @@ import uvicorn
|
|
16 |
import asyncio
|
17 |
from io import BytesIO
|
18 |
from transformers import pipeline
|
|
|
19 |
|
20 |
AWS_ACCESS_KEY_ID = os.getenv("AWS_ACCESS_KEY_ID")
|
21 |
AWS_SECRET_ACCESS_KEY = os.getenv("AWS_SECRET_ACCESS_KEY")
|
@@ -193,15 +194,17 @@ async def stream_text(model, tokenizer, input_text,
|
|
193 |
chunked_output = [final_output[i:i+10] for i in range(0, len(final_output), 10)]
|
194 |
|
195 |
for chunk in chunked_output:
|
196 |
-
|
|
|
197 |
await asyncio.sleep(chunk_delay)
|
198 |
|
|
|
199 |
break
|
200 |
|
201 |
else:
|
202 |
chunked_output = [new_text[i:i+10] for i in range(0, len(new_text), 10)]
|
203 |
for chunk in chunked_output:
|
204 |
-
yield chunk
|
205 |
await asyncio.sleep(chunk_delay)
|
206 |
|
207 |
|
@@ -210,8 +213,9 @@ async def stream_text(model, tokenizer, input_text,
|
|
210 |
chunked_output = [output_text[i:i+10] for i in range(0, len(output_text), 10)]
|
211 |
|
212 |
for chunk in chunked_output:
|
213 |
-
yield chunk
|
214 |
await asyncio.sleep(chunk_delay)
|
|
|
215 |
break
|
216 |
|
217 |
encoded_input = tokenizer(output_text,
|
|
|
16 |
import asyncio
|
17 |
from io import BytesIO
|
18 |
from transformers import pipeline
|
19 |
+
import json
|
20 |
|
21 |
AWS_ACCESS_KEY_ID = os.getenv("AWS_ACCESS_KEY_ID")
|
22 |
AWS_SECRET_ACCESS_KEY = os.getenv("AWS_SECRET_ACCESS_KEY")
|
|
|
194 |
chunked_output = [final_output[i:i+10] for i in range(0, len(final_output), 10)]
|
195 |
|
196 |
for chunk in chunked_output:
|
197 |
+
|
198 |
+
yield json.dumps({"text": chunk, "is_end": False}) + "\n"
|
199 |
await asyncio.sleep(chunk_delay)
|
200 |
|
201 |
+
yield json.dumps({"text": "", "is_end": True}) + "\n"
|
202 |
break
|
203 |
|
204 |
else:
|
205 |
chunked_output = [new_text[i:i+10] for i in range(0, len(new_text), 10)]
|
206 |
for chunk in chunked_output:
|
207 |
+
yield json.dumps({"text": chunk, "is_end": False}) + "\n"
|
208 |
await asyncio.sleep(chunk_delay)
|
209 |
|
210 |
|
|
|
213 |
chunked_output = [output_text[i:i+10] for i in range(0, len(output_text), 10)]
|
214 |
|
215 |
for chunk in chunked_output:
|
216 |
+
yield json.dumps({"text": chunk, "is_end": False}) + "\n"
|
217 |
await asyncio.sleep(chunk_delay)
|
218 |
+
yield json.dumps({"text": "", "is_end": True}) + "\n"
|
219 |
break
|
220 |
|
221 |
encoded_input = tokenizer(output_text,
|