Spaces:
Runtime error
Runtime error
ffreemt
commited on
Commit
·
0e7c90d
1
Parent(s):
1132eaf
Update block.load to update buff Textbox
Browse files
app.py
CHANGED
@@ -219,6 +219,12 @@ logger.info("start dl")
|
|
219 |
_ = """full url: https://huggingface.co/TheBloke/mpt-30B-chat-GGML/blob/main/mpt-30b-chat.ggmlv0.q4_1.bin"""
|
220 |
|
221 |
repo_id = "TheBloke/mpt-30B-chat-GGML"
|
|
|
|
|
|
|
|
|
|
|
|
|
222 |
model_filename = "mpt-30b-chat.ggmlv0.q4_1.bin"
|
223 |
destination_folder = "models"
|
224 |
|
@@ -431,8 +437,8 @@ with gr.Blocks(
|
|
431 |
show_progress="full",
|
432 |
)
|
433 |
|
434 |
-
# update buff Textbox
|
435 |
-
block.
|
436 |
|
437 |
# concurrency_count=5, max_size=20
|
438 |
# max_size=36, concurrency_count=14
|
|
|
219 |
_ = """full url: https://huggingface.co/TheBloke/mpt-30B-chat-GGML/blob/main/mpt-30b-chat.ggmlv0.q4_1.bin"""
|
220 |
|
221 |
repo_id = "TheBloke/mpt-30B-chat-GGML"
|
222 |
+
|
223 |
+
# https://huggingface.co/TheBloke/mpt-30B-chat-GGML
|
224 |
+
_ = """
|
225 |
+
mpt-30b-chat.ggmlv0.q4_0.bin q4_0 4 16.85 GB 19.35 GB 4-bit.
|
226 |
+
mpt-30b-chat.ggmlv0.q4_1.bin q4_1 4 18.73 GB 21.23 GB 4-bit. Higher accuracy than q4_0 but not as high as q5_0. However has quicker inference than q5 models.
|
227 |
+
"""
|
228 |
model_filename = "mpt-30b-chat.ggmlv0.q4_1.bin"
|
229 |
destination_folder = "models"
|
230 |
|
|
|
437 |
show_progress="full",
|
438 |
)
|
439 |
|
440 |
+
# update buff Textbox, every: units in seconds)
|
441 |
+
block.run_forever(lambda: ns.response, None, [buff], every=1)
|
442 |
|
443 |
# concurrency_count=5, max_size=20
|
444 |
# max_size=36, concurrency_count=14
|