Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -13,32 +13,8 @@ from ctransformers import AutoModelForCausalLM
|
|
13 |
from dl_hf_model import dl_hf_model
|
14 |
from loguru import logger
|
15 |
|
16 |
-
filename_list = [
|
17 |
-
"Wizard-Vicuna-7B-Uncensored.ggmlv3.q2_K.bin",
|
18 |
-
"Wizard-Vicuna-7B-Uncensored.ggmlv3.q3_K_L.bin",
|
19 |
-
"Wizard-Vicuna-7B-Uncensored.ggmlv3.q3_K_M.bin",
|
20 |
-
"Wizard-Vicuna-7B-Uncensored.ggmlv3.q3_K_S.bin",
|
21 |
-
"Wizard-Vicuna-7B-Uncensored.ggmlv3.q4_0.bin",
|
22 |
-
"Wizard-Vicuna-7B-Uncensored.ggmlv3.q4_1.bin",
|
23 |
-
"Wizard-Vicuna-7B-Uncensored.ggmlv3.q4_K_M.bin",
|
24 |
-
"Wizard-Vicuna-7B-Uncensored.ggmlv3.q4_K_S.bin",
|
25 |
-
"Wizard-Vicuna-7B-Uncensored.ggmlv3.q5_0.bin",
|
26 |
-
"Wizard-Vicuna-7B-Uncensored.ggmlv3.q5_1.bin",
|
27 |
-
"Wizard-Vicuna-7B-Uncensored.ggmlv3.q5_K_M.bin",
|
28 |
-
"Wizard-Vicuna-7B-Uncensored.ggmlv3.q5_K_S.bin",
|
29 |
-
"Wizard-Vicuna-7B-Uncensored.ggmlv3.q6_K.bin",
|
30 |
-
"Wizard-Vicuna-7B-Uncensored.ggmlv3.q8_0.bin",
|
31 |
-
]
|
32 |
-
|
33 |
-
URL = "https://huggingface.co/TheBloke/Wizard-Vicuna-7B-Uncensored-GGML/raw/main/Wizard-Vicuna-7B-Uncensored.ggmlv3.q4_K_M.bin" # 4.05G
|
34 |
-
|
35 |
-
url = "https://huggingface.co/savvamadar/ggml-gpt4all-j-v1.3-groovy/blob/main/ggml-gpt4all-j-v1.3-groovy.bin"
|
36 |
-
url = "https://huggingface.co/TheBloke/Llama-2-13B-GGML/blob/main/llama-2-13b.ggmlv3.q4_K_S.bin" # 7.37G
|
37 |
-
# url = "https://huggingface.co/TheBloke/Llama-2-13B-chat-GGML/blob/main/llama-2-13b-chat.ggmlv3.q3_K_L.bin"
|
38 |
-
url = "https://huggingface.co/TheBloke/Llama-2-13B-chat-GGML/blob/main/llama-2-13b-chat.ggmlv3.q3_K_L.bin" # 6.93G
|
39 |
-
# url = "https://huggingface.co/TheBloke/Llama-2-13B-chat-GGML/blob/main/llama-2-13b-chat.ggmlv3.q3_K_L.binhttps://huggingface.co/TheBloke/Llama-2-13B-chat-GGML/blob/main/llama-2-13b-chat.ggmlv3.q4_K_M.bin" # 7.87G
|
40 |
|
41 |
-
url = "https://huggingface.co/
|
42 |
|
43 |
_ = (
|
44 |
"golay" in platform.node()
|
@@ -51,7 +27,6 @@ _ = (
|
|
51 |
if _:
|
52 |
# url = "https://huggingface.co/TheBloke/Llama-2-13B-chat-GGML/blob/main/llama-2-13b-chat.ggmlv3.q2_K.bin"
|
53 |
url = "https://huggingface.co/TheBloke/Llama-2-7B-Chat-GGML/blob/main/llama-2-7b-chat.ggmlv3.q2_K.bin" # 2.87G
|
54 |
-
url = "https://huggingface.co/TheBloke/Llama-2-7B-Chat-GGML/blob/main/llama-2-7b-chat.ggmlv3.q4_K_M.bin" # 2.87G
|
55 |
|
56 |
|
57 |
prompt_template = """Below is an instruction that describes a task. Write a response that appropriately completes the request.
|
|
|
13 |
from dl_hf_model import dl_hf_model
|
14 |
from loguru import logger
|
15 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
16 |
|
17 |
+
url = "https://huggingface.co/TheBloke/Llama-2-7B-Chat-GGML/blob/main/llama-2-7b-chat.ggmlv3.q2_K.bin" # 7.37G
|
18 |
|
19 |
_ = (
|
20 |
"golay" in platform.node()
|
|
|
27 |
if _:
|
28 |
# url = "https://huggingface.co/TheBloke/Llama-2-13B-chat-GGML/blob/main/llama-2-13b-chat.ggmlv3.q2_K.bin"
|
29 |
url = "https://huggingface.co/TheBloke/Llama-2-7B-Chat-GGML/blob/main/llama-2-7b-chat.ggmlv3.q2_K.bin" # 2.87G
|
|
|
30 |
|
31 |
|
32 |
prompt_template = """Below is an instruction that describes a task. Write a response that appropriately completes the request.
|