Update app.py
Browse files
app.py
CHANGED
@@ -9,14 +9,14 @@ model_id = "01-ai/Yi-34B-200K"
|
|
9 |
|
10 |
os.environ['PYTORCH_CUDA_ALLOC_CONF'] = 'max_split_size_mb:50'
|
11 |
device = "cuda" if torch.cuda.is_available() else "cpu"
|
12 |
-
offload_directory = './model_offload'
|
13 |
-
if not os.path.exists(offload_directory):
|
14 |
-
|
|
|
|
|
15 |
|
16 |
tokenizer = YiTokenizer(vocab_file="./tokenizer.model")
|
17 |
-
model = AutoModelForCausalLM.from_pretrained(model_id, device_map="auto", load_in_8bit=True,
|
18 |
-
# model = AutoModelForCausalLM.from_pretrained(model_id, trust_remote_code=True)
|
19 |
-
# model = model.to(device)
|
20 |
|
21 |
def run(message, chat_history, max_new_tokens=4056, temperature=3.5, top_p=0.9, top_k=800):
|
22 |
prompt = get_prompt(message, chat_history)
|
|
|
9 |
|
10 |
os.environ['PYTORCH_CUDA_ALLOC_CONF'] = 'max_split_size_mb:50'
|
11 |
device = "cuda" if torch.cuda.is_available() else "cpu"
|
12 |
+
# offload_directory = './model_offload'
|
13 |
+
# if not os.path.exists(offload_directory):
|
14 |
+
# os.makedirs(offload_directory)
|
15 |
+
# model = AutoModelForCausalLM.from_pretrained(model_id, device_map="auto", load_in_8bit=True, load_in_8bit_fp32_cpu_offload=True, offload_folder=offload_directory, trust_remote_code=True)
|
16 |
+
# model = model.to(device)
|
17 |
|
18 |
tokenizer = YiTokenizer(vocab_file="./tokenizer.model")
|
19 |
+
model = AutoModelForCausalLM.from_pretrained(model_id, device_map="auto", load_in_8bit=True, trust_remote_code=True)
|
|
|
|
|
20 |
|
21 |
def run(message, chat_history, max_new_tokens=4056, temperature=3.5, top_p=0.9, top_k=800):
|
22 |
prompt = get_prompt(message, chat_history)
|