1inkusFace commited on
Commit
dd12d34
·
verified ·
1 Parent(s): b7ad091

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +11 -9
app.py CHANGED
@@ -27,15 +27,17 @@ def generate_code(prompt):
27
  add_generation_prompt=True
28
  )
29
  model_inputs = tokenizer([text], return_tensors="pt").to(model.device)
30
- with torch.no_grad():
31
- generated_ids = model.generate(
32
- **model_inputs,
33
- max_new_tokens=1024
34
- )
35
- generated_ids = [
36
- output_ids[len(input_ids):] for input_ids, output_ids in zip(model_inputs.input_ids, generated_ids)
37
- ]
38
- response = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)[0]
 
 
39
  return response
40
 
41
  with gr.Blocks(title="Qwen 14b") as demo: # Updated title
 
27
  add_generation_prompt=True
28
  )
29
  model_inputs = tokenizer([text], return_tensors="pt").to(model.device)
30
+ #with torch.no_grad():
31
+ generated_ids = model.generate(
32
+ **model_inputs,
33
+ max_length = 1024,
34
+ min_new_tokens = 256,
35
+ #low_memory = True,
36
+ token_healing = True,
37
+ guidance_scale = 3.8,
38
+ )
39
+ generated_ids = [output_ids[len(input_ids):] for input_ids, output_ids in zip(model_inputs.input_ids, generated_ids)]
40
+ response = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)[0]
41
  return response
42
 
43
  with gr.Blocks(title="Qwen 14b") as demo: # Updated title