Tuchuanhuhuhu commited on
Commit
2342c7b
·
1 Parent(s): 33cbbdb

支持高级参数调节

Browse files
Files changed (1) hide show
  1. modules/models/StableLM.py +8 -5
modules/models/StableLM.py CHANGED
@@ -39,6 +39,9 @@ class StableLM_Client(BaseLLMModel):
39
  - StableAssistant is excited to be able to help the user, but will refuse to do anything that could be considered harmful to the user.
40
  - StableAssistant is more than just an information source, StableAssistant is also able to write poetry, short stories, and make jokes.
41
  - StableAssistant will refuse to participate in anything that could harm a human."""
 
 
 
42
 
43
  def _get_stablelm_style_input(self):
44
  history = self.history + [{"role": "assistant", "content": ""}]
@@ -50,8 +53,8 @@ class StableLM_Client(BaseLLMModel):
50
 
51
  def _generate(self, text, bad_text=None):
52
  stop = StopOnTokens()
53
- result = self.generator(text, max_new_tokens=1024, num_return_sequences=1, num_beams=1, do_sample=True,
54
- temperature=1.0, top_p=0.95, top_k=1000, stopping_criteria=StoppingCriteriaList([stop]))
55
  return result[0]["generated_text"].replace(text, "")
56
 
57
  def get_answer_at_once(self):
@@ -68,11 +71,11 @@ class StableLM_Client(BaseLLMModel):
68
  generate_kwargs = dict(
69
  model_inputs,
70
  streamer=streamer,
71
- max_new_tokens=1024,
72
  do_sample=True,
73
- top_p=0.95,
74
  top_k=1000,
75
- temperature=1.0,
76
  num_beams=1,
77
  stopping_criteria=StoppingCriteriaList([stop])
78
  )
 
39
  - StableAssistant is excited to be able to help the user, but will refuse to do anything that could be considered harmful to the user.
40
  - StableAssistant is more than just an information source, StableAssistant is also able to write poetry, short stories, and make jokes.
41
  - StableAssistant will refuse to participate in anything that could harm a human."""
42
+ self.max_generation_token = 1024
43
+ self.top_p = 0.95
44
+ self.temperature = 1.0
45
 
46
  def _get_stablelm_style_input(self):
47
  history = self.history + [{"role": "assistant", "content": ""}]
 
53
 
54
  def _generate(self, text, bad_text=None):
55
  stop = StopOnTokens()
56
+ result = self.generator(text, max_new_tokens=self.max_generation_token, num_return_sequences=1, num_beams=1, do_sample=True,
57
+ temperature=self.temperature, top_p=self.top_p, top_k=1000, stopping_criteria=StoppingCriteriaList([stop]))
58
  return result[0]["generated_text"].replace(text, "")
59
 
60
  def get_answer_at_once(self):
 
71
  generate_kwargs = dict(
72
  model_inputs,
73
  streamer=streamer,
74
+ max_new_tokens=self.max_generation_token,
75
  do_sample=True,
76
+ top_p=self.top_p,
77
  top_k=1000,
78
+ temperature=self.temperature,
79
  num_beams=1,
80
  stopping_criteria=StoppingCriteriaList([stop])
81
  )