Spaces:
Running
Running
Update app.py
Browse files
app.py
CHANGED
@@ -21,8 +21,8 @@ top_p=0.99
|
|
21 |
repetition_penalty=1.0
|
22 |
|
23 |
def compress_history(history,seed,temperature=temperature,top_p=top_p,repetition_penalty=repetition_penalty):
|
24 |
-
|
25 |
-
client = InferenceClient("abacusai/Slerp-CM-mist-dpo")
|
26 |
|
27 |
print("COMPRESSING")
|
28 |
formatted_prompt=f"{COMPRESS_HISTORY.format(history=history)}"
|
@@ -44,8 +44,8 @@ MAX_HISTORY=100
|
|
44 |
opts=[]
|
45 |
def generate(prompt, history,max_new_tokens,health,seed,temperature=temperature,top_p=top_p,repetition_penalty=repetition_penalty):
|
46 |
opts.clear()
|
47 |
-
|
48 |
-
client = InferenceClient("abacusai/Slerp-CM-mist-dpo")
|
49 |
|
50 |
temperature = float(temperature)
|
51 |
if temperature < 1e-2:
|
|
|
21 |
repetition_penalty=1.0
|
22 |
|
23 |
def compress_history(history,seed,temperature=temperature,top_p=top_p,repetition_penalty=repetition_penalty):
|
24 |
+
client = InferenceClient("mistralai/Mixtral-8x7B-Instruct-v0.1")
|
25 |
+
#client = InferenceClient("abacusai/Slerp-CM-mist-dpo")
|
26 |
|
27 |
print("COMPRESSING")
|
28 |
formatted_prompt=f"{COMPRESS_HISTORY.format(history=history)}"
|
|
|
44 |
opts=[]
|
45 |
def generate(prompt, history,max_new_tokens,health,seed,temperature=temperature,top_p=top_p,repetition_penalty=repetition_penalty):
|
46 |
opts.clear()
|
47 |
+
client = InferenceClient("mistralai/Mixtral-8x7B-Instruct-v0.1")
|
48 |
+
#client = InferenceClient("abacusai/Slerp-CM-mist-dpo")
|
49 |
|
50 |
temperature = float(temperature)
|
51 |
if temperature < 1e-2:
|