Spaces:
Sleeping
Sleeping
update
Browse files- config.yaml +14 -0
config.yaml
CHANGED
@@ -3,6 +3,20 @@ model_list:
|
|
3 |
litellm_params:
|
4 |
model: ollama/mistral:7b
|
5 |
api_base: https://zhengr-ollama.hf.space
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
6 |
|
7 |
litellm_settings: # module level litellm settings - https://github.com/BerriAI/litellm/blob/main/litellm/__init__.py
|
8 |
drop_params: True
|
|
|
3 |
litellm_params:
|
4 |
model: ollama/mistral:7b
|
5 |
api_base: https://zhengr-ollama.hf.space
|
6 |
+
# Model-specific parameters
|
7 |
+
#model: "huggingface/mistralai/Mistral-7B-Instruct-v0.1"
|
8 |
+
#api_base: "<your-api-base>"
|
9 |
+
#api_key: "<your-api-key>" # [OPTIONAL] for hf inference endpoints
|
10 |
+
#initial_prompt_value: "\n"
|
11 |
+
#roles: {"system":{"pre_message":"<|im_start|>system\n", "post_message":"<|im_end|>"}, "assistant":{"pre_message":"<|im_start|>assistant\n","post_message":"<|im_end|>"}, "user":{"pre_message":"<|im_start|>user\n","post_message":"<|im_end|>"}}
|
12 |
+
#final_prompt_value: "\n"
|
13 |
+
#bos_token: "<s>"
|
14 |
+
#eos_token: "</s>"
|
15 |
+
#max_tokens: 4096
|
16 |
+
- model_name: xinference-llama-3-instruct
|
17 |
+
litellm_params:
|
18 |
+
model: xinference/llama-3-instruct
|
19 |
+
api_base: https://zhengr-xinference.hf.space
|
20 |
|
21 |
litellm_settings: # module level litellm settings - https://github.com/BerriAI/litellm/blob/main/litellm/__init__.py
|
22 |
drop_params: True
|