Update README.md
Browse files
README.md
CHANGED
@@ -5,7 +5,7 @@ library_name: transformers
|
|
5 |
tags:
|
6 |
- mergekit
|
7 |
- merge
|
8 |
-
|
9 |
---
|
10 |
# Lumina-5.5-Instruct
|
11 |
Lumina-5.5-Instruct is a Mixture of Experts (MoE) made with [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing). This model uses a context window of up to 32k.
|
@@ -35,4 +35,4 @@ messages = [{"role": "user", "content": "Explain what a Mixture of Experts is in
|
|
35 |
prompt = pipeline.tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
|
36 |
outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
|
37 |
print(outputs[0]["generated_text"])
|
38 |
-
```
|
|
|
5 |
tags:
|
6 |
- mergekit
|
7 |
- merge
|
8 |
+
license: apache-2.0
|
9 |
---
|
10 |
# Lumina-5.5-Instruct
|
11 |
Lumina-5.5-Instruct is a Mixture of Experts (MoE) made with [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing). This model uses a context window of up to 32k.
|
|
|
35 |
prompt = pipeline.tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
|
36 |
outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
|
37 |
print(outputs[0]["generated_text"])
|
38 |
+
```
|