Update README.md
Browse files
README.md
CHANGED
@@ -1,3 +1,14 @@
|
|
1 |
-
---
|
2 |
-
license: llama3
|
3 |
-
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: llama3
|
3 |
+
---
|
4 |
+
|
5 |
+
This is an experimental 2x8B moe with random gates, using the following 2 models
|
6 |
+
|
7 |
+
- Hermes-2-Theta-l3-8B by Nous Research https://huggingface.co/NousResearch/Hermes-2-Theta-Llama-3-8B
|
8 |
+
|
9 |
+
- llama-3-cat-8B-instruct-V1 by TheSkullery https://huggingface.co/TheSkullery/llama-3-cat-8b-instruct-v1
|
10 |
+
|
11 |
+
|
12 |
+
***Important***
|
13 |
+
|
14 |
+
Make sure to add `</s>` a stop sequence as it uses llama-3-cat-8B-instruct-V1 as the base model.
|