MarsupialAI
commited on
Commit
•
cafad1a
1
Parent(s):
77234f9
Update README.md
Browse files
README.md
CHANGED
@@ -3,8 +3,8 @@ license: llama2
|
|
3 |
language:
|
4 |
- en
|
5 |
tags:
|
6 |
-
-
|
7 |
-
-
|
8 |
- chat
|
9 |
- storywriting
|
10 |
---
|
@@ -14,7 +14,7 @@ tags:
|
|
14 |
This model is a rotating-stack merge of three 70b models in a 103b (120 layer) configuration inspired by Venus 103b. The result of
|
15 |
this "frankenmerge" is a large model that contains a little bit of everything - including the kitchen sink. RP, chat, storywriting,
|
16 |
and instruct are all well supported. It may or may not code well - I lack the expertise to test it in that capacity, but considering
|
17 |
-
the source models, it is unlikely.
|
18 |
|
19 |
Component models for the rotating stack are
|
20 |
- royallab/Aetheria-L2-70B
|
@@ -24,8 +24,13 @@ Component models for the rotating stack are
|
|
24 |
Components of those models are purported to include: Nous-Hermes-Llama2-70b, Xwin-LM-7B-V0.1, Mythospice-70b, Euryale-1.3-L2-70B,
|
25 |
tulu-2-dpo-70b, GOAT-70B-Storytelling, Platypus2-70B-instruct, Lila-70B, SunsetBoulevard, and some private LoRAs.
|
26 |
|
|
|
|
|
|
|
27 |
This model is uncensored and perfectly capable of generating objectionable material. However, it is not an explicitely-NSFW model,
|
28 |
-
and it has never "gone rogue" and tried to insert NSFW content into SFW prompts in my experience.
|
|
|
|
|
29 |
|
30 |
|
31 |
|
|
|
3 |
language:
|
4 |
- en
|
5 |
tags:
|
6 |
+
- rp
|
7 |
+
- erp
|
8 |
- chat
|
9 |
- storywriting
|
10 |
---
|
|
|
14 |
This model is a rotating-stack merge of three 70b models in a 103b (120 layer) configuration inspired by Venus 103b. The result of
|
15 |
this "frankenmerge" is a large model that contains a little bit of everything - including the kitchen sink. RP, chat, storywriting,
|
16 |
and instruct are all well supported. It may or may not code well - I lack the expertise to test it in that capacity, but considering
|
17 |
+
the source models, I suspect it is unlikely.
|
18 |
|
19 |
Component models for the rotating stack are
|
20 |
- royallab/Aetheria-L2-70B
|
|
|
24 |
Components of those models are purported to include: Nous-Hermes-Llama2-70b, Xwin-LM-7B-V0.1, Mythospice-70b, Euryale-1.3-L2-70B,
|
25 |
tulu-2-dpo-70b, GOAT-70B-Storytelling, Platypus2-70B-instruct, Lila-70B, SunsetBoulevard, and some private LoRAs.
|
26 |
|
27 |
+
As all components are based on Llama2 70b, native context length is 4k tokens. Coherency out to 8k is extremely good with rope scaling,
|
28 |
+
but starts to decline beyond that.
|
29 |
+
|
30 |
This model is uncensored and perfectly capable of generating objectionable material. However, it is not an explicitely-NSFW model,
|
31 |
+
and it has never "gone rogue" and tried to insert NSFW content into SFW prompts in my experience. As with any LLM, no factual claims
|
32 |
+
made by the model should be taken at face value. You know that boilerplate safety disclaimer that most professional models have?
|
33 |
+
Assume this has it too. This model is for entertainment purposes only.
|
34 |
|
35 |
|
36 |
|