Steelskull commited on
Commit
27bfbe0
1 Parent(s): 23d2a0c

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +103 -0
README.md ADDED
@@ -0,0 +1,103 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Lumosia-v2-MoE-4x10.7
2
+
3
+ "Lumosia" was selected as its a MoE of Multiple SOLAR Merges so it really "Lights the way".... its 3am.
4
+
5
+ This is a very experimantal model. its a MoE of all good performing Solar models (based off of personal experiance not open leaderboard),
6
+ The models goal was to make a good all rounder, in chat/logic/rp
7
+
8
+ Why? Dunno whated to see what would happen
9
+
10
+ context is 4k but coherent up to 16k
11
+
12
+ A Lumosia Personality tavern card has been added
13
+
14
+ Come join the Discord:
15
+ [ConvexAI](https://discord.gg/yYqmNmg7Wj)
16
+
17
+
18
+ Template:
19
+ ```
20
+ ### System:
21
+
22
+ ### USER:{prompt}
23
+
24
+ ### Assistant:
25
+ ```
26
+
27
+
28
+ Settings:
29
+ ```
30
+ Temp: 1.0
31
+ min-p: 0.02-0.1
32
+ ```
33
+
34
+ ## Evals:
35
+
36
+ * Avg:
37
+ * ARC:
38
+ * HellaSwag:
39
+ * MMLU:
40
+ * T-QA:
41
+ * Winogrande:
42
+ * GSM8K:
43
+
44
+ ## Examples:
45
+ ```
46
+ Example 1:
47
+
48
+ User:
49
+
50
+ Lumosia:
51
+
52
+ ```
53
+ ```
54
+ Example 2:
55
+
56
+ User:
57
+
58
+ Lumosia:
59
+
60
+ ```
61
+
62
+ ## 🧩 Configuration
63
+
64
+ ```
65
+ yaml
66
+ base_model: DopeorNope/SOLARC-M-10.7B
67
+ gate_mode: hidden
68
+ dtype: bfloat16
69
+ experts:
70
+ - source_model: DopeorNope/SOLARC-M-10.7B
71
+ positive_prompts: [""]
72
+ - source_model: maywell/PiVoT-10.7B-Mistral-v0.2-RP
73
+ positive_prompts: [""]
74
+ - source_model: kyujinpy/Sakura-SOLAR-Instruct
75
+ positive_prompts: [""]
76
+ - source_model: jeonsworld/CarbonVillain-en-10.7B-v1
77
+ positive_prompts: [""]
78
+ ```
79
+
80
+ ## 💻 Usage
81
+
82
+ ```
83
+ python
84
+ !pip install -qU transformers bitsandbytes accelerate
85
+
86
+ from transformers import AutoTokenizer
87
+ import transformers
88
+ import torch
89
+
90
+ model = "Steelskull/Lumosia-MoE-4x10.7"
91
+
92
+ tokenizer = AutoTokenizer.from_pretrained(model)
93
+ pipeline = transformers.pipeline(
94
+ "text-generation",
95
+ model=model,
96
+ model_kwargs={"torch_dtype": torch.float16, "load_in_4bit": True},
97
+ )
98
+
99
+ messages = [{"role": "user", "content": "Explain what a Mixture of Experts is in less than 100 words."}]
100
+ prompt = pipeline.tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
101
+ outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
102
+ print(outputs[0]["generated_text"])
103
+ ```