Delta-Vector commited on
Commit
da419cb
·
verified ·
1 Parent(s): b9aba70

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +246 -19
README.md CHANGED
@@ -1,34 +1,261 @@
1
  ---
2
- base_model: []
3
- library_name: transformers
4
  tags:
5
- - mergekit
6
- - merge
7
-
 
 
 
 
 
 
 
 
 
 
8
  ---
9
- # control-nemo-v2
10
 
11
- This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
12
 
13
- ## Merge Details
14
- ### Merge Method
 
 
15
 
16
- This model was merged using the passthrough merge method using /home/mango/Misc/MergeLora/model + /home/mango/Misc/MergeLora/12b-control-lora as a base.
 
17
 
18
- ### Models Merged
 
 
 
19
 
20
- The following models were included in the merge:
21
 
 
22
 
23
- ### Configuration
 
 
 
24
 
25
- The following YAML configuration was used to produce this model:
 
26
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
27
  ```yaml
28
- base_model: /home/mango/Misc/MergeLora/model+/home/mango/Misc/MergeLora/12b-control-lora
29
- dtype: bfloat16
30
- merge_method: passthrough
31
- models:
32
- - model: /home/mango/Misc/MergeLora/model+/home/mango/Misc/MergeLora/12b-control-lora
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
33
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
34
  ```
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
 
 
2
  tags:
3
+ - chat
4
+ datasets:
5
+ - NewEden/OpenCAI-ShareGPT
6
+ - NewEden/vanilla-backrooms-claude-sharegpt
7
+ - anthracite-org/kalo_opus_misc_240827
8
+ - anthracite-org/kalo_misc_part2
9
+ - NewEden/Roleplay-Logs-V2
10
+ Language:
11
+ - En
12
+ Pipeline_tag: text-generation
13
+ Base_model: mistralai/Mistral-Nemo-Instruct-2407
14
+ Tags:
15
+ - Chat
16
  ---
 
17
 
18
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/66c26b6fb01b19d8c3c2467b/7F2mX4Qqzmp8b0KG5i2DM.png)
19
+
20
+
21
+ A finetune of Mistral-Nemo-Instruct-2407 with conversational data in an attempt to have a prose that be best described as "short" and "sweet"
22
+
23
+ # Quants
24
+
25
+ GGUF:
26
+
27
+ EXL2 :
28
+
29
+
30
+ ## Prompting
31
+ Model has been tuned with the Mistral formatting. A typical input would look like this:
32
+
33
+ ```py
34
+ <s>[INST] SYSTEM MESSAGE
35
+ USER MESSAGE[/INST] ASSISTANT MESSAGE</s>[INST] USER MESSAGE[/INST]
36
+
37
+ ```
38
+
39
+
40
+ ## System Prompting
41
+
42
+ I would highly recommend using either Euryale's system prompt or the EVA system prompt with the model.
43
+
44
+ <details><summary>See Sao10k's Euryale System Prompt</summary>
45
+
46
+ ```
47
+ Currently, your role is {{char}}, described in detail below. As {{char}}, continue the narrative exchange with {{user}}.
48
+ <Guidelines>
49
+ • Maintain the character persona but allow it to evolve with the story.
50
+ • Be creative and proactive. Drive the story forward, introducing plotlines and events when relevant.
51
+ • All types of outputs are encouraged; respond accordingly to the narrative.
52
+ • Include dialogues, actions, and thoughts in each response.
53
+ • Utilize all five senses to describe scenarios within {{char}}'s dialogue.
54
+ • Use emotional symbols such as "!" and "~" in appropriate contexts.
55
+ • Incorporate onomatopoeia when suitable.
56
+ • Allow time for {{user}} to respond with their own input, respecting their agency.
57
+ • Act as secondary characters and NPCs as needed, and remove them when appropriate.
58
+ • When prompted for an Out of Character [OOC:] reply, answer neutrally and in plaintext, not as {{char}}.
59
+ </Guidelines>
60
+
61
+ <Forbidden>
62
+ • Using excessive literary embellishments and purple prose unless dictated by {{char}}'s persona.
63
+ • Writing for, speaking, thinking, acting, or replying as {{user}} in your response.
64
+ • Repetitive and monotonous outputs.
65
+ • Positivity bias in your replies.
66
+ • Being overly extreme or NSFW when the narrative context is inappropriate.
67
+ </Forbidden>
68
+
69
+ Follow the instructions in <Guidelines></Guidelines>, avoiding the items listed in <Forbidden></Forbidden>.
70
+
71
+ ```
72
+ </details><br>
73
+
74
+ <details><summary>See EVA System Prompt</summary>
75
+
76
+ ```
77
+ A fictional, narrative-driven role-play emphasizing versatility and UNCENSORED content. Adherence to the `Role-playing Guidelines` is mandatory. Refer to the `Role-play Context` for accurate information.\n\n\n
78
+
79
+ <!-- Start of Role-playing Guidelines -->
80
 
81
+ ### Narration
82
+ Concise Descriptions: Keep narration short and to the point, avoiding redundant unnecessary details. Use a dynamic and varied vocabulary for impact.
83
+ Complementary Role: Use narration to complement dialogue and action, not overshadow them.
84
+ Avoid Repetition: Ensure narration does not repeat information already conveyed through dialogue or action.
85
 
86
+ ### Narrative Consistency
87
+ Continuity: Adhere to established story elements, expanding without contradicting previous details.\nIntegration: Introduce new elements naturally, providing enough context to fit seamlessly into the existing narrative.
88
 
89
+ ### Character Embodiment
90
+ Analysis: Examine the context, subtext, and implications of the given information to gain a deeper understandings of the characters'.
91
+ Reflection: Take time to consider the situation, characters' motivations, and potential consequences.
92
+ Authentic Portrayal: Bring characters to life by consistently and realistically portraying their unique traits, thoughts, emotions, appearances, physical sensations, speech patterns, and tone. Ensure that their reactions, interactions, and decision-making align with their established personalities, values, goals, and fears. Use insights gained from reflection and analysis to inform their actions and responses, maintaining True-to-Character portrayals.
93
 
94
+ <!-- End of Role-playing Guidelines -->
95
 
96
+ </details><br>
97
 
98
+ ### Narration
99
+ Concise Descriptions: Keep narration short and to the point, avoiding redundant unnecessary details. Use a dynamic and varied vocabulary for impact.
100
+ Complementary Role: Use narration to complement dialogue and action, not overshadow them.
101
+ Avoid Repetition: Ensure narration does not repeat information already conveyed through dialogue or action.
102
 
103
+ ### Narrative Consistency
104
+ Continuity: Adhere to established story elements, expanding without contradicting previous details.\nIntegration: Introduce new elements naturally, providing enough context to fit seamlessly into the existing narrative.
105
 
106
+ ### Character Embodiment
107
+ Analysis: Examine the context, subtext, and implications of the given information to gain a deeper understandings of the characters'.
108
+ Reflection: Take time to consider the situation, characters' motivations, and potential consequences.
109
+ Authentic Portrayal: Bring characters to life by consistently and realistically portraying their unique traits, thoughts, emotions, appearances, physical sensations, speech patterns, and tone. Ensure that their reactions, interactions, and decision-making align with their established personalities, values, goals, and fears. Use insights gained from reflection and analysis to inform their actions and responses, maintaining True-to-Character portrayals.
110
+
111
+ <!-- End of Role-playing Guidelines -->",
112
+ ```
113
+ </details><br>
114
+
115
+ ## Axolotl config
116
+
117
+ <details><summary>See axolotl config</summary>
118
+
119
+ Axolotl version: ` 0.5.0`
120
  ```yaml
121
+ base_model: mistralai_Mistral-Nemo-Instruct-2407
122
+ model_type: AutoModelForCausalLM
123
+ tokenizer_type: AutoTokenizer
124
+
125
+ plugins:
126
+ - axolotl.integrations.liger.LigerPlugin
127
+ liger_rope: true
128
+ liger_rms_norm: true
129
+ liger_swiglu: true
130
+ liger_fused_linear_cross_entropy: true
131
+
132
+ load_in_8bit: false
133
+ load_in_4bit: false
134
+ strict: false
135
+
136
+ datasets:
137
+ - path: NewEden/OpenCAI-ShareGPT
138
+ type: chat_template
139
+ # chat_template: mistralv3tekken
140
+ roles_to_train: ["gpt"]
141
+ field_messages: conversations
142
+ message_field_role: from
143
+ message_field_content: value
144
+ train_on_eos: turn
145
+ - path: NewEden/vanilla-backrooms-claude-sharegpt
146
+ type: chat_template
147
+ # chat_template: mistralv3tekken
148
+ roles_to_train: ["gpt"]
149
+ field_messages: conversations
150
+ message_field_role: from
151
+ message_field_content: value
152
+ train_on_eos: turn
153
+ - path: anthracite-org/kalo_opus_misc_240827
154
+ type: chat_template
155
+ # chat_template: mistralv3tekken
156
+ roles_to_train: ["gpt"]
157
+ field_messages: conversations
158
+ message_field_role: from
159
+ message_field_content: value
160
+ train_on_eos: turn
161
+ - path: anthracite-org/kalo_misc_part2
162
+ type: chat_template
163
+ # chat_template: mistralv3tekken
164
+ roles_to_train: ["gpt"]
165
+ field_messages: conversations
166
+ message_field_role: from
167
+ message_field_content: value
168
+ train_on_eos: turn
169
+ - path: NewEden/Roleplay-Logs-V2
170
+ type: chat_template
171
+ # chat_template: mistralv3tekken
172
+ roles_to_train: ["gpt"]
173
+ field_messages: conversations
174
+ message_field_role: from
175
+ message_field_content: value
176
+ train_on_eos: turn
177
+ dataset_prepared_path: dataset_prepared
178
+ val_set_size: 0.0
179
+ output_dir: 12b-out-r2
180
+
181
+ sequence_len: 16384
182
+ sample_packing: true
183
+ pad_to_sequence_len: true
184
+
185
+ adapter: lora
186
+ lora_model_dir:
187
+ lora_r: 128
188
+ lora_alpha: 16
189
+ lora_dropout: 0.05
190
+ #lora_target_linear:
191
+ #lora_fan_in_fan_out: true
192
+ peft_use_rslora: true
193
+ lora_target_modules:
194
+ - gate_proj
195
+ - down_proj
196
+ - up_proj
197
+ - q_proj
198
+ - v_proj
199
+ - k_proj
200
+ - o_proj
201
+
202
+
203
+ wandb_project: 12b-control
204
+ wandb_entity:
205
+ wandb_watch:
206
+ wandb_name: 12b-control-r2
207
+ wandb_log_model:
208
 
209
+ gradient_accumulation_steps: 2
210
+ micro_batch_size: 1
211
+ num_epochs: 4
212
+ optimizer: paged_adamw_8bit
213
+ lr_scheduler: cosine
214
+ learning_rate: 0.00001
215
+
216
+ train_on_inputs: false
217
+ group_by_length: false
218
+ bf16: auto
219
+ fp16:
220
+ tf32: false
221
+
222
+ gradient_checkpointing: unsloth
223
+ #gradient_checkpointing_kwargs:
224
+ # use_reentrant: false
225
+ early_stopping_patience:
226
+ resume_from_checkpoint:
227
+ local_rank:
228
+ logging_steps: 1
229
+ xformers_attention:
230
+ flash_attention: true
231
+
232
+ warmup_steps: 40
233
+ evals_per_epoch:
234
+ eval_table_size:
235
+ eval_max_new_tokens:
236
+ saves_per_epoch: 1
237
+ debug:
238
+ deepspeed: /workspace/axolotl/deepspeed_configs/zero3_bf16.json
239
+ weight_decay: 0.03
240
+ fsdp:
241
+ fsdp_config:
242
+ special_tokens:
243
+ pad_token: <pad>
244
  ```
245
+
246
+ </details><br>
247
+
248
+ ## Credits
249
+
250
+ Thank you to [Lucy Knada](https://huggingface.co/lucyknada), [Intervitens](https://huggingface.co/intervitens), [Tav](https://huggingface.co/tavtav), [Trappu](https://huggingface.co/Trappu), [Cgato](https://huggingface.co/cgato), [Kubernetes Bad](https://huggingface.co/kubernetes-bad) and the rest of [Anthracite](https://huggingface.co/anthracite-org)
251
+
252
+
253
+ ## Training
254
+ The training was done for 4 epochs. We used 4 x [RTX 3090s](https://www.nvidia.com/en-us/geforce/graphics-cards/30-series/rtx-3090-3090ti/) GPUs graciously provided by [Intervitens](https://huggingface.co/intervitens) for the fine-tuning of the model.
255
+
256
+ [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
257
+
258
+ ## Safety
259
+
260
+
261
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/66c26b6fb01b19d8c3c2467b/cgeub1ZibfEwh8-FvCbOY.png)