awoo
Browse filesSigned-off-by: Balazs Horvath <[email protected]>
README.md
CHANGED
@@ -32,13 +32,16 @@ The Yiff Toolkit is a comprehensive set of tools designed to enhance your creati
|
|
32 |
- [Sample Prompt File](#sample-prompt-file)
|
33 |
- [`--lowram`](#--lowram)
|
34 |
- [`--pretrained_model_name_or_path`](#--pretrained_model_name_or_path)
|
|
|
35 |
- [`--train_data_dir`](#--train_data_dir)
|
36 |
- [`--resolution`](#--resolution)
|
|
|
|
|
37 |
- [`--optimizer_type`](#--optimizer_type)
|
38 |
- [`--dataset_repeats`](#--dataset_repeats)
|
39 |
- [`--max_train_steps`](#--max_train_steps)
|
40 |
- [`--shuffle_caption`](#--shuffle_caption)
|
41 |
-
- [`--sdpa`](#--sdpa)
|
42 |
- [`--sample_prompts --sample_sampler --sample_every_n_steps`](#--sample_prompts---sample_sampler---sample_every_n_steps)
|
43 |
- [CosXL Training](#cosxl-training)
|
44 |
- [Embeddings for 1.5 and SDXL](#embeddings-for-15-and-sdxl)
|
@@ -157,6 +160,14 @@ The directory containing the checkpoint you just downloaded. I recommend closing
|
|
157 |
--pretrained_model_name_or_path="/ponydiffusers/" \
|
158 |
```
|
159 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
160 |
#### `--train_data_dir`
|
161 |
|
162 |
The directory containing the dataset. We prepared this earlier together.
|
@@ -170,7 +181,24 @@ The directory containing the dataset. We prepared this earlier together.
|
|
170 |
Always set this to match the model's resolution, which in Pony's case it is 1024x1024. If you can't fit into the VRAM, you can decrease it to `512,512` as a last resort.
|
171 |
|
172 |
```py
|
173 |
-
--resolution="
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
174 |
```
|
175 |
|
176 |
#### `--optimizer_type`
|
@@ -201,13 +229,13 @@ Specify the number of steps or epochs to train. If both `--max_train_steps` and
|
|
201 |
|
202 |
Shuffles the captions set by `--caption_separator`, it is a comma `,` by default which will work perfectly for our case since our captions look like this:
|
203 |
|
204 |
-
> rating_questionable, 5 fingers, anthro, bent over, big breasts, blue eyes, blue hair, breasts, butt, claws, curved horn, female, finger claws, fingers, fur, hair, huge breasts, looking at viewer, looking back, looking back at viewer, nipples, nude, pink body, pink hair, pink nipples, rear view, solo, tail, tail tuft, tuft, by lunarii, by x-leon-x, mythology, krystal \(darkmaster781\), dragon, scalie, wickerbeast, The image showcases a pink-scaled wickerbeast a furred dragon creature with blue eyes., She has large breasts and a thick tail., Her blue and pink horns are curved and pointy and she has a slight smiling expression on her face., Her scales are shiny and she has a blue and pink pattern on her body., Her hair is a mix of pink and blue., She is looking back at the viewer with a curious expression., She has a slight blush.,
|
205 |
|
206 |
As you can tell, I have separated the caption part not just the tags with a `,` to make sure everything gets shuffled. I'm at this point pretty certain this is beneficial especially when your caption file contains more than 77 tokens.
|
207 |
|
208 |
-
#### `--sdpa`
|
209 |
|
210 |
-
The choice between `--xformers` and `--spda` will depend on your GPU. You can benchmark it by repeating a training with
|
211 |
|
212 |
#### `--sample_prompts --sample_sampler --sample_every_n_steps`
|
213 |
|
|
|
32 |
- [Sample Prompt File](#sample-prompt-file)
|
33 |
- [`--lowram`](#--lowram)
|
34 |
- [`--pretrained_model_name_or_path`](#--pretrained_model_name_or_path)
|
35 |
+
- [`--output_dir`](#--output_dir)
|
36 |
- [`--train_data_dir`](#--train_data_dir)
|
37 |
- [`--resolution`](#--resolution)
|
38 |
+
- [`--enable_bucket`](#--enable_bucket)
|
39 |
+
- [`--min_bucket_reso` `--max_bucket_reso`](#--min_bucket_reso---max_bucket_reso)
|
40 |
- [`--optimizer_type`](#--optimizer_type)
|
41 |
- [`--dataset_repeats`](#--dataset_repeats)
|
42 |
- [`--max_train_steps`](#--max_train_steps)
|
43 |
- [`--shuffle_caption`](#--shuffle_caption)
|
44 |
+
- [`--sdpa` or `--xformers` or `--mem_eff_attn`](#--sdpa-or---xformers-or---mem_eff_attn)
|
45 |
- [`--sample_prompts --sample_sampler --sample_every_n_steps`](#--sample_prompts---sample_sampler---sample_every_n_steps)
|
46 |
- [CosXL Training](#cosxl-training)
|
47 |
- [Embeddings for 1.5 and SDXL](#embeddings-for-15-and-sdxl)
|
|
|
160 |
--pretrained_model_name_or_path="/ponydiffusers/" \
|
161 |
```
|
162 |
|
163 |
+
#### `--output_dir`
|
164 |
+
|
165 |
+
This is where all the saved epochs or steps will be saved, including the last one. If y
|
166 |
+
|
167 |
+
```py
|
168 |
+
--output_dir="/output_dir" \
|
169 |
+
```
|
170 |
+
|
171 |
#### `--train_data_dir`
|
172 |
|
173 |
The directory containing the dataset. We prepared this earlier together.
|
|
|
181 |
Always set this to match the model's resolution, which in Pony's case it is 1024x1024. If you can't fit into the VRAM, you can decrease it to `512,512` as a last resort.
|
182 |
|
183 |
```py
|
184 |
+
--resolution="1024,1024" \
|
185 |
+
```
|
186 |
+
|
187 |
+
#### `--enable_bucket`
|
188 |
+
|
189 |
+
⚠️
|
190 |
+
|
191 |
+
```py
|
192 |
+
--enable_bucket \
|
193 |
+
```
|
194 |
+
|
195 |
+
#### `--min_bucket_reso` `--max_bucket_reso`
|
196 |
+
|
197 |
+
⚠️
|
198 |
+
|
199 |
+
```py
|
200 |
+
--min_bucket_reso=256 \
|
201 |
+
--max_bucket_reso=1024 \
|
202 |
```
|
203 |
|
204 |
#### `--optimizer_type`
|
|
|
229 |
|
230 |
Shuffles the captions set by `--caption_separator`, it is a comma `,` by default which will work perfectly for our case since our captions look like this:
|
231 |
|
232 |
+
> rating_questionable, 5 fingers, anthro, bent over, big breasts, blue eyes, blue hair, breasts, butt, claws, curved horn, female, finger claws, fingers, fur, hair, huge breasts, looking at viewer, looking back, looking back at viewer, nipples, nude, pink body, pink hair, pink nipples, rear view, solo, tail, tail tuft, tuft, by lunarii, by x-leon-x, mythology, krystal \(darkmaster781\), dragon, scalie, wickerbeast, The image showcases a pink-scaled wickerbeast a furred dragon creature with blue eyes., She has large breasts and a thick tail., Her blue and pink horns are curved and pointy and she has a slight smiling expression on her face., Her scales are shiny and she has a blue and pink pattern on her body., Her hair is a mix of pink and blue., She is looking back at the viewer with a curious expression., She has a slight blush.,
|
233 |
|
234 |
As you can tell, I have separated the caption part not just the tags with a `,` to make sure everything gets shuffled. I'm at this point pretty certain this is beneficial especially when your caption file contains more than 77 tokens.
|
235 |
|
236 |
+
#### `--sdpa` or `--xformers` or `--mem_eff_attn`
|
237 |
|
238 |
+
The choice between `--xformers` or `--mem_eff_attn` and `--spda` will depend on your GPU. You can benchmark it by repeating a training with them!
|
239 |
|
240 |
#### `--sample_prompts --sample_sampler --sample_every_n_steps`
|
241 |
|