Update README.md
Browse files
README.md
CHANGED
@@ -22,7 +22,6 @@ I`m trying to find a way to optimize the captioning of identifiers. I`ll write d
|
|
22 |
## Ningguang/εε
|
23 |
### Brief intro
|
24 |
LORA of Ningguang, with two costumes in game. civitAI page [Download](https://civitai.com/models/8546/ningguang)
|
25 |
-
|
26 |
### Training dataset
|
27 |
#### Default costume
|
28 |
72 images in total, in folder "30_Ningguang"
|
@@ -36,29 +35,45 @@ LORA of Ningguang, with two costumes in game. civitAI page [Download](https://ci
|
|
36 |
* 15 normal 360 3D model snapshots
|
37 |
* 2 nude illustrations
|
38 |
* 15 nude 360 3D model snapshots
|
39 |
-
|
40 |
### Captioning
|
41 |
WD14 captioning instead of the deepdanbooru caption was used, since the former one will not crop/resize the images. Threshold are usually set to 0.75-0.8. since I don't like to have a very long and sometimes inaccurate caption for my training data. After captionin is done, I added "ningguang \ \(genshin impact\ \)" after "1girl" to every caption file of the default costume, and "ningguang \ \(orchid's evening gown\ \) \ \(genshin impact\ \)" to the orchid costume. Some of the caption files were empty so I have to manually type the words.
|
42 |
-
|
43 |
### Training setup
|
44 |
Trained with Kohya_SS stable diffusion trainer Base model was [Anything V3.0 full](https://huggingface.co/Linaqruf/anything-v3.0/blob/main/anything-v3-fp32-pruned.safetensors) Trainig process consist of two phases. The first one with default parameters of:
|
45 |
* learning_rate: 0.0001
|
46 |
* text_encoder_lr: 5e-5
|
47 |
* unet_lr: 0.0001 and 6 epoch,
|
48 |
After phase 1, choose the one with the best result (a little bit underfitting, no over fitting, and the two costume are seperated), which is the 6th one. Then trained with 1/10 of the original LR for another 7 epochs.
|
49 |
-
|
50 |
### Result
|
51 |
![sample1](https://imagecache.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/9573a553-c456-4c36-c029-f2955fe52800/width=480)
|
52 |
![sample2](https://imagecache.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/c7709515-4537-4501-fe87-296734995700/width=480)
|
53 |
![sample3](https://imagecache.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/38d47c4a-6ba5-4925-5a56-e8701856a100/width=480)
|
54 |
![sample4](https://imagecache.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/b60aa7f4-6f63-46fb-381f-05b11f4afe00/width=480)
|
55 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
56 |
|
57 |
# Charater from Honkai impact
|
58 |
## Elysia/η±θεΈι
|
59 |
### Brief intro
|
60 |
LORA of Elysia, with 4 costumes in game. civitAI page [Download](https://civitai.com/models/14616)
|
61 |
-
|
62 |
### Training dataset
|
63 |
#### Default costume/Miss Pink Elf
|
64 |
70 images in total, in folder "14_Elysia (miss pink elf) 1girl"
|
@@ -77,16 +92,13 @@ LORA of Elysia, with 4 costumes in game. civitAI page [Download](https://civitai
|
|
77 |
* 14 normal illustrations, non-nude
|
78 |
* 60 normal 360 3D model snapshots
|
79 |
In addition, I have also included 12 images with non-official costumes in a new folder "10_Elysia 1girl"
|
80 |
-
|
81 |
### Captioning
|
82 |
WD14 captioning instead of the deepdanbooru caption was used, since the former one will not crop/resize the images. Threshold are usually set to 0.75-0.8. since I don't like to have a very long and sometimes inaccurate caption for my training data. After captionin is done, I added "elysia \ \(miss pink elf\ \) \ \(honkai impact\ \)", "elysia \ \(herrscher of human:ego\ \) \ \(honkai impact\ \)", "Elysia-maid", "Elysia-swimsuit" and "1girl, elysia \ \(honkai impact\ \)" to the captioning respectively as identifiers.
|
83 |
-
|
84 |
### Training setup
|
85 |
Trained with Kohya_SS stable diffusion trainer Base model was [Anything V3.0 full](https://huggingface.co/Linaqruf/anything-v3.0/blob/main/anything-v3-fp32-pruned.safetensors) Trainig process consist of two phases. The first one with default parameters of:
|
86 |
* learning_rate: 0.0001
|
87 |
* text_encoder_lr: 5e-5
|
88 |
* unet_lr: 0.0001 and 4 epoch,
|
89 |
After phase 1, choose the one with the best result (a little bit underfitting, no over fitting, and the two costume are seperated), which is the 6th one. Then trained with 1/10 of the original LR for another 8 epochs.
|
90 |
-
|
91 |
### Result
|
92 |
![sample1](https://imagecache.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/9a699f92-b026-4efb-9714-6d6e2675f400/width=800/174757)
|
|
|
22 |
## Ningguang/εε
|
23 |
### Brief intro
|
24 |
LORA of Ningguang, with two costumes in game. civitAI page [Download](https://civitai.com/models/8546/ningguang)
|
|
|
25 |
### Training dataset
|
26 |
#### Default costume
|
27 |
72 images in total, in folder "30_Ningguang"
|
|
|
35 |
* 15 normal 360 3D model snapshots
|
36 |
* 2 nude illustrations
|
37 |
* 15 nude 360 3D model snapshots
|
|
|
38 |
### Captioning
|
39 |
WD14 captioning instead of the deepdanbooru caption was used, since the former one will not crop/resize the images. Threshold are usually set to 0.75-0.8. since I don't like to have a very long and sometimes inaccurate caption for my training data. After captionin is done, I added "ningguang \ \(genshin impact\ \)" after "1girl" to every caption file of the default costume, and "ningguang \ \(orchid's evening gown\ \) \ \(genshin impact\ \)" to the orchid costume. Some of the caption files were empty so I have to manually type the words.
|
|
|
40 |
### Training setup
|
41 |
Trained with Kohya_SS stable diffusion trainer Base model was [Anything V3.0 full](https://huggingface.co/Linaqruf/anything-v3.0/blob/main/anything-v3-fp32-pruned.safetensors) Trainig process consist of two phases. The first one with default parameters of:
|
42 |
* learning_rate: 0.0001
|
43 |
* text_encoder_lr: 5e-5
|
44 |
* unet_lr: 0.0001 and 6 epoch,
|
45 |
After phase 1, choose the one with the best result (a little bit underfitting, no over fitting, and the two costume are seperated), which is the 6th one. Then trained with 1/10 of the original LR for another 7 epochs.
|
|
|
46 |
### Result
|
47 |
![sample1](https://imagecache.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/9573a553-c456-4c36-c029-f2955fe52800/width=480)
|
48 |
![sample2](https://imagecache.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/c7709515-4537-4501-fe87-296734995700/width=480)
|
49 |
![sample3](https://imagecache.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/38d47c4a-6ba5-4925-5a56-e8701856a100/width=480)
|
50 |
![sample4](https://imagecache.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/b60aa7f4-6f63-46fb-381f-05b11f4afe00/width=480)
|
51 |
|
52 |
+
## Barbara/θθζ
|
53 |
+
### Brief intro
|
54 |
+
LORA of Barbara, with two costumes in game.
|
55 |
+
### Training dataset
|
56 |
+
#### Default costume
|
57 |
+
164 images in total, in folder "10_barbara_(genshin_impact) 1girl"
|
58 |
+
* 104 illustrations, bothSFW and NSFW, handpicked to ensure best quality
|
59 |
+
* 30 normal 360 3D model snapshots
|
60 |
+
* 30 nude 360 3D model snapshots
|
61 |
+
#### Summertime swimsuit
|
62 |
+
94 imges in total, in folder "16_barbara_(summertime_sparkle)_(genshin_impact) 1girl"
|
63 |
+
* 64 illustrations, bothSFW and NSFW, handpicked to ensure best quality
|
64 |
+
* 30 normal 360 3D model snapshots
|
65 |
+
### Captioning
|
66 |
+
It was the first time that the standard Danbooru style prompt was used for captioning. "barbara_(genshin_impact)" and "barbara_(summertime_sparkle)_(genshin_impact)" were added to each costume respectively.
|
67 |
+
### Training setup
|
68 |
+
Defalut LR fo 4 epochs, then 1/10 default LR for another 8 epochs.
|
69 |
+
Trainig basing on anything v3.
|
70 |
+
Total steps is: (4+8)x(164x10+94x16)=37,728
|
71 |
+
### results
|
72 |
|
73 |
# Charater from Honkai impact
|
74 |
## Elysia/η±θεΈι
|
75 |
### Brief intro
|
76 |
LORA of Elysia, with 4 costumes in game. civitAI page [Download](https://civitai.com/models/14616)
|
|
|
77 |
### Training dataset
|
78 |
#### Default costume/Miss Pink Elf
|
79 |
70 images in total, in folder "14_Elysia (miss pink elf) 1girl"
|
|
|
92 |
* 14 normal illustrations, non-nude
|
93 |
* 60 normal 360 3D model snapshots
|
94 |
In addition, I have also included 12 images with non-official costumes in a new folder "10_Elysia 1girl"
|
|
|
95 |
### Captioning
|
96 |
WD14 captioning instead of the deepdanbooru caption was used, since the former one will not crop/resize the images. Threshold are usually set to 0.75-0.8. since I don't like to have a very long and sometimes inaccurate caption for my training data. After captionin is done, I added "elysia \ \(miss pink elf\ \) \ \(honkai impact\ \)", "elysia \ \(herrscher of human:ego\ \) \ \(honkai impact\ \)", "Elysia-maid", "Elysia-swimsuit" and "1girl, elysia \ \(honkai impact\ \)" to the captioning respectively as identifiers.
|
|
|
97 |
### Training setup
|
98 |
Trained with Kohya_SS stable diffusion trainer Base model was [Anything V3.0 full](https://huggingface.co/Linaqruf/anything-v3.0/blob/main/anything-v3-fp32-pruned.safetensors) Trainig process consist of two phases. The first one with default parameters of:
|
99 |
* learning_rate: 0.0001
|
100 |
* text_encoder_lr: 5e-5
|
101 |
* unet_lr: 0.0001 and 4 epoch,
|
102 |
After phase 1, choose the one with the best result (a little bit underfitting, no over fitting, and the two costume are seperated), which is the 6th one. Then trained with 1/10 of the original LR for another 8 epochs.
|
|
|
103 |
### Result
|
104 |
![sample1](https://imagecache.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/9a699f92-b026-4efb-9714-6d6e2675f400/width=800/174757)
|