okingjo commited on
Commit
295e219
β€’
1 Parent(s): a1babe2

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +20 -8
README.md CHANGED
@@ -22,7 +22,6 @@ I`m trying to find a way to optimize the captioning of identifiers. I`ll write d
22
  ## Ningguang/凝光
23
  ### Brief intro
24
  LORA of Ningguang, with two costumes in game. civitAI page [Download](https://civitai.com/models/8546/ningguang)
25
-
26
  ### Training dataset
27
  #### Default costume
28
  72 images in total, in folder "30_Ningguang"
@@ -36,29 +35,45 @@ LORA of Ningguang, with two costumes in game. civitAI page [Download](https://ci
36
  * 15 normal 360 3D model snapshots
37
  * 2 nude illustrations
38
  * 15 nude 360 3D model snapshots
39
-
40
  ### Captioning
41
  WD14 captioning instead of the deepdanbooru caption was used, since the former one will not crop/resize the images. Threshold are usually set to 0.75-0.8. since I don't like to have a very long and sometimes inaccurate caption for my training data. After captionin is done, I added "ningguang \ \(genshin impact\ \)" after "1girl" to every caption file of the default costume, and "ningguang \ \(orchid's evening gown\ \) \ \(genshin impact\ \)" to the orchid costume. Some of the caption files were empty so I have to manually type the words.
42
-
43
  ### Training setup
44
  Trained with Kohya_SS stable diffusion trainer Base model was [Anything V3.0 full](https://huggingface.co/Linaqruf/anything-v3.0/blob/main/anything-v3-fp32-pruned.safetensors) Trainig process consist of two phases. The first one with default parameters of:
45
  * learning_rate: 0.0001
46
  * text_encoder_lr: 5e-5
47
  * unet_lr: 0.0001 and 6 epoch,
48
  After phase 1, choose the one with the best result (a little bit underfitting, no over fitting, and the two costume are seperated), which is the 6th one. Then trained with 1/10 of the original LR for another 7 epochs.
49
-
50
  ### Result
51
  ![sample1](https://imagecache.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/9573a553-c456-4c36-c029-f2955fe52800/width=480)
52
  ![sample2](https://imagecache.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/c7709515-4537-4501-fe87-296734995700/width=480)
53
  ![sample3](https://imagecache.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/38d47c4a-6ba5-4925-5a56-e8701856a100/width=480)
54
  ![sample4](https://imagecache.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/b60aa7f4-6f63-46fb-381f-05b11f4afe00/width=480)
55
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
56
 
57
  # Charater from Honkai impact
58
  ## Elysia/ηˆ±θŽ‰εΈŒι›…
59
  ### Brief intro
60
  LORA of Elysia, with 4 costumes in game. civitAI page [Download](https://civitai.com/models/14616)
61
-
62
  ### Training dataset
63
  #### Default costume/Miss Pink Elf
64
  70 images in total, in folder "14_Elysia (miss pink elf) 1girl"
@@ -77,16 +92,13 @@ LORA of Elysia, with 4 costumes in game. civitAI page [Download](https://civitai
77
  * 14 normal illustrations, non-nude
78
  * 60 normal 360 3D model snapshots
79
  In addition, I have also included 12 images with non-official costumes in a new folder "10_Elysia 1girl"
80
-
81
  ### Captioning
82
  WD14 captioning instead of the deepdanbooru caption was used, since the former one will not crop/resize the images. Threshold are usually set to 0.75-0.8. since I don't like to have a very long and sometimes inaccurate caption for my training data. After captionin is done, I added "elysia \ \(miss pink elf\ \) \ \(honkai impact\ \)", "elysia \ \(herrscher of human:ego\ \) \ \(honkai impact\ \)", "Elysia-maid", "Elysia-swimsuit" and "1girl, elysia \ \(honkai impact\ \)" to the captioning respectively as identifiers.
83
-
84
  ### Training setup
85
  Trained with Kohya_SS stable diffusion trainer Base model was [Anything V3.0 full](https://huggingface.co/Linaqruf/anything-v3.0/blob/main/anything-v3-fp32-pruned.safetensors) Trainig process consist of two phases. The first one with default parameters of:
86
  * learning_rate: 0.0001
87
  * text_encoder_lr: 5e-5
88
  * unet_lr: 0.0001 and 4 epoch,
89
  After phase 1, choose the one with the best result (a little bit underfitting, no over fitting, and the two costume are seperated), which is the 6th one. Then trained with 1/10 of the original LR for another 8 epochs.
90
-
91
  ### Result
92
  ![sample1](https://imagecache.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/9a699f92-b026-4efb-9714-6d6e2675f400/width=800/174757)
 
22
  ## Ningguang/凝光
23
  ### Brief intro
24
  LORA of Ningguang, with two costumes in game. civitAI page [Download](https://civitai.com/models/8546/ningguang)
 
25
  ### Training dataset
26
  #### Default costume
27
  72 images in total, in folder "30_Ningguang"
 
35
  * 15 normal 360 3D model snapshots
36
  * 2 nude illustrations
37
  * 15 nude 360 3D model snapshots
 
38
  ### Captioning
39
  WD14 captioning instead of the deepdanbooru caption was used, since the former one will not crop/resize the images. Threshold are usually set to 0.75-0.8. since I don't like to have a very long and sometimes inaccurate caption for my training data. After captionin is done, I added "ningguang \ \(genshin impact\ \)" after "1girl" to every caption file of the default costume, and "ningguang \ \(orchid's evening gown\ \) \ \(genshin impact\ \)" to the orchid costume. Some of the caption files were empty so I have to manually type the words.
 
40
  ### Training setup
41
  Trained with Kohya_SS stable diffusion trainer Base model was [Anything V3.0 full](https://huggingface.co/Linaqruf/anything-v3.0/blob/main/anything-v3-fp32-pruned.safetensors) Trainig process consist of two phases. The first one with default parameters of:
42
  * learning_rate: 0.0001
43
  * text_encoder_lr: 5e-5
44
  * unet_lr: 0.0001 and 6 epoch,
45
  After phase 1, choose the one with the best result (a little bit underfitting, no over fitting, and the two costume are seperated), which is the 6th one. Then trained with 1/10 of the original LR for another 7 epochs.
 
46
  ### Result
47
  ![sample1](https://imagecache.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/9573a553-c456-4c36-c029-f2955fe52800/width=480)
48
  ![sample2](https://imagecache.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/c7709515-4537-4501-fe87-296734995700/width=480)
49
  ![sample3](https://imagecache.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/38d47c4a-6ba5-4925-5a56-e8701856a100/width=480)
50
  ![sample4](https://imagecache.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/b60aa7f4-6f63-46fb-381f-05b11f4afe00/width=480)
51
 
52
+ ## Barbara/θŠ­θŠ­ζ‹‰
53
+ ### Brief intro
54
+ LORA of Barbara, with two costumes in game.
55
+ ### Training dataset
56
+ #### Default costume
57
+ 164 images in total, in folder "10_barbara_(genshin_impact) 1girl"
58
+ * 104 illustrations, bothSFW and NSFW, handpicked to ensure best quality
59
+ * 30 normal 360 3D model snapshots
60
+ * 30 nude 360 3D model snapshots
61
+ #### Summertime swimsuit
62
+ 94 imges in total, in folder "16_barbara_(summertime_sparkle)_(genshin_impact) 1girl"
63
+ * 64 illustrations, bothSFW and NSFW, handpicked to ensure best quality
64
+ * 30 normal 360 3D model snapshots
65
+ ### Captioning
66
+ It was the first time that the standard Danbooru style prompt was used for captioning. "barbara_(genshin_impact)" and "barbara_(summertime_sparkle)_(genshin_impact)" were added to each costume respectively.
67
+ ### Training setup
68
+ Defalut LR fo 4 epochs, then 1/10 default LR for another 8 epochs.
69
+ Trainig basing on anything v3.
70
+ Total steps is: (4+8)x(164x10+94x16)=37,728
71
+ ### results
72
 
73
  # Charater from Honkai impact
74
  ## Elysia/ηˆ±θŽ‰εΈŒι›…
75
  ### Brief intro
76
  LORA of Elysia, with 4 costumes in game. civitAI page [Download](https://civitai.com/models/14616)
 
77
  ### Training dataset
78
  #### Default costume/Miss Pink Elf
79
  70 images in total, in folder "14_Elysia (miss pink elf) 1girl"
 
92
  * 14 normal illustrations, non-nude
93
  * 60 normal 360 3D model snapshots
94
  In addition, I have also included 12 images with non-official costumes in a new folder "10_Elysia 1girl"
 
95
  ### Captioning
96
  WD14 captioning instead of the deepdanbooru caption was used, since the former one will not crop/resize the images. Threshold are usually set to 0.75-0.8. since I don't like to have a very long and sometimes inaccurate caption for my training data. After captionin is done, I added "elysia \ \(miss pink elf\ \) \ \(honkai impact\ \)", "elysia \ \(herrscher of human:ego\ \) \ \(honkai impact\ \)", "Elysia-maid", "Elysia-swimsuit" and "1girl, elysia \ \(honkai impact\ \)" to the captioning respectively as identifiers.
 
97
  ### Training setup
98
  Trained with Kohya_SS stable diffusion trainer Base model was [Anything V3.0 full](https://huggingface.co/Linaqruf/anything-v3.0/blob/main/anything-v3-fp32-pruned.safetensors) Trainig process consist of two phases. The first one with default parameters of:
99
  * learning_rate: 0.0001
100
  * text_encoder_lr: 5e-5
101
  * unet_lr: 0.0001 and 4 epoch,
102
  After phase 1, choose the one with the best result (a little bit underfitting, no over fitting, and the two costume are seperated), which is the 6th one. Then trained with 1/10 of the original LR for another 8 epochs.
 
103
  ### Result
104
  ![sample1](https://imagecache.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/9a699f92-b026-4efb-9714-6d6e2675f400/width=800/174757)