Crosstyan commited on
Commit
30b3c66
1 Parent(s): 1cee318
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -22,7 +22,7 @@ Why is the Model even existing? There are loads of Stable Diffusion model out th
22
  Well, is there any models trained with resolution base resolution (`base_res`) 768 even 1024 before? Don't think so.
23
  Here it is, the BPModel, a Stable Diffusion model you may love or hate.
24
  Trained with 5k high quality images that suit my taste (not necessary yours unfortunately) from [Sankaku Complex](https://chan.sankakucomplex.com) with annotations.
25
- The dataset is public in [Crosstyan/BPDataset](https://huggingface.co/datasets/Crosstyan/BPDataset) for full disclosure.
26
  Pure combination of tags may not be the optimal way to describe the image,
27
  but I don't need to do extra work.
28
  And no, I won't feed any AI generated image
@@ -97,7 +97,7 @@ better than some artist style DreamBooth model which only train with a few
97
  hundred images or even less. I also oppose changing style by merging model since You
98
  could apply different style by training with proper captions and prompting.
99
 
100
- Besides some of images in my dataset has the artist name in the caption, however some artist name will
101
  be misinterpreted by CLIP when tokenizing. For example, *as109* will be tokenized as `[as, 1, 0, 9]` and
102
  *fuzichoco* will become `[fu, z, ic, hoco]`. Romanized Japanese suffers from the problem a lot and
103
  I don't have a good solution to fix it other than changing the artist name in the caption, which is
 
22
  Well, is there any models trained with resolution base resolution (`base_res`) 768 even 1024 before? Don't think so.
23
  Here it is, the BPModel, a Stable Diffusion model you may love or hate.
24
  Trained with 5k high quality images that suit my taste (not necessary yours unfortunately) from [Sankaku Complex](https://chan.sankakucomplex.com) with annotations.
25
+ The dataset is public in [Crosstyan/BPDataset](https://huggingface.co/datasets/Crosstyan/BPDataset) for the sake of full disclosure .
26
  Pure combination of tags may not be the optimal way to describe the image,
27
  but I don't need to do extra work.
28
  And no, I won't feed any AI generated image
 
97
  hundred images or even less. I also oppose changing style by merging model since You
98
  could apply different style by training with proper captions and prompting.
99
 
100
+ Besides some of images in my dataset have the artist name in the caption, however some artist name will
101
  be misinterpreted by CLIP when tokenizing. For example, *as109* will be tokenized as `[as, 1, 0, 9]` and
102
  *fuzichoco* will become `[fu, z, ic, hoco]`. Romanized Japanese suffers from the problem a lot and
103
  I don't have a good solution to fix it other than changing the artist name in the caption, which is