ZeroCool94 commited on
Commit
16ac9dc
1 Parent(s): c2c81cc

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +14 -13
README.md CHANGED
@@ -23,13 +23,13 @@ pipeline_tag: text-to-image
23
 
24
  <!-- Provide a quick summary of what the model is/does. -->
25
 
26
- This model is based in [Muse](https://muse-model.github.io/) and trained using the code hosted on [ZeroCool940711/muse-maskgit-pytorch](https://github.com/ZeroCool940711/muse-maskgit-pytorch), which is based on [`lucidrains/muse-maskgit-pytorch`](https://github.com/lucidrains/muse-maskgit-pytorch).
27
 
28
  # Model Details
29
- This model is a new model trained from scratch based on [Muse](https://muse-model.github.io/), trained on the [Imaginary Network Expanded Dataset](https://github.com/Sygil-Dev/INE-dataset), with the big advantage of allowing the use of multiple namespaces (labeled tags) to control various parts of the final generation.
30
  The use of namespaces (eg. “species:seal” or “studio:dc”) stops the model from misinterpreting a seal as the singer Seal, or DC Comics as Washington DC.
31
 
32
- Note: As of right now, only the first VAE and MaskGit has been trained, we still need to train the Super Resolution VAE for the model to be usable even tho we might be able to reuse the first VAE depending on the quality of it once the training progresses more.
33
 
34
  If you find my work useful, please consider supporting me on [GitHub Sponsors](https://github.com/sponsors/ZeroCool940711)!
35
 
@@ -44,11 +44,12 @@ This model is still in its infancy and it's meant to be constantly updated and t
44
  - [vae.sygil_muse_v0.1.pt](https://huggingface.co/Sygil/Sygil-Muse/blob/main/vae.sygil_muse_v0.1.pt): Trained from scratch for 3.0M steps with **dim: 128** and **vq_codebook_size: 256**.
45
  - [maskgit.sygil_muse_v0.1.pt](https://huggingface.co/Sygil/Sygil-Muse/blob/main/maskgit.sygil_muse_v0.1.pt): Maskgit trained from the VAE for 3.46M steps
46
  - [vae.sygil_muse_v0.5.pt](https://huggingface.co/Sygil/Sygil-Muse/blob/main/vae.sygil_muse_v0.5.pt): Trained from scratch for 1.99M steps with **dim: 128** and **vq_codebook_size: 8192**.
 
 
47
  - #### Beta:
48
- - [vae.263000.pt](https://huggingface.co/Sygil/Sygil-Muse/blob/main/vae.263000.pt): Trained from scratch for 263K steps and higher **vq_codebook_dim** and **vq_codebook_size** than before.
49
- - [maskgit.23739.pt](https://huggingface.co/Sygil/Sygil-Muse/blob/main/maskgit.23739.pt): Maskgit trained from the VAE for 23K steps.
50
 
51
- Note: Checkpoints under the Beta section are updated daily or at least 3-4 times a week. While the beta checkpoints can be used as they are only the latest version is kept on the repo and the older checkpoints are removed when a new one
52
  is uploaded to keep the repo clean.
53
 
54
  ## Training
@@ -60,26 +61,26 @@ The model was trained on the following dataset:
60
  **Hardware and others**
61
  - **Hardware:** 1 x Nvidia RTX 3050 GPU
62
  - **Hours Trained:** NaN.
63
- - **Gradient Accumulations**: 20
64
  - **Batch:** 1
65
  - **Learning Rate:** 1e-5
66
  - **Learning Rate Scheduler:** `constant_with_warmup`
67
- - **Scheduler Power:** 0.5
68
  - **Optimizer:** Adam
69
  - **Warmup Steps:** 10,000
70
  - **Number of Cycles:** 1
71
  - **Resolution/Image Size**: First trained at a resolution of 64x64, then increased to 256x256 and then to 512x512. Check the notes down below for more details on this.
72
- - **Dimension:** 128
73
- - **vq_codebook_dim:** 4096
74
- - **vq_codebook_size:** 16384
75
  - **heads:** 8
76
  - **depth:** 4
77
  - **Random Crop:** True
78
- - **Total Training Steps:** 263,000
79
 
80
  Note: On Muse we can change the image_size or resolution at any time without having to train the model from scratch again, this allows us to first train the model at low resolution using the same `dim` and `vq_codebook_size` to train faster and then we can increase the `image_size` and use a higher resolution once the model has trained enough.
81
 
82
- Developed by: [ZeroCool](https://github.com/ZeroCool940711) at [Sygil-Dev](https://github.com/Sygil-Dev/)
83
 
84
  # License
85
  This model is open access and available to all, with a CreativeML Open RAIL++-M License further specifying rights and usage.
 
23
 
24
  <!-- Provide a quick summary of what the model is/does. -->
25
 
26
+ This model is based in [Muse](https://muse-model.github.io/) and trained using the code hosted on [Sygil-Dev/muse-maskgit-pytorch](https://github.com/Sygil-Dev/muse-maskgit-pytorch), which is based on [`lucidrains/muse-maskgit-pytorch`](https://github.com/lucidrains/muse-maskgit-pytorch) and a collaboration between the [Sygil-Dev](https://github.com/Sygil-Dev) and [ShoukanLabs](https://github.com/ShoukanLabs) teams.
27
 
28
  # Model Details
29
+ This model is a new model trained from scratch based on [Muse](https://muse-model.github.io/), trained on a subset of the [Imaginary Network Expanded Dataset](https://github.com/Sygil-Dev/INE-dataset), with the big advantage of allowing the use of multiple namespaces (labeled tags) to control various parts of the final generation.
30
  The use of namespaces (eg. “species:seal” or “studio:dc”) stops the model from misinterpreting a seal as the singer Seal, or DC Comics as Washington DC.
31
 
32
+ Note: As of right now, only the first VAE and MaskGit has been trained with different configuration, we are trying to find the best balance between quality, performance and vram usage so Muse can be used on all kind of devices, we still need to train the Super Resolution VAE for the model to be usable even tho we might be able to reuse the first VAE depending on the quality of it once the training progresses more.
33
 
34
  If you find my work useful, please consider supporting me on [GitHub Sponsors](https://github.com/sponsors/ZeroCool940711)!
35
 
 
44
  - [vae.sygil_muse_v0.1.pt](https://huggingface.co/Sygil/Sygil-Muse/blob/main/vae.sygil_muse_v0.1.pt): Trained from scratch for 3.0M steps with **dim: 128** and **vq_codebook_size: 256**.
45
  - [maskgit.sygil_muse_v0.1.pt](https://huggingface.co/Sygil/Sygil-Muse/blob/main/maskgit.sygil_muse_v0.1.pt): Maskgit trained from the VAE for 3.46M steps
46
  - [vae.sygil_muse_v0.5.pt](https://huggingface.co/Sygil/Sygil-Muse/blob/main/vae.sygil_muse_v0.5.pt): Trained from scratch for 1.99M steps with **dim: 128** and **vq_codebook_size: 8192**.
47
+ - [vae.sygil_muse_v0.7.pt](https://huggingface.co/Sygil/Sygil-Muse/blob/main/vae.sygil_muse_v0.7.pt): Trained from scratch for 440K steps with **dim: 128**, **vq_codebook_dim 4096** and **vq_codebook_size: 16384**.
48
+ - [maskgit.sygil_muse_v0.7.pt](https://huggingface.co/Sygil/Sygil-Muse/blob/main/maskgit.sygil_muse_v0.7.pt): Maskgit trained from the **vae.sygil_muse_v0.7.pt** VAE for 108K steps.
49
  - #### Beta:
50
+ - [vae.190000.pt](https://huggingface.co/Sygil/Sygil-Muse/blob/main/vae.190000.pt): Trained from scratch for 190K steps with **dim: 32**,**vq_codebook_dim: 8192** and **vq_codebook_size: 8192**.
 
51
 
52
+ Note: Checkpoints under the Beta section are updated daily or at least 3-4 times a week. While the beta checkpoints can be used as they are, only the latest version is kept on the repo and the older checkpoints are removed when a new one
53
  is uploaded to keep the repo clean.
54
 
55
  ## Training
 
61
  **Hardware and others**
62
  - **Hardware:** 1 x Nvidia RTX 3050 GPU
63
  - **Hours Trained:** NaN.
64
+ - **Gradient Accumulations**: 10
65
  - **Batch:** 1
66
  - **Learning Rate:** 1e-5
67
  - **Learning Rate Scheduler:** `constant_with_warmup`
68
+ - **Scheduler Power:** 1.0
69
  - **Optimizer:** Adam
70
  - **Warmup Steps:** 10,000
71
  - **Number of Cycles:** 1
72
  - **Resolution/Image Size**: First trained at a resolution of 64x64, then increased to 256x256 and then to 512x512. Check the notes down below for more details on this.
73
+ - **Dimension:** 32
74
+ - **vq_codebook_dim:** 8192
75
+ - **vq_codebook_size:** 8192
76
  - **heads:** 8
77
  - **depth:** 4
78
  - **Random Crop:** True
79
+ - **Total Training Steps:** 190,000
80
 
81
  Note: On Muse we can change the image_size or resolution at any time without having to train the model from scratch again, this allows us to first train the model at low resolution using the same `dim` and `vq_codebook_size` to train faster and then we can increase the `image_size` and use a higher resolution once the model has trained enough.
82
 
83
+ Developed by: [ZeroCool](https://github.com/ZeroCool940711) at [Sygil-Dev](https://github.com/Sygil-Dev/).
84
 
85
  # License
86
  This model is open access and available to all, with a CreativeML Open RAIL++-M License further specifying rights and usage.