bghira commited on
Commit
fdaebef
·
verified ·
1 Parent(s): b9fc826

Trained for 0 epochs and 1000 steps.

Browse files

Trained with datasets ['text-embeds', 'sfwbooru-crop']
Learning rate 1.0, batch size 4, and 1 gradient accumulation steps.
Used DDPM noise scheduler for training with epsilon prediction type and rescaled_betas_zero_snr=False
Using 'trailing' timestep spacing.
Base model: stabilityai/stable-diffusion-3.5-medium
VAE: None

.gitattributes CHANGED
@@ -89,3 +89,4 @@ assets/image_6_0.png filter=lfs diff=lfs merge=lfs -text
89
  assets/image_7_0.png filter=lfs diff=lfs merge=lfs -text
90
  assets/image_8_0.png filter=lfs diff=lfs merge=lfs -text
91
  assets/image_9_0.png filter=lfs diff=lfs merge=lfs -text
 
 
89
  assets/image_7_0.png filter=lfs diff=lfs merge=lfs -text
90
  assets/image_8_0.png filter=lfs diff=lfs merge=lfs -text
91
  assets/image_9_0.png filter=lfs diff=lfs merge=lfs -text
92
+ training_state-sfwbooru-crop.json filter=lfs diff=lfs merge=lfs -text
README.md CHANGED
@@ -329,13 +329,13 @@ You may reuse the base model text encoder for inference.
329
  ## Training settings
330
 
331
  - Training epochs: 0
332
- - Training steps: 100
333
  - Learning rate: 1.0
334
  - Learning rate schedule: constant
335
  - Warmup steps: 500
336
- - Max grad norm: 0.01
337
- - Effective batch size: 6
338
- - Micro-batch size: 6
339
  - Gradient accumulation steps: 1
340
  - Number of GPUs: 1
341
  - Gradient checkpointing: True
@@ -358,15 +358,6 @@ You may reuse the base model text encoder for inference.
358
  - Crop style: random
359
  - Crop aspect: square
360
  - Used for regularisation data: No
361
- ### sfwbooru
362
- - Repeats: 0
363
- - Total number of images: 469191
364
- - Total number of aspect buckets: 70
365
- - Resolution: 1.048576 megapixels
366
- - Cropped: False
367
- - Crop style: None
368
- - Crop aspect: None
369
- - Used for regularisation data: No
370
 
371
 
372
  ## Inference
@@ -397,3 +388,11 @@ image.save("output.png", format="PNG")
397
 
398
 
399
 
 
 
 
 
 
 
 
 
 
329
  ## Training settings
330
 
331
  - Training epochs: 0
332
+ - Training steps: 1000
333
  - Learning rate: 1.0
334
  - Learning rate schedule: constant
335
  - Warmup steps: 500
336
+ - Max grad norm: 0.0
337
+ - Effective batch size: 4
338
+ - Micro-batch size: 4
339
  - Gradient accumulation steps: 1
340
  - Number of GPUs: 1
341
  - Gradient checkpointing: True
 
358
  - Crop style: random
359
  - Crop aspect: square
360
  - Used for regularisation data: No
 
 
 
 
 
 
 
 
 
361
 
362
 
363
  ## Inference
 
388
 
389
 
390
 
391
+ ## Exponential Moving Average (EMA)
392
+
393
+ SimpleTuner generates a safetensors variant of the EMA weights and a pt file.
394
+
395
+ The safetensors file is intended to be used for inference, and the pt file is for continuing finetuning.
396
+
397
+ The EMA model may provide a more well-rounded result, but typically will feel undertrained compared to the full model as it is a running decayed average of the model weights.
398
+
optimizer.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6274f068dddc60934117c2fe0fdc19194d7c7ef632ebb79d6656a6bcc0954f50
3
+ size 5313950865
random_states_0.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c3396fb45d05f866cede6ac17b1016a0ea31e7c16413cb4924c4312c77f29a03
3
+ size 14344
training_state-sfwbooru-crop.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:82435ef38f168d82561ef3ed42940e3cfb7233af3dd51d6d2ff06ef22b5c05f0
3
+ size 14740962
training_state.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"global_step": 1000, "epoch_step": 1000, "epoch": 1, "exhausted_backends": [], "repeats": {}}
transformer/diffusion_pytorch_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6f58e2d4094eaf093c27eed83243547d8d01510fcac23e0a80d88049a31f7a1a
3
  size 4939433672
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f14fdae48f7932fd7481a2373d8507ce9d451295dcd044d2ca98188975c1baf8
3
  size 4939433672
transformer_ema/config.json ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "SD3Transformer2DModel",
3
+ "_diffusers_version": "0.32.0.dev0",
4
+ "_name_or_path": "stabilityai/stable-diffusion-3.5-medium",
5
+ "attention_head_dim": 64,
6
+ "caption_projection_dim": 1536,
7
+ "decay": 0.995,
8
+ "dual_attention_layers": [
9
+ 0,
10
+ 1,
11
+ 2,
12
+ 3,
13
+ 4,
14
+ 5,
15
+ 6,
16
+ 7,
17
+ 8,
18
+ 9,
19
+ 10,
20
+ 11,
21
+ 12
22
+ ],
23
+ "in_channels": 16,
24
+ "inv_gamma": 1.0,
25
+ "joint_attention_dim": 4096,
26
+ "min_decay": 0.0,
27
+ "num_attention_heads": 24,
28
+ "num_layers": 24,
29
+ "optimization_step": 1000,
30
+ "out_channels": 16,
31
+ "patch_size": 2,
32
+ "pooled_projection_dim": 2048,
33
+ "pos_embed_max_size": 384,
34
+ "power": 0.6666666666666666,
35
+ "qk_norm": "rms_norm",
36
+ "sample_size": 128,
37
+ "update_after_step": 0,
38
+ "use_ema_warmup": false
39
+ }
transformer_ema/diffusion_pytorch_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:94a7ec3026986d8026933cea83042ad4959ffde3cb2da707557b65f9747ff43a
3
+ size 9878760864
transformer_ema/ema_model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cd018af71196ba782d63cca6bb9fd86be0fd14ba2ca9d2e37aefa5f3bb5fbf12
3
+ size 4486612394