Upload decoder/decoder_config.json with huggingface_hub
Browse files
decoder/decoder_config.json
CHANGED
@@ -33,8 +33,8 @@
|
|
33 |
},
|
34 |
"data": {
|
35 |
"webdataset_base_url": "pipe:aws s3 cp --quiet s3://s-datasets/laion5b/laion2B-data/{}.tar -",
|
36 |
-
"num_workers":
|
37 |
-
"batch_size":
|
38 |
"start_shard": 0,
|
39 |
"end_shard": 231349,
|
40 |
"shard_width": 6,
|
@@ -61,6 +61,7 @@
|
|
61 |
"wd": 0.01,
|
62 |
"max_grad_norm": 0.5,
|
63 |
"save_every_n_samples": 5000000,
|
|
|
64 |
"n_sample_images": 10,
|
65 |
"device": "cuda:0",
|
66 |
"epoch_samples": 10000000,
|
@@ -85,11 +86,9 @@
|
|
85 |
|
86 |
"log": {
|
87 |
"log_type": "wandb",
|
88 |
-
|
89 |
"wandb_entity": "nousr_laion",
|
90 |
"wandb_project": "h_14_decoder",
|
91 |
"wandb_resume": false,
|
92 |
-
|
93 |
"auto_resume": true,
|
94 |
"verbose": true
|
95 |
},
|
|
|
33 |
},
|
34 |
"data": {
|
35 |
"webdataset_base_url": "pipe:aws s3 cp --quiet s3://s-datasets/laion5b/laion2B-data/{}.tar -",
|
36 |
+
"num_workers": 6,
|
37 |
+
"batch_size": 25,
|
38 |
"start_shard": 0,
|
39 |
"end_shard": 231349,
|
40 |
"shard_width": 6,
|
|
|
61 |
"wd": 0.01,
|
62 |
"max_grad_norm": 0.5,
|
63 |
"save_every_n_samples": 5000000,
|
64 |
+
"find_unused_parameters": false,
|
65 |
"n_sample_images": 10,
|
66 |
"device": "cuda:0",
|
67 |
"epoch_samples": 10000000,
|
|
|
86 |
|
87 |
"log": {
|
88 |
"log_type": "wandb",
|
|
|
89 |
"wandb_entity": "nousr_laion",
|
90 |
"wandb_project": "h_14_decoder",
|
91 |
"wandb_resume": false,
|
|
|
92 |
"auto_resume": true,
|
93 |
"verbose": true
|
94 |
},
|