SeonghoonYu commited on
Commit
8f39ff9
1 Parent(s): 8693284

Upload 3 files

Browse files
laion_cc3m/checkpoints/epoch_1.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:94865986f6be13dbb84774e59ad2a4a6a1896c6e800446a20d6b151debfef531
3
+ size 7662183430
laion_cc3m/out.log ADDED
The diff for this file is too large to render. See raw diff
 
laion_cc3m/params.txt ADDED
@@ -0,0 +1,91 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ accum_freq: 1
2
+ aug_cfg: {}
3
+ batch_size: 16
4
+ beta1: 0.9
5
+ beta2: 0.98
6
+ checkpoint_path: ./logs/2023_04_03-18_34_41-model_coca_ViT-L-14-lr_1e-05-b_16-j_4-p_amp/checkpoints
7
+ coca_caption_loss_weight: 2.0
8
+ coca_contrastive_loss_weight: 1.0
9
+ copy_codebase: False
10
+ csv_caption_key: title
11
+ csv_img_key: filepath
12
+ csv_separator:
13
+ dataset_resampled: False
14
+ dataset_type: webdataset
15
+ ddp_static_graph: False
16
+ debug: False
17
+ delete_previous_checkpoint: False
18
+ device: cuda:0
19
+ dist_backend: nccl
20
+ dist_url: env://
21
+ distill: False
22
+ distill_model: None
23
+ distill_pretrained: None
24
+ distributed: True
25
+ epochs: 1
26
+ epochs_cooldown: None
27
+ eps: 1e-06
28
+ force_custom_text: False
29
+ force_image_size: None
30
+ force_patch_dropout: None
31
+ force_quick_gelu: False
32
+ gather_with_grad: False
33
+ grad_checkpointing: False
34
+ grad_clip_norm: None
35
+ horovod: False
36
+ image_mean: None
37
+ image_std: None
38
+ imagenet_v2: None
39
+ imagenet_val: None
40
+ local_loss: False
41
+ local_rank: 0
42
+ lock_image: False
43
+ lock_image_freeze_bn_stats: False
44
+ lock_image_unlocked_groups: 0
45
+ lock_text: False
46
+ lock_text_freeze_layer_norm: False
47
+ lock_text_unlocked_layers: 0
48
+ log_every_n_steps: 100
49
+ log_level: 20
50
+ log_local: False
51
+ log_path: ./logs/2023_04_03-18_34_41-model_coca_ViT-L-14-lr_1e-05-b_16-j_4-p_amp/out.log
52
+ logs: ./logs/
53
+ lr: 1e-05
54
+ lr_cooldown_end: 0.0
55
+ lr_cooldown_power: 1.0
56
+ lr_scheduler: cosine
57
+ model: coca_ViT-L-14
58
+ name: 2023_04_03-18_34_41-model_coca_ViT-L-14-lr_1e-05-b_16-j_4-p_amp
59
+ no_set_device_rank: False
60
+ precision: amp
61
+ pretrained: laion2B-s13B-b90k
62
+ pretrained_image: False
63
+ rank: 0
64
+ remote_sync: None
65
+ remote_sync_frequency: 300
66
+ remote_sync_protocol: s3
67
+ report_to:
68
+ resume: None
69
+ save_frequency: 1
70
+ save_most_recent: False
71
+ seed: 0
72
+ skip_scheduler: False
73
+ tensorboard: False
74
+ tensorboard_path:
75
+ torchscript: False
76
+ trace: False
77
+ train_data: /home/vision/seonghoon/caption_dataset/cc3m/{00000..00331}.tar
78
+ train_data_upsampling_factors: None
79
+ train_num_samples: 2339077
80
+ use_bn_sync: False
81
+ val_data: None
82
+ val_frequency: 1
83
+ val_num_samples: None
84
+ wandb: False
85
+ wandb_notes:
86
+ wandb_project_name: open-clip
87
+ warmup: 1000
88
+ wd: 0.1
89
+ workers: 4
90
+ world_size: 4
91
+ zeroshot_frequency: 2