tobiaslee commited on
Commit
afbc238
1 Parent(s): 8ec44e9

update stage1

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. stage1_qformer2_128/checkpoints/latest-checkpoint.pt +3 -0
  2. stage1_qformer2_128/checkpoints/step-002180-epoch-00-loss=2.4867.pt +3 -0
  3. stage1_qformer2_128/config.json +60 -0
  4. stage1_qformer2_128/config.yaml +53 -0
  5. stage1_qformer2_128/run-metrics.jsonl +1 -0
  6. stage1_qformer2_128/stage1_qformer2_128.jsonl +0 -0
  7. stage1_qformer2_128/wandb/debug-internal.log +1 -0
  8. stage1_qformer2_128/wandb/debug.log +1 -0
  9. stage1_qformer2_128/wandb/latest-run +1 -0
  10. stage1_qformer2_128/wandb/run-20240313_070332-tkilnh6o/files/conda-environment.yaml +194 -0
  11. stage1_qformer2_128/wandb/run-20240313_070332-tkilnh6o/files/config.yaml +119 -0
  12. stage1_qformer2_128/wandb/run-20240313_070332-tkilnh6o/files/output.log +1745 -0
  13. stage1_qformer2_128/wandb/run-20240313_070332-tkilnh6o/files/requirements.txt +122 -0
  14. stage1_qformer2_128/wandb/run-20240313_070332-tkilnh6o/files/wandb-metadata.json +1362 -0
  15. stage1_qformer2_128/wandb/run-20240313_070332-tkilnh6o/files/wandb-summary.json +1 -0
  16. stage1_qformer2_128/wandb/run-20240313_070332-tkilnh6o/logs/debug-internal.log +0 -0
  17. stage1_qformer2_128/wandb/run-20240313_070332-tkilnh6o/logs/debug.log +34 -0
  18. stage1_qformer2_128/wandb/run-20240313_070332-tkilnh6o/run-tkilnh6o.wandb +3 -0
  19. stage1_qformer2_256/checkpoints/latest-checkpoint.pt +3 -0
  20. stage1_qformer2_256/checkpoints/step-002180-epoch-00-loss=2.5032.pt +3 -0
  21. stage1_qformer2_256/config.json +60 -0
  22. stage1_qformer2_256/config.yaml +53 -0
  23. stage1_qformer2_256/run-metrics.jsonl +1 -0
  24. stage1_qformer2_256/stage1_qformer2_256.jsonl +0 -0
  25. stage1_qformer2_256/wandb/debug-internal.log +1 -0
  26. stage1_qformer2_256/wandb/debug.log +1 -0
  27. stage1_qformer2_256/wandb/latest-run +1 -0
  28. stage1_qformer2_256/wandb/run-20240313_070329-pog6cg3f/files/conda-environment.yaml +194 -0
  29. stage1_qformer2_256/wandb/run-20240313_070329-pog6cg3f/files/config.yaml +119 -0
  30. stage1_qformer2_256/wandb/run-20240313_070329-pog6cg3f/files/output.log +2180 -0
  31. stage1_qformer2_256/wandb/run-20240313_070329-pog6cg3f/files/requirements.txt +122 -0
  32. stage1_qformer2_256/wandb/run-20240313_070329-pog6cg3f/files/wandb-metadata.json +1362 -0
  33. stage1_qformer2_256/wandb/run-20240313_070329-pog6cg3f/files/wandb-summary.json +1 -0
  34. stage1_qformer2_256/wandb/run-20240313_070329-pog6cg3f/logs/debug-internal.log +0 -0
  35. stage1_qformer2_256/wandb/run-20240313_070329-pog6cg3f/logs/debug.log +34 -0
  36. stage1_qformer2_256/wandb/run-20240313_070329-pog6cg3f/run-pog6cg3f.wandb +3 -0
  37. stage1_qformer2_32/checkpoints/latest-checkpoint.pt +3 -0
  38. stage1_qformer2_32/checkpoints/step-002180-epoch-00-loss=2.4617.pt +3 -0
  39. stage1_qformer2_32/config.json +60 -0
  40. stage1_qformer2_32/config.yaml +53 -0
  41. stage1_qformer2_32/run-metrics.jsonl +1 -0
  42. stage1_qformer2_32/stage1_qformer2_32.jsonl +0 -0
  43. stage1_qformer2_32/wandb/debug-internal.log +1 -0
  44. stage1_qformer2_32/wandb/debug.log +1 -0
  45. stage1_qformer2_32/wandb/latest-run +1 -0
  46. stage1_qformer2_32/wandb/run-20240313_070204-8nxg74lb/files/conda-environment.yaml +194 -0
  47. stage1_qformer2_32/wandb/run-20240313_070204-8nxg74lb/files/config.yaml +119 -0
  48. stage1_qformer2_32/wandb/run-20240313_070204-8nxg74lb/files/output.log +1031 -0
  49. stage1_qformer2_32/wandb/run-20240313_070204-8nxg74lb/files/requirements.txt +122 -0
  50. stage1_qformer2_32/wandb/run-20240313_070204-8nxg74lb/files/wandb-metadata.json +1362 -0
stage1_qformer2_128/checkpoints/latest-checkpoint.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7d2abe3f487797cc2e97eb3ec0345d81bbeb9ba5478e0a3f6a6829413bf10af9
3
+ size 1866539184
stage1_qformer2_128/checkpoints/step-002180-epoch-00-loss=2.4867.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7d2abe3f487797cc2e97eb3ec0345d81bbeb9ba5478e0a3f6a6829413bf10af9
3
+ size 1866539184
stage1_qformer2_128/config.json ADDED
@@ -0,0 +1,60 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset": {
3
+ "align_stage_components": [
4
+ "download/llava-laion-cc-sbu-558k/chat.json",
5
+ "download/llava-laion-cc-sbu-558k"
6
+ ],
7
+ "dataset_id": "llava-v15",
8
+ "dataset_root_dir": "data",
9
+ "finetune_stage_components": [
10
+ "download/llava-v1.5-instruct/llava_v1_5_mix665k.json",
11
+ "download/llava-v1.5-instruct"
12
+ ],
13
+ "type": "llava-v15"
14
+ },
15
+ "hf_token": ".hf_token",
16
+ "max_length": 4096,
17
+ "model": {
18
+ "align_epochs": 1,
19
+ "align_global_batch_size": 256,
20
+ "align_learning_rate": 0.001,
21
+ "align_lr_scheduler_type": "linear-warmup+cosine-decay",
22
+ "align_max_grad_norm": 1.0,
23
+ "align_max_steps": null,
24
+ "align_per_device_batch_size": 16,
25
+ "align_train_strategy": "fsdp-shard-grad-op",
26
+ "align_warmup_ratio": 0.03,
27
+ "align_weight_decay": 0.0,
28
+ "arch_specifier": "qformer2_128",
29
+ "enable_gradient_checkpointing": true,
30
+ "enable_mixed_precision_training": true,
31
+ "finetune_epochs": 1,
32
+ "finetune_global_batch_size": 128,
33
+ "finetune_learning_rate": 2e-05,
34
+ "finetune_lr_scheduler_type": "linear-warmup+cosine-decay",
35
+ "finetune_max_grad_norm": 1.0,
36
+ "finetune_max_steps": null,
37
+ "finetune_per_device_batch_size": 16,
38
+ "finetune_train_strategy": "fsdp-full-shard",
39
+ "finetune_warmup_ratio": 0.03,
40
+ "finetune_weight_decay": 0.1,
41
+ "image_resize_strategy": "letterbox",
42
+ "llm_backbone_id": "vicuna-v15-7b",
43
+ "llm_max_length": 2048,
44
+ "model_id": "one-stage+7b_qformer2_128",
45
+ "reduce_in_full_precision": false,
46
+ "type": "one-stage+7b",
47
+ "vision_backbone_id": "clip-vit-l-336px"
48
+ },
49
+ "pretrained_checkpoint": null,
50
+ "run_id": "stage1_qformer2_128",
51
+ "run_root_dir": "runs",
52
+ "seed": 7,
53
+ "stage": "align",
54
+ "trackers": [
55
+ "jsonl",
56
+ "wandb"
57
+ ],
58
+ "wandb_entity": "lilei_stones",
59
+ "wandb_project": "hf-vlms"
60
+ }
stage1_qformer2_128/config.yaml ADDED
@@ -0,0 +1,53 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ dataset:
2
+ align_stage_components:
3
+ - download/llava-laion-cc-sbu-558k/chat.json
4
+ - download/llava-laion-cc-sbu-558k
5
+ dataset_id: llava-v15
6
+ dataset_root_dir: data
7
+ finetune_stage_components:
8
+ - download/llava-v1.5-instruct/llava_v1_5_mix665k.json
9
+ - download/llava-v1.5-instruct
10
+ type: llava-v15
11
+ hf_token: .hf_token
12
+ max_length: 4096
13
+ model:
14
+ align_epochs: 1
15
+ align_global_batch_size: 256
16
+ align_learning_rate: 0.001
17
+ align_lr_scheduler_type: linear-warmup+cosine-decay
18
+ align_max_grad_norm: 1.0
19
+ align_max_steps: null
20
+ align_per_device_batch_size: 16
21
+ align_train_strategy: fsdp-shard-grad-op
22
+ align_warmup_ratio: 0.03
23
+ align_weight_decay: 0.0
24
+ arch_specifier: qformer2_128
25
+ enable_gradient_checkpointing: true
26
+ enable_mixed_precision_training: true
27
+ finetune_epochs: 1
28
+ finetune_global_batch_size: 128
29
+ finetune_learning_rate: 2.0e-05
30
+ finetune_lr_scheduler_type: linear-warmup+cosine-decay
31
+ finetune_max_grad_norm: 1.0
32
+ finetune_max_steps: null
33
+ finetune_per_device_batch_size: 16
34
+ finetune_train_strategy: fsdp-full-shard
35
+ finetune_warmup_ratio: 0.03
36
+ finetune_weight_decay: 0.1
37
+ image_resize_strategy: letterbox
38
+ llm_backbone_id: vicuna-v15-7b
39
+ llm_max_length: 2048
40
+ model_id: one-stage+7b_qformer2_128
41
+ reduce_in_full_precision: false
42
+ type: one-stage+7b
43
+ vision_backbone_id: clip-vit-l-336px
44
+ pretrained_checkpoint: null
45
+ run_id: stage1_qformer2_128
46
+ run_root_dir: runs
47
+ seed: 7
48
+ stage: align
49
+ trackers:
50
+ - jsonl
51
+ - wandb
52
+ wandb_entity: lilei_stones
53
+ wandb_project: hf-vlms
stage1_qformer2_128/run-metrics.jsonl ADDED
@@ -0,0 +1 @@
 
 
1
+ {"hparams": {"dataset": {"align_stage_components": ["download/llava-laion-cc-sbu-558k/chat.json", "download/llava-laion-cc-sbu-558k"], "dataset_id": "llava-v15", "dataset_root_dir": "data", "finetune_stage_components": ["download/llava-v1.5-instruct/llava_v1_5_mix665k.json", "download/llava-v1.5-instruct"], "type": "llava-v15"}, "hf_token": ".hf_token", "max_length": 4096, "model": {"align_epochs": 1, "align_global_batch_size": 256, "align_learning_rate": 0.001, "align_lr_scheduler_type": "linear-warmup+cosine-decay", "align_max_grad_norm": 1.0, "align_max_steps": null, "align_per_device_batch_size": 16, "align_train_strategy": "fsdp-shard-grad-op", "align_warmup_ratio": 0.03, "align_weight_decay": 0.0, "arch_specifier": "qformer2_128", "enable_gradient_checkpointing": true, "enable_mixed_precision_training": true, "finetune_epochs": 1, "finetune_global_batch_size": 128, "finetune_learning_rate": 2e-05, "finetune_lr_scheduler_type": "linear-warmup+cosine-decay", "finetune_max_grad_norm": 1.0, "finetune_max_steps": null, "finetune_per_device_batch_size": 16, "finetune_train_strategy": "fsdp-full-shard", "finetune_warmup_ratio": 0.03, "finetune_weight_decay": 0.1, "image_resize_strategy": "letterbox", "llm_backbone_id": "vicuna-v15-7b", "llm_max_length": 2048, "model_id": "one-stage+7b_qformer2_128", "reduce_in_full_precision": false, "type": "one-stage+7b", "vision_backbone_id": "clip-vit-l-336px"}, "pretrained_checkpoint": null, "run_id": "stage1_qformer2_128", "run_root_dir": "runs", "seed": 7, "stage": "align", "trackers": ["jsonl", "wandb"], "wandb_entity": "lilei_stones", "wandb_project": "hf-vlms"}, "run_id": "stage1_qformer2_128"}
stage1_qformer2_128/stage1_qformer2_128.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
stage1_qformer2_128/wandb/debug-internal.log ADDED
@@ -0,0 +1 @@
 
 
1
+ run-20240313_070332-tkilnh6o/logs/debug-internal.log
stage1_qformer2_128/wandb/debug.log ADDED
@@ -0,0 +1 @@
 
 
1
+ run-20240313_070332-tkilnh6o/logs/debug.log
stage1_qformer2_128/wandb/latest-run ADDED
@@ -0,0 +1 @@
 
 
1
+ run-20240313_070332-tkilnh6o
stage1_qformer2_128/wandb/run-20240313_070332-tkilnh6o/files/conda-environment.yaml ADDED
@@ -0,0 +1,194 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ name: pvlm
2
+ channels:
3
+ - pytorch
4
+ - nvidia
5
+ - defaults
6
+ dependencies:
7
+ - _libgcc_mutex=0.1=main
8
+ - _openmp_mutex=5.1=1_gnu
9
+ - blas=1.0=mkl
10
+ - bzip2=1.0.8=h7b6447c_0
11
+ - ca-certificates=2023.12.12=h06a4308_0
12
+ - certifi=2024.2.2=py310h06a4308_0
13
+ - charset-normalizer=2.0.4=pyhd3eb1b0_0
14
+ - cuda-cudart=11.8.89=0
15
+ - cuda-cupti=11.8.87=0
16
+ - cuda-libraries=11.8.0=0
17
+ - cuda-nvrtc=11.8.89=0
18
+ - cuda-nvtx=11.8.86=0
19
+ - cuda-runtime=11.8.0=0
20
+ - ffmpeg=4.3=hf484d3e_0
21
+ - filelock=3.13.1=py310h06a4308_0
22
+ - freetype=2.12.1=h4a9f257_0
23
+ - gmp=6.2.1=h295c915_3
24
+ - gmpy2=2.1.2=py310heeb90bb_0
25
+ - gnutls=3.6.15=he1e5248_0
26
+ - idna=3.4=py310h06a4308_0
27
+ - intel-openmp=2023.1.0=hdb19cb5_46306
28
+ - jinja2=3.1.3=py310h06a4308_0
29
+ - jpeg=9e=h5eee18b_1
30
+ - lame=3.100=h7b6447c_0
31
+ - lcms2=2.12=h3be6417_0
32
+ - ld_impl_linux-64=2.38=h1181459_1
33
+ - lerc=3.0=h295c915_0
34
+ - libcublas=11.11.3.6=0
35
+ - libcufft=10.9.0.58=0
36
+ - libcufile=1.8.1.2=0
37
+ - libcurand=10.3.4.107=0
38
+ - libcusolver=11.4.1.48=0
39
+ - libcusparse=11.7.5.86=0
40
+ - libdeflate=1.17=h5eee18b_1
41
+ - libffi=3.4.4=h6a678d5_0
42
+ - libgcc-ng=11.2.0=h1234567_1
43
+ - libgomp=11.2.0=h1234567_1
44
+ - libiconv=1.16=h7f8727e_2
45
+ - libidn2=2.3.4=h5eee18b_0
46
+ - libjpeg-turbo=2.0.0=h9bf148f_0
47
+ - libnpp=11.8.0.86=0
48
+ - libnvjpeg=11.9.0.86=0
49
+ - libpng=1.6.39=h5eee18b_0
50
+ - libstdcxx-ng=11.2.0=h1234567_1
51
+ - libtasn1=4.19.0=h5eee18b_0
52
+ - libtiff=4.5.1=h6a678d5_0
53
+ - libunistring=0.9.10=h27cfd23_0
54
+ - libuuid=1.41.5=h5eee18b_0
55
+ - libwebp-base=1.3.2=h5eee18b_0
56
+ - llvm-openmp=14.0.6=h9e868ea_0
57
+ - lz4-c=1.9.4=h6a678d5_0
58
+ - markupsafe=2.1.3=py310h5eee18b_0
59
+ - mkl=2023.1.0=h213fc3f_46344
60
+ - mkl-service=2.4.0=py310h5eee18b_1
61
+ - mkl_fft=1.3.8=py310h5eee18b_0
62
+ - mkl_random=1.2.4=py310hdb19cb5_0
63
+ - mpc=1.1.0=h10f8cd9_1
64
+ - mpfr=4.0.2=hb69a4c5_1
65
+ - mpmath=1.3.0=py310h06a4308_0
66
+ - ncurses=6.4=h6a678d5_0
67
+ - nettle=3.7.3=hbbd107a_1
68
+ - networkx=3.1=py310h06a4308_0
69
+ - numpy=1.26.4=py310h5f9d8c6_0
70
+ - numpy-base=1.26.4=py310hb5e798b_0
71
+ - openh264=2.1.1=h4ff587b_0
72
+ - openjpeg=2.4.0=h3ad879b_0
73
+ - openssl=3.0.13=h7f8727e_0
74
+ - pillow=10.2.0=py310h5eee18b_0
75
+ - pip=23.3.1=py310h06a4308_0
76
+ - python=3.10.13=h955ad1f_0
77
+ - pytorch=2.2.1=py3.10_cuda11.8_cudnn8.7.0_0
78
+ - pytorch-cuda=11.8=h7e8668a_5
79
+ - pytorch-mutex=1.0=cuda
80
+ - pyyaml=6.0.1=py310h5eee18b_0
81
+ - readline=8.2=h5eee18b_0
82
+ - requests=2.31.0=py310h06a4308_1
83
+ - setuptools=68.2.2=py310h06a4308_0
84
+ - sqlite=3.41.2=h5eee18b_0
85
+ - sympy=1.12=py310h06a4308_0
86
+ - tbb=2021.8.0=hdb19cb5_0
87
+ - tk=8.6.12=h1ccaba5_0
88
+ - torchaudio=2.2.1=py310_cu118
89
+ - torchtriton=2.2.0=py310
90
+ - torchvision=0.17.1=py310_cu118
91
+ - typing_extensions=4.9.0=py310h06a4308_1
92
+ - urllib3=2.1.0=py310h06a4308_0
93
+ - wheel=0.41.2=py310h06a4308_0
94
+ - xz=5.4.5=h5eee18b_0
95
+ - yaml=0.2.5=h7b6447c_0
96
+ - zlib=1.2.13=h5eee18b_0
97
+ - zstd=1.5.5=hc292b87_0
98
+ - pip:
99
+ - accelerate==0.27.2
100
+ - aiohttp==3.9.3
101
+ - aiosignal==1.3.1
102
+ - annotated-types==0.6.0
103
+ - appdirs==1.4.4
104
+ - async-timeout==4.0.3
105
+ - attrs==23.2.0
106
+ - beautifulsoup4==4.12.3
107
+ - blis==0.7.11
108
+ - catalogue==2.0.10
109
+ - click==8.1.7
110
+ - cloudpathlib==0.16.0
111
+ - confection==0.1.4
112
+ - contourpy==1.2.0
113
+ - cycler==0.12.1
114
+ - cymem==2.0.8
115
+ - datasets==2.18.0
116
+ - deepspeed==0.13.4
117
+ - dill==0.3.8
118
+ - docker-pycreds==0.4.0
119
+ - draccus==0.7.1
120
+ - einops==0.7.0
121
+ - flash-attn==2.5.5
122
+ - fonttools==4.49.0
123
+ - frozenlist==1.4.1
124
+ - fsspec==2024.2.0
125
+ - gdown==5.1.0
126
+ - gitdb==4.0.11
127
+ - gitpython==3.1.42
128
+ - hjson==3.1.0
129
+ - huggingface-hub==0.20.3
130
+ - install==1.3.5
131
+ - jsonlines==4.0.0
132
+ - kiwisolver==1.4.5
133
+ - langcodes==3.3.0
134
+ - markdown-it-py==3.0.0
135
+ - matplotlib==3.8.3
136
+ - mdurl==0.1.2
137
+ - mergedeep==1.3.4
138
+ - multidict==6.0.5
139
+ - multiprocess==0.70.16
140
+ - murmurhash==1.0.10
141
+ - mypy-extensions==1.0.0
142
+ - ninja==1.11.1.1
143
+ - packaging==23.2
144
+ - pandas==2.2.1
145
+ - preshed==3.0.9
146
+ - prismatic==0.0.1
147
+ - protobuf==4.25.3
148
+ - psutil==5.9.8
149
+ - py-cpuinfo==9.0.0
150
+ - pyarrow==15.0.0
151
+ - pyarrow-hotfix==0.6
152
+ - pycocoevalcap==1.2
153
+ - pycocotools==2.0.7
154
+ - pydantic==2.6.3
155
+ - pydantic-core==2.16.3
156
+ - pygments==2.17.2
157
+ - pynvml==11.5.0
158
+ - pyparsing==3.1.1
159
+ - pysocks==1.7.1
160
+ - pythainlp==5.0.1
161
+ - python-dateutil==2.9.0.post0
162
+ - pytz==2024.1
163
+ - pyyaml-include==1.3.2
164
+ - regex==2023.12.25
165
+ - rich==13.7.0
166
+ - safetensors==0.4.2
167
+ - sentencepiece==0.2.0
168
+ - sentry-sdk==1.40.5
169
+ - setproctitle==1.3.3
170
+ - six==1.16.0
171
+ - smart-open==6.4.0
172
+ - smmap==5.0.1
173
+ - soupsieve==2.5
174
+ - spacy==3.7.4
175
+ - spacy-legacy==3.0.12
176
+ - spacy-loggers==1.0.5
177
+ - srsly==2.4.8
178
+ - sudachidict-core==20240109
179
+ - sudachipy==0.6.8
180
+ - thinc==8.2.3
181
+ - timm==0.9.16
182
+ - tokenizers==0.15.2
183
+ - toml==0.10.2
184
+ - tqdm==4.66.2
185
+ - transformers==4.38.1
186
+ - typer==0.9.0
187
+ - typing-inspect==0.9.0
188
+ - tzdata==2024.1
189
+ - wandb==0.16.3
190
+ - wasabi==1.1.2
191
+ - weasel==0.3.4
192
+ - xxhash==3.4.1
193
+ - yarl==1.9.4
194
+ prefix: /home/lilei/anaconda3/envs/pvlm
stage1_qformer2_128/wandb/run-20240313_070332-tkilnh6o/files/config.yaml ADDED
@@ -0,0 +1,119 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wandb_version: 1
2
+
3
+ model:
4
+ desc: null
5
+ value:
6
+ type: one-stage+7b
7
+ model_id: one-stage+7b_qformer2_128
8
+ arch_specifier: qformer2_128
9
+ vision_backbone_id: clip-vit-l-336px
10
+ llm_backbone_id: vicuna-v15-7b
11
+ image_resize_strategy: letterbox
12
+ llm_max_length: 2048
13
+ align_epochs: 1
14
+ align_max_steps: null
15
+ align_global_batch_size: 256
16
+ align_per_device_batch_size: 16
17
+ align_learning_rate: 0.001
18
+ align_weight_decay: 0.0
19
+ align_max_grad_norm: 1.0
20
+ align_lr_scheduler_type: linear-warmup+cosine-decay
21
+ align_warmup_ratio: 0.03
22
+ align_train_strategy: fsdp-shard-grad-op
23
+ finetune_epochs: 1
24
+ finetune_max_steps: null
25
+ finetune_global_batch_size: 128
26
+ finetune_per_device_batch_size: 16
27
+ finetune_learning_rate: 2.0e-05
28
+ finetune_weight_decay: 0.1
29
+ finetune_max_grad_norm: 1.0
30
+ finetune_lr_scheduler_type: linear-warmup+cosine-decay
31
+ finetune_warmup_ratio: 0.03
32
+ finetune_train_strategy: fsdp-full-shard
33
+ enable_gradient_checkpointing: true
34
+ enable_mixed_precision_training: true
35
+ reduce_in_full_precision: false
36
+ dataset:
37
+ desc: null
38
+ value:
39
+ type: llava-v15
40
+ dataset_id: llava-v15
41
+ align_stage_components:
42
+ - download/llava-laion-cc-sbu-558k/chat.json
43
+ - download/llava-laion-cc-sbu-558k
44
+ finetune_stage_components:
45
+ - download/llava-v1.5-instruct/llava_v1_5_mix665k.json
46
+ - download/llava-v1.5-instruct
47
+ dataset_root_dir: data
48
+ stage:
49
+ desc: null
50
+ value: align
51
+ pretrained_checkpoint:
52
+ desc: null
53
+ value: null
54
+ run_id:
55
+ desc: null
56
+ value: stage1_qformer2_128
57
+ run_root_dir:
58
+ desc: null
59
+ value: runs
60
+ seed:
61
+ desc: null
62
+ value: 7
63
+ hf_token:
64
+ desc: null
65
+ value: .hf_token
66
+ trackers:
67
+ desc: null
68
+ value:
69
+ - jsonl
70
+ - wandb
71
+ wandb_project:
72
+ desc: null
73
+ value: hf-vlms
74
+ wandb_entity:
75
+ desc: null
76
+ value: lilei_stones
77
+ max_length:
78
+ desc: null
79
+ value: 4096
80
+ _wandb:
81
+ desc: null
82
+ value:
83
+ python_version: 3.10.13
84
+ cli_version: 0.16.3
85
+ framework: huggingface
86
+ huggingface_version: 4.38.1
87
+ is_jupyter_run: false
88
+ is_kaggle_kernel: false
89
+ start_time: 1710313412.489003
90
+ t:
91
+ 1:
92
+ - 1
93
+ - 11
94
+ - 41
95
+ - 49
96
+ - 51
97
+ - 55
98
+ - 63
99
+ - 71
100
+ 2:
101
+ - 1
102
+ - 11
103
+ - 41
104
+ - 49
105
+ - 51
106
+ - 55
107
+ - 63
108
+ - 71
109
+ 3:
110
+ - 2
111
+ - 13
112
+ - 16
113
+ - 23
114
+ 4: 3.10.13
115
+ 5: 0.16.3
116
+ 6: 4.38.1
117
+ 8:
118
+ - 5
119
+ 13: linux-x86_64
stage1_qformer2_128/wandb/run-20240313_070332-tkilnh6o/files/output.log ADDED
@@ -0,0 +1,1745 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ 03/13 [07:03:39] INFO  | >> [*] Starting Training Loop ]8;id=72103;file:///home/lilei/prismatic-vlms/scripts/pretrain.py\pretrain.py]8;;\:]8;id=880770;file:///home/lilei/prismatic-vlms/scripts/pretrain.py#226\226]8;;\
3
+
4
+
5
+
6
+
7
+
8
+
9
+
10
+
11
+
12
+
13
+
14
+
15
+
16
+
17
+
18
+
19
+
20
+
21
+
22
+
23
+
24
+
25
+
26
+
27
+
28
+
29
+
30
+
31
+
32
+
33
+
34
+
35
+
36
+
37
+
38
+
39
+
40
+
41
+
42
+
43
+
44
+
45
+
46
+
47
+
48
+
49
+
50
+
51
+
52
+
53
+
54
+
55
+
56
+
57
+
58
+
59
+
60
+
61
+
62
+
63
+
64
+
65
+
66
+
67
+
68
+
69
+
70
+
71
+
72
+
73
+
74
+
75
+
76
+
77
+
78
+
79
+
80
+
81
+
82
+
83
+
84
+
85
+
86
+
87
+
88
+
89
+
90
+
91
+
92
+
93
+
94
+
95
+
96
+
97
+
98
+
99
+
100
+
101
+
102
+
103
+
104
+
105
+
106
+
107
+
108
+
109
+
110
+
111
+
112
+
113
+
114
+
115
+
116
+
117
+
118
+
119
+
120
+
121
+
122
+
123
+
124
+
125
+
126
+
127
+
128
+
129
+
130
+
131
+
132
+
133
+
134
+
135
+
136
+
137
+
138
+
139
+
140
+
141
+
142
+
143
+
144
+
145
+
146
+
147
+
148
+
149
+
150
+
151
+
152
+
153
+
154
+
155
+
156
+
157
+
158
+
159
+
160
+
161
+
162
+
163
+
164
+
165
+
166
+
167
+
168
+
169
+
170
+
171
+
172
+
173
+
174
+
175
+
176
+
177
+
178
+
179
+
180
+
181
+
182
+
183
+
184
+
185
+
186
+
187
+
188
+
189
+
190
+
191
+
192
+
193
+
194
+
195
+
196
+
197
+
198
+
199
+
200
+
201
+
202
+
203
+
204
+
205
+
206
+
207
+
208
+
209
+
210
+
211
+
212
+
213
+
214
+
215
+
216
+
217
+
218
+
219
+
220
+
221
+
222
+
223
+
224
+
225
+
226
+
227
+
228
+
229
+
230
+
231
+
232
+
233
+
234
+
235
+
236
+
237
+
238
+
239
+
240
+
241
+
242
+
243
+
244
+
245
+
246
+
247
+
248
+
249
+
250
+
251
+
252
+
253
+
254
+
255
+
256
+
257
+
258
+
259
+
260
+
261
+
262
+
263
+
264
+
265
+
266
+
267
+
268
+
269
+
270
+
271
+
272
+
273
+
274
+
275
+
276
+
277
+
278
+
279
+
280
+
281
+
282
+
283
+
284
+
285
+
286
+
287
+
288
+
289
+
290
+
291
+
292
+
293
+
294
+
295
+
296
+
297
+
298
+
299
+
300
+
301
+
302
+
303
+
304
+
305
+
306
+
307
+
308
+
309
+
310
+
311
+
312
+
313
+
314
+
315
+
316
+
317
+
318
+
319
+
320
+
321
+
322
+
323
+
324
+
325
+
326
+
327
+
328
+
329
+
330
+
331
+
332
+
333
+
334
+
335
+
336
+
337
+
338
+
339
+
340
+
341
+
342
+
343
+
344
+
345
+
346
+
347
+
348
+
349
+
350
+
351
+
352
+
353
+
354
+
355
+
356
+
357
+
358
+
359
+
360
+
361
+
362
+
363
+
364
+
365
+
366
+
367
+
368
+
369
+
370
+
371
+
372
+
373
+
374
+
375
+
376
+
377
+
378
+
379
+
380
+
381
+
382
+
383
+
384
+
385
+
386
+
387
+
388
+
389
+
390
+
391
+
392
+
393
+
394
+
395
+
396
+
397
+
398
+
399
+
400
+
401
+
402
+
403
+
404
+
405
+
406
+
407
+
408
+
409
+
410
+
411
+
412
+
413
+
414
+
415
+
416
+
417
+
418
+
419
+
420
+
421
+
422
+
423
+
424
+
425
+
426
+
427
+
428
+
429
+
430
+
431
+
432
+
433
+
434
+
435
+
436
+
437
+
438
+
439
+
440
+
441
+
442
+
443
+
444
+
445
+
446
+
447
+
448
+
449
+
450
+
451
+
452
+
453
+
454
+
455
+
456
+
457
+
458
+
459
+
460
+
461
+
462
+
463
+
464
+
465
+
466
+
467
+
468
+
469
+
470
+
471
+
472
+
473
+
474
+
475
+
476
+
477
+
478
+
479
+
480
+
481
+
482
+
483
+
484
+
485
+
486
+
487
+
488
+
489
+
490
+
491
+
492
+
493
+
494
+
495
+
496
+
497
+
498
+
499
+
500
+
501
+
502
+
503
+
504
+
505
+
506
+
507
+
508
+
509
+
510
+
511
+
512
+
513
+
514
+
515
+
516
+
517
+
518
+
519
+
520
+
521
+
522
+
523
+
524
+
525
+
526
+
527
+
528
+
529
+
530
+
531
+
532
+
533
+
534
+
535
+
536
+
537
+
538
+
539
+
540
+
541
+
542
+
543
+
544
+
545
+
546
+
547
+
548
+
549
+
550
+
551
+
552
+
553
+
554
+
555
+
556
+
557
+
558
+
559
+
560
+
561
+
562
+
563
+
564
+
565
+
566
+
567
+
568
+
569
+
570
+
571
+
572
+
573
+
574
+
575
+
576
+
577
+
578
+
579
+
580
+
581
+
582
+
583
+
584
+
585
+
586
+
587
+
588
+
589
+
590
+
591
+
592
+
593
+
594
+
595
+
596
+
597
+
598
+
599
+
600
+
601
+
602
+
603
+
604
+
605
+
606
+
607
+
608
+
609
+
610
+
611
+
612
+
613
+
614
+
615
+
616
+
617
+
618
+
619
+
620
+
621
+
622
+
623
+
624
+
625
+
626
+
627
+
628
+
629
+
630
+
631
+
632
+
633
+
634
+
635
+
636
+
637
+
638
+
639
+
640
+
641
+
642
+
643
+
644
+
645
+
646
+
647
+
648
+
649
+
650
+
651
+
652
+
653
+
654
+
655
+
656
+
657
+
658
+
659
+
660
+
661
+
662
+
663
+
664
+
665
+
666
+
667
+
668
+
669
+
670
+
671
+
672
+
673
+
674
+
675
+
676
+
677
+
678
+
679
+
680
+
681
+
682
+
683
+
684
+
685
+
686
+
687
+
688
+
689
+
690
+
691
+
692
+
693
+
694
+
695
+
696
+
697
+
698
+
699
+
700
+
701
+
702
+
703
+
704
+
705
+
706
+
707
+
708
+
709
+
710
+
711
+
712
+
713
+
714
+
715
+
716
+
717
+
718
+
719
+
720
+
721
+
722
+
723
+
724
+
725
+
726
+
727
+
728
+
729
+
730
+
731
+
732
+
733
+
734
+
735
+
736
+
737
+
738
+
739
+
740
+
741
+
742
+
743
+
744
+
745
+
746
+
747
+
748
+
749
+
750
+
751
+
752
+
753
+
754
+
755
+
756
+
757
+
758
+
759
+
760
+
761
+
762
+
763
+
764
+
765
+
766
+
767
+
768
+
769
+
770
+
771
+
772
+
773
+
774
+
775
+
776
+
777
+
778
+
779
+
780
+
781
+
782
+
783
+
784
+
785
+
786
+
787
+
788
+
789
+
790
+
791
+
792
+
793
+
794
+
795
+
796
+
797
+
798
+
799
+
800
+
801
+
802
+
803
+
804
+
805
+
806
+
807
+
808
+
809
+
810
+
811
+
812
+
813
+
814
+
815
+
816
+
817
+
818
+
819
+
820
+
821
+
822
+
823
+
824
+
825
+
826
+
827
+
828
+
829
+
830
+
831
+
832
+
833
+
834
+
835
+
836
+
837
+
838
+
839
+
840
+
841
+
842
+
843
+
844
+
845
+
846
+
847
+
848
+
849
+
850
+
851
+
852
+
853
+
854
+
855
+
856
+
857
+
858
+
859
+
860
+
861
+
862
+
863
+
864
+
865
+
866
+
867
+
868
+
869
+
870
+
871
+
872
+
873
+
874
+
875
+
876
+
877
+
878
+
879
+
880
+
881
+
882
+
883
+
884
+
885
+
886
+
887
+
888
+
889
+
890
+
891
+
892
+
893
+
894
+
895
+
896
+
897
+
898
+
899
+
900
+
901
+
902
+
903
+
904
+
905
+
906
+
907
+
908
+
909
+
910
+
911
+
912
+
913
+
914
+
915
+
916
+
917
+
918
+
919
+
920
+
921
+
922
+
923
+
924
+
925
+
926
+
927
+
928
+
929
+
930
+
931
+
932
+
933
+
934
+
935
+
936
+
937
+
938
+
939
+
940
+
941
+
942
+
943
+
944
+
945
+
946
+
947
+
948
+
949
+
950
+
951
+
952
+
953
+
954
+
955
+
956
+
957
+
958
+
959
+
960
+
961
+
962
+
963
+
964
+
965
+
966
+
967
+
968
+
969
+
970
+
971
+
972
+
973
+
974
+
975
+
976
+
977
+
978
+
979
+
980
+
981
+
982
+
983
+
984
+
985
+
986
+
987
+
988
+
989
+
990
+
991
+
992
+
993
+
994
+
995
+
996
+
997
+
998
+
999
+
1000
+
1001
+
1002
+
1003
+
1004
+
1005
+
1006
+
1007
+
1008
+
1009
+
1010
+
1011
+
1012
+
1013
+
1014
+
1015
+
1016
+
1017
+
1018
+
1019
+
1020
+
1021
+
1022
+
1023
+
1024
+
1025
+
1026
+
1027
+
1028
+
1029
+
1030
+
1031
+
1032
+
1033
+
1034
+
1035
+
1036
+
1037
+
1038
+
1039
+
1040
+
1041
+
1042
+
1043
+
1044
+
1045
+
1046
+
1047
+
1048
+
1049
+
1050
+
1051
+
1052
+
1053
+
1054
+
1055
+
1056
+
1057
+
1058
+
1059
+
1060
+
1061
+
1062
+
1063
+
1064
+
1065
+
1066
+
1067
+
1068
+
1069
+
1070
+
1071
+
1072
+
1073
+
1074
+
1075
+
1076
+
1077
+
1078
+
1079
+
1080
+
1081
+
1082
+
1083
+
1084
+
1085
+
1086
+
1087
+
1088
+
1089
+
1090
+
1091
+
1092
+
1093
+
1094
+
1095
+
1096
+
1097
+
1098
+
1099
+
1100
+
1101
+
1102
+
1103
+
1104
+
1105
+
1106
+
1107
+
1108
+
1109
+
1110
+
1111
+
1112
+
1113
+
1114
+
1115
+
1116
+
1117
+
1118
+
1119
+
1120
+
1121
+
1122
+
1123
+
1124
+
1125
+
1126
+
1127
+
1128
+
1129
+
1130
+
1131
+
1132
+
1133
+
1134
+
1135
+
1136
+
1137
+
1138
+
1139
+
1140
+
1141
+
1142
+
1143
+
1144
+
1145
+
1146
+
1147
+
1148
+
1149
+
1150
+
1151
+
1152
+
1153
+
1154
+
1155
+
1156
+
1157
+
1158
+
1159
+
1160
+
1161
+
1162
+
1163
+
1164
+
1165
+
1166
+
1167
+
1168
+
1169
+
1170
+
1171
+
1172
+
1173
+
1174
+
1175
+
1176
+
1177
+
1178
+
1179
+
1180
+
1181
+
1182
+
1183
+
1184
+
1185
+
1186
+
1187
+
1188
+
1189
+
1190
+
1191
+
1192
+
1193
+
1194
+
1195
+
1196
+
1197
+
1198
+
1199
+
1200
+
1201
+
1202
+
1203
+
1204
+
1205
+
1206
+
1207
+
1208
+
1209
+
1210
+
1211
+
1212
+
1213
+
1214
+
1215
+
1216
+
1217
+
1218
+
1219
+
1220
+
1221
+
1222
+
1223
+
1224
+
1225
+
1226
+
1227
+
1228
+
1229
+
1230
+
1231
+
1232
+
1233
+
1234
+
1235
+
1236
+
1237
+
1238
+
1239
+
1240
+
1241
+
1242
+
1243
+
1244
+
1245
+
1246
+
1247
+
1248
+
1249
+
1250
+
1251
+
1252
+
1253
+
1254
+
1255
+
1256
+
1257
+
1258
+
1259
+
1260
+
1261
+
1262
+
1263
+
1264
+
1265
+
1266
+
1267
+
1268
+
1269
+
1270
+
1271
+
1272
+
1273
+
1274
+
1275
+
1276
+
1277
+
1278
+
1279
+
1280
+
1281
+
1282
+
1283
+
1284
+
1285
+
1286
+
1287
+
1288
+
1289
+
1290
+
1291
+
1292
+
1293
+
1294
+
1295
+
1296
+
1297
+
1298
+
1299
+
1300
+
1301
+
1302
+
1303
+
1304
+
1305
+
1306
+
1307
+
1308
+
1309
+
1310
+
1311
+
1312
+
1313
+
1314
+
1315
+
1316
+
1317
+
1318
+
1319
+
1320
+
1321
+
1322
+
1323
+
1324
+
1325
+
1326
+
1327
+
1328
+
1329
+
1330
+
1331
+
1332
+
1333
+
1334
+
1335
+
1336
+
1337
+
1338
+
1339
+
1340
+
1341
+
1342
+
1343
+
1344
+
1345
+
1346
+
1347
+
1348
+
1349
+
1350
+
1351
+
1352
+
1353
+
1354
+
1355
+
1356
+
1357
+
1358
+
1359
+
1360
+
1361
+
1362
+
1363
+
1364
+
1365
+
1366
+
1367
+
1368
+
1369
+
1370
+
1371
+
1372
+
1373
+
1374
+
1375
+
1376
+
1377
+
1378
+
1379
+
1380
+
1381
+
1382
+
1383
+
1384
+
1385
+
1386
+
1387
+
1388
+
1389
+
1390
+
1391
+
1392
+
1393
+
1394
+
1395
+
1396
+
1397
+
1398
+
1399
+
1400
+
1401
+
1402
+
1403
+
1404
+
1405
+
1406
+
1407
+
1408
+
1409
+
1410
+
1411
+
1412
+
1413
+
1414
+
1415
+
1416
+
1417
+
1418
+
1419
+
1420
+
1421
+
1422
+
1423
+
1424
+
1425
+
1426
+
1427
+
1428
+
1429
+
1430
+
1431
+
1432
+
1433
+
1434
+
1435
+
1436
+
1437
+
1438
+
1439
+
1440
+
1441
+
1442
+
1443
+
1444
+
1445
+
1446
+
1447
+
1448
+
1449
+
1450
+
1451
+
1452
+
1453
+
1454
+
1455
+
1456
+
1457
+
1458
+
1459
+
1460
+
1461
+
1462
+
1463
+
1464
+
1465
+
1466
+
1467
+
1468
+
1469
+
1470
+
1471
+
1472
+
1473
+
1474
+
1475
+
1476
+
1477
+
1478
+
1479
+
1480
+
1481
+
1482
+
1483
+
1484
+
1485
+
1486
+
1487
+
1488
+
1489
+
1490
+
1491
+
1492
+
1493
+
1494
+
1495
+
1496
+
1497
+
1498
+
1499
+
1500
+
1501
+
1502
+
1503
+
1504
+
1505
+
1506
+
1507
+
1508
+
1509
+
1510
+
1511
+
1512
+
1513
+
1514
+
1515
+
1516
+
1517
+
1518
+
1519
+
1520
+
1521
+
1522
+
1523
+
1524
+
1525
+
1526
+
1527
+
1528
+
1529
+
1530
+
1531
+
1532
+
1533
+
1534
+
1535
+
1536
+
1537
+
1538
+
1539
+
1540
+
1541
+
1542
+
1543
+
1544
+
1545
+
1546
+
1547
+
1548
+
1549
+
1550
+
1551
+
1552
+
1553
+
1554
+
1555
+
1556
+
1557
+
1558
+
1559
+
1560
+
1561
+
1562
+
1563
+
1564
+
1565
+
1566
+
1567
+
1568
+
1569
+
1570
+
1571
+
1572
+
1573
+
1574
+
1575
+
1576
+
1577
+
1578
+
1579
+
1580
+
1581
+
1582
+
1583
+
1584
+
1585
+
1586
+
1587
+
1588
+
1589
+
1590
+
1591
+
1592
+
1593
+
1594
+
1595
+
1596
+
1597
+
1598
+
1599
+
1600
+
1601
+
1602
+
1603
+
1604
+
1605
+
1606
+
1607
+
1608
+
1609
+
1610
+
1611
+
1612
+
1613
+
1614
+
1615
+
1616
+
1617
+
1618
+
1619
+
1620
+
1621
+
1622
+
1623
+
1624
+
1625
+
1626
+
1627
+
1628
+
1629
+
1630
+
1631
+
1632
+
1633
+
1634
+
1635
+
1636
+
1637
+
1638
+
1639
+
1640
+
1641
+
1642
+
1643
+
1644
+
1645
+
1646
+
1647
+
1648
+
1649
+
1650
+
1651
+
1652
+
1653
+
1654
+
1655
+
1656
+
1657
+
1658
+
1659
+
1660
+
1661
+
1662
+
1663
+
1664
+
1665
+
1666
+
1667
+
1668
+
1669
+
1670
+
1671
+
1672
+
1673
+
1674
+
1675
+
1676
+
1677
+
1678
+
1679
+
1680
+
1681
+
1682
+
1683
+
1684
+
1685
+
1686
+
1687
+
1688
+
1689
+
1690
+
1691
+
1692
+
1693
+
1694
+
1695
+
1696
+
1697
+
1698
+
1699
+
1700
+
1701
+
1702
+
1703
+
1704
+
1705
+
1706
+
1707
+
1708
+
1709
+
1710
+
1711
+
1712
+
1713
+
1714
+
1715
+
1716
+
1717
+
1718
+
1719
+
1720
+
1721
+
1722
+
1723
+
1724
+
1725
+
1726
+
1727
+
1728
+
1729
+
1730
+
1731
+
1732
+
1733
+
1734
+
1735
+
1736
+
1737
+
1738
+
1739
+
1740
+
1741
+
1742
+
1743
+ =>> [Global Step] 002180 =>> LR :: 0.000000 -- Loss :: 3.2349: 100%|█| 2180/2180 [58:26<00:00, 1.59s/i
1744
+
1745
+ 03/13 [08:02:43] INFO  | >> [*] Done with Training =>> Finalizing Metrics ]8;id=696414;file:///home/lilei/prismatic-vlms/scripts/pretrain.py\pretrain.py]8;;\:]8;id=68157;file:///home/lilei/prismatic-vlms/scripts/pretrain.py#230\230]8;;\
stage1_qformer2_128/wandb/run-20240313_070332-tkilnh6o/files/requirements.txt ADDED
@@ -0,0 +1,122 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ accelerate==0.27.2
2
+ aiohttp==3.9.3
3
+ aiosignal==1.3.1
4
+ annotated-types==0.6.0
5
+ appdirs==1.4.4
6
+ async-timeout==4.0.3
7
+ attrs==23.2.0
8
+ beautifulsoup4==4.12.3
9
+ blis==0.7.11
10
+ catalogue==2.0.10
11
+ certifi==2024.2.2
12
+ charset-normalizer==2.0.4
13
+ click==8.1.7
14
+ cloudpathlib==0.16.0
15
+ confection==0.1.4
16
+ contourpy==1.2.0
17
+ cycler==0.12.1
18
+ cymem==2.0.8
19
+ datasets==2.18.0
20
+ deepspeed==0.13.4
21
+ dill==0.3.8
22
+ docker-pycreds==0.4.0
23
+ draccus==0.7.1
24
+ einops==0.7.0
25
+ filelock==3.13.1
26
+ flash-attn==2.5.5
27
+ fonttools==4.49.0
28
+ frozenlist==1.4.1
29
+ fsspec==2024.2.0
30
+ gdown==5.1.0
31
+ gitdb==4.0.11
32
+ gitpython==3.1.42
33
+ gmpy2==2.1.2
34
+ hjson==3.1.0
35
+ huggingface-hub==0.20.3
36
+ idna==3.4
37
+ install==1.3.5
38
+ jinja2==3.1.3
39
+ jsonlines==4.0.0
40
+ kiwisolver==1.4.5
41
+ langcodes==3.3.0
42
+ markdown-it-py==3.0.0
43
+ markupsafe==2.1.3
44
+ matplotlib==3.8.3
45
+ mdurl==0.1.2
46
+ mergedeep==1.3.4
47
+ mkl-fft==1.3.8
48
+ mkl-random==1.2.4
49
+ mkl-service==2.4.0
50
+ mpmath==1.3.0
51
+ multidict==6.0.5
52
+ multiprocess==0.70.16
53
+ murmurhash==1.0.10
54
+ mypy-extensions==1.0.0
55
+ networkx==3.1
56
+ ninja==1.11.1.1
57
+ numpy==1.26.4
58
+ ocifs==1.3.1
59
+ packaging==23.2
60
+ pandas==2.2.1
61
+ pillow==10.2.0
62
+ pip==23.3.1
63
+ preshed==3.0.9
64
+ prismatic==0.0.1
65
+ protobuf==4.25.3
66
+ psutil==5.9.8
67
+ py-cpuinfo==9.0.0
68
+ pyarrow-hotfix==0.6
69
+ pyarrow==15.0.0
70
+ pycocoevalcap==1.2
71
+ pycocotools==2.0.7
72
+ pydantic-core==2.16.3
73
+ pydantic==2.6.3
74
+ pygments==2.17.2
75
+ pynvml==11.5.0
76
+ pyparsing==3.1.1
77
+ pysocks==1.7.1
78
+ pythainlp==5.0.1
79
+ python-dateutil==2.9.0.post0
80
+ pytz==2024.1
81
+ pyyaml-include==1.3.2
82
+ pyyaml==6.0.1
83
+ regex==2023.12.25
84
+ requests==2.31.0
85
+ rich==13.7.0
86
+ safetensors==0.4.2
87
+ sentencepiece==0.2.0
88
+ sentry-sdk==1.40.5
89
+ setproctitle==1.3.3
90
+ setuptools==68.2.2
91
+ six==1.16.0
92
+ smart-open==6.4.0
93
+ smmap==5.0.1
94
+ soupsieve==2.5
95
+ spacy-legacy==3.0.12
96
+ spacy-loggers==1.0.5
97
+ spacy==3.7.4
98
+ srsly==2.4.8
99
+ sudachidict-core==20240109
100
+ sudachipy==0.6.8
101
+ sympy==1.12
102
+ thinc==8.2.3
103
+ timm==0.9.16
104
+ tokenizers==0.15.2
105
+ toml==0.10.2
106
+ torch==2.2.1
107
+ torchaudio==2.2.1
108
+ torchvision==0.17.1
109
+ tqdm==4.66.2
110
+ transformers==4.38.1
111
+ triton==2.2.0
112
+ typer==0.9.0
113
+ typing-extensions==4.9.0
114
+ typing-inspect==0.9.0
115
+ tzdata==2024.1
116
+ urllib3==2.1.0
117
+ wandb==0.16.3
118
+ wasabi==1.1.2
119
+ weasel==0.3.4
120
+ wheel==0.41.2
121
+ xxhash==3.4.1
122
+ yarl==1.9.4
stage1_qformer2_128/wandb/run-20240313_070332-tkilnh6o/files/wandb-metadata.json ADDED
@@ -0,0 +1,1362 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.15.0-1027-oracle-x86_64-with-glibc2.31",
3
+ "python": "3.10.13",
4
+ "heartbeatAt": "2024-03-13T07:03:33.267957",
5
+ "startedAt": "2024-03-13T07:03:32.435991",
6
+ "docker": null,
7
+ "cuda": null,
8
+ "args": [
9
+ "--model.type",
10
+ "one-stage+7b",
11
+ "--stage",
12
+ "align",
13
+ "--model.model_id",
14
+ "one-stage+7b_qformer2_128",
15
+ "--run_id",
16
+ "stage1_qformer2_128",
17
+ "--model.arch_specifier",
18
+ "qformer2_128"
19
+ ],
20
+ "state": "running",
21
+ "program": "/home/lilei/prismatic-vlms/scripts/pretrain.py",
22
+ "codePathLocal": "scripts/pretrain.py",
23
+ "codePath": "scripts/pretrain.py",
24
+ "git": {
25
+ "remote": "https://github.com/TRI-ML/prismatic-vlms.git",
26
+ "commit": "e87c4d04915bf8fad0d90f2ae79454df778ee2d3"
27
+ },
28
+ "email": "tobiaslee@qq.com",
29
+ "root": "/home/lilei/prismatic-vlms",
30
+ "host": "compute-permanent-node-262",
31
+ "username": "lilei",
32
+ "executable": "/home/lilei/anaconda3/envs/pvlm/bin/python",
33
+ "cpu_count": 128,
34
+ "cpu_count_logical": 255,
35
+ "cpu_freq": {
36
+ "current": 2572.1024235294117,
37
+ "min": 1500.0,
38
+ "max": 2550.0
39
+ },
40
+ "cpu_freq_per_core": [
41
+ {
42
+ "current": 3248.751,
43
+ "min": 1500.0,
44
+ "max": 2550.0
45
+ },
46
+ {
47
+ "current": 2550.0,
48
+ "min": 1500.0,
49
+ "max": 2550.0
50
+ },
51
+ {
52
+ "current": 2550.0,
53
+ "min": 1500.0,
54
+ "max": 2550.0
55
+ },
56
+ {
57
+ "current": 2550.0,
58
+ "min": 1500.0,
59
+ "max": 2550.0
60
+ },
61
+ {
62
+ "current": 2550.0,
63
+ "min": 1500.0,
64
+ "max": 2550.0
65
+ },
66
+ {
67
+ "current": 2550.0,
68
+ "min": 1500.0,
69
+ "max": 2550.0
70
+ },
71
+ {
72
+ "current": 2550.0,
73
+ "min": 1500.0,
74
+ "max": 2550.0
75
+ },
76
+ {
77
+ "current": 2550.0,
78
+ "min": 1500.0,
79
+ "max": 2550.0
80
+ },
81
+ {
82
+ "current": 2550.0,
83
+ "min": 1500.0,
84
+ "max": 2550.0
85
+ },
86
+ {
87
+ "current": 2550.0,
88
+ "min": 1500.0,
89
+ "max": 2550.0
90
+ },
91
+ {
92
+ "current": 3227.828,
93
+ "min": 1500.0,
94
+ "max": 2550.0
95
+ },
96
+ {
97
+ "current": 2550.0,
98
+ "min": 1500.0,
99
+ "max": 2550.0
100
+ },
101
+ {
102
+ "current": 2550.0,
103
+ "min": 1500.0,
104
+ "max": 2550.0
105
+ },
106
+ {
107
+ "current": 2550.0,
108
+ "min": 1500.0,
109
+ "max": 2550.0
110
+ },
111
+ {
112
+ "current": 2550.0,
113
+ "min": 1500.0,
114
+ "max": 2550.0
115
+ },
116
+ {
117
+ "current": 2550.0,
118
+ "min": 1500.0,
119
+ "max": 2550.0
120
+ },
121
+ {
122
+ "current": 2550.0,
123
+ "min": 1500.0,
124
+ "max": 2550.0
125
+ },
126
+ {
127
+ "current": 2550.0,
128
+ "min": 1500.0,
129
+ "max": 2550.0
130
+ },
131
+ {
132
+ "current": 2550.0,
133
+ "min": 1500.0,
134
+ "max": 2550.0
135
+ },
136
+ {
137
+ "current": 2550.0,
138
+ "min": 1500.0,
139
+ "max": 2550.0
140
+ },
141
+ {
142
+ "current": 2550.0,
143
+ "min": 1500.0,
144
+ "max": 2550.0
145
+ },
146
+ {
147
+ "current": 2550.0,
148
+ "min": 1500.0,
149
+ "max": 2550.0
150
+ },
151
+ {
152
+ "current": 2550.0,
153
+ "min": 1500.0,
154
+ "max": 2550.0
155
+ },
156
+ {
157
+ "current": 2550.0,
158
+ "min": 1500.0,
159
+ "max": 2550.0
160
+ },
161
+ {
162
+ "current": 2550.0,
163
+ "min": 1500.0,
164
+ "max": 2550.0
165
+ },
166
+ {
167
+ "current": 2550.0,
168
+ "min": 1500.0,
169
+ "max": 2550.0
170
+ },
171
+ {
172
+ "current": 2550.0,
173
+ "min": 1500.0,
174
+ "max": 2550.0
175
+ },
176
+ {
177
+ "current": 2550.0,
178
+ "min": 1500.0,
179
+ "max": 2550.0
180
+ },
181
+ {
182
+ "current": 2550.0,
183
+ "min": 1500.0,
184
+ "max": 2550.0
185
+ },
186
+ {
187
+ "current": 2550.0,
188
+ "min": 1500.0,
189
+ "max": 2550.0
190
+ },
191
+ {
192
+ "current": 2550.0,
193
+ "min": 1500.0,
194
+ "max": 2550.0
195
+ },
196
+ {
197
+ "current": 2550.0,
198
+ "min": 1500.0,
199
+ "max": 2550.0
200
+ },
201
+ {
202
+ "current": 2550.0,
203
+ "min": 1500.0,
204
+ "max": 2550.0
205
+ },
206
+ {
207
+ "current": 2550.0,
208
+ "min": 1500.0,
209
+ "max": 2550.0
210
+ },
211
+ {
212
+ "current": 2550.0,
213
+ "min": 1500.0,
214
+ "max": 2550.0
215
+ },
216
+ {
217
+ "current": 2550.0,
218
+ "min": 1500.0,
219
+ "max": 2550.0
220
+ },
221
+ {
222
+ "current": 2550.0,
223
+ "min": 1500.0,
224
+ "max": 2550.0
225
+ },
226
+ {
227
+ "current": 2550.0,
228
+ "min": 1500.0,
229
+ "max": 2550.0
230
+ },
231
+ {
232
+ "current": 2550.0,
233
+ "min": 1500.0,
234
+ "max": 2550.0
235
+ },
236
+ {
237
+ "current": 2550.0,
238
+ "min": 1500.0,
239
+ "max": 2550.0
240
+ },
241
+ {
242
+ "current": 3249.655,
243
+ "min": 1500.0,
244
+ "max": 2550.0
245
+ },
246
+ {
247
+ "current": 2550.0,
248
+ "min": 1500.0,
249
+ "max": 2550.0
250
+ },
251
+ {
252
+ "current": 2550.0,
253
+ "min": 1500.0,
254
+ "max": 2550.0
255
+ },
256
+ {
257
+ "current": 2550.0,
258
+ "min": 1500.0,
259
+ "max": 2550.0
260
+ },
261
+ {
262
+ "current": 2550.0,
263
+ "min": 1500.0,
264
+ "max": 2550.0
265
+ },
266
+ {
267
+ "current": 2550.0,
268
+ "min": 1500.0,
269
+ "max": 2550.0
270
+ },
271
+ {
272
+ "current": 2550.0,
273
+ "min": 1500.0,
274
+ "max": 2550.0
275
+ },
276
+ {
277
+ "current": 2550.0,
278
+ "min": 1500.0,
279
+ "max": 2550.0
280
+ },
281
+ {
282
+ "current": 2550.0,
283
+ "min": 1500.0,
284
+ "max": 2550.0
285
+ },
286
+ {
287
+ "current": 2550.0,
288
+ "min": 1500.0,
289
+ "max": 2550.0
290
+ },
291
+ {
292
+ "current": 2550.0,
293
+ "min": 1500.0,
294
+ "max": 2550.0
295
+ },
296
+ {
297
+ "current": 2550.0,
298
+ "min": 1500.0,
299
+ "max": 2550.0
300
+ },
301
+ {
302
+ "current": 3249.383,
303
+ "min": 1500.0,
304
+ "max": 2550.0
305
+ },
306
+ {
307
+ "current": 2550.0,
308
+ "min": 1500.0,
309
+ "max": 2550.0
310
+ },
311
+ {
312
+ "current": 2550.0,
313
+ "min": 1500.0,
314
+ "max": 2550.0
315
+ },
316
+ {
317
+ "current": 2550.0,
318
+ "min": 1500.0,
319
+ "max": 2550.0
320
+ },
321
+ {
322
+ "current": 2550.0,
323
+ "min": 1500.0,
324
+ "max": 2550.0
325
+ },
326
+ {
327
+ "current": 2550.0,
328
+ "min": 1500.0,
329
+ "max": 2550.0
330
+ },
331
+ {
332
+ "current": 2550.0,
333
+ "min": 1500.0,
334
+ "max": 2550.0
335
+ },
336
+ {
337
+ "current": 2550.0,
338
+ "min": 1500.0,
339
+ "max": 2550.0
340
+ },
341
+ {
342
+ "current": 2550.0,
343
+ "min": 1500.0,
344
+ "max": 2550.0
345
+ },
346
+ {
347
+ "current": 3249.559,
348
+ "min": 1500.0,
349
+ "max": 2550.0
350
+ },
351
+ {
352
+ "current": 2550.0,
353
+ "min": 1500.0,
354
+ "max": 2550.0
355
+ },
356
+ {
357
+ "current": 2550.0,
358
+ "min": 1500.0,
359
+ "max": 2550.0
360
+ },
361
+ {
362
+ "current": 2550.0,
363
+ "min": 1500.0,
364
+ "max": 2550.0
365
+ },
366
+ {
367
+ "current": 2550.0,
368
+ "min": 1500.0,
369
+ "max": 2550.0
370
+ },
371
+ {
372
+ "current": 2550.0,
373
+ "min": 1500.0,
374
+ "max": 2550.0
375
+ },
376
+ {
377
+ "current": 2550.0,
378
+ "min": 1500.0,
379
+ "max": 2550.0
380
+ },
381
+ {
382
+ "current": 2550.0,
383
+ "min": 1500.0,
384
+ "max": 2550.0
385
+ },
386
+ {
387
+ "current": 2550.0,
388
+ "min": 1500.0,
389
+ "max": 2550.0
390
+ },
391
+ {
392
+ "current": 2550.0,
393
+ "min": 1500.0,
394
+ "max": 2550.0
395
+ },
396
+ {
397
+ "current": 2550.0,
398
+ "min": 1500.0,
399
+ "max": 2550.0
400
+ },
401
+ {
402
+ "current": 2550.0,
403
+ "min": 1500.0,
404
+ "max": 2550.0
405
+ },
406
+ {
407
+ "current": 2550.0,
408
+ "min": 1500.0,
409
+ "max": 2550.0
410
+ },
411
+ {
412
+ "current": 2550.0,
413
+ "min": 1500.0,
414
+ "max": 2550.0
415
+ },
416
+ {
417
+ "current": 2550.0,
418
+ "min": 1500.0,
419
+ "max": 2550.0
420
+ },
421
+ {
422
+ "current": 2550.0,
423
+ "min": 1500.0,
424
+ "max": 2550.0
425
+ },
426
+ {
427
+ "current": 2550.0,
428
+ "min": 1500.0,
429
+ "max": 2550.0
430
+ },
431
+ {
432
+ "current": 2550.0,
433
+ "min": 1500.0,
434
+ "max": 2550.0
435
+ },
436
+ {
437
+ "current": 2550.0,
438
+ "min": 1500.0,
439
+ "max": 2550.0
440
+ },
441
+ {
442
+ "current": 2550.0,
443
+ "min": 1500.0,
444
+ "max": 2550.0
445
+ },
446
+ {
447
+ "current": 2550.0,
448
+ "min": 1500.0,
449
+ "max": 2550.0
450
+ },
451
+ {
452
+ "current": 2550.0,
453
+ "min": 1500.0,
454
+ "max": 2550.0
455
+ },
456
+ {
457
+ "current": 2550.0,
458
+ "min": 1500.0,
459
+ "max": 2550.0
460
+ },
461
+ {
462
+ "current": 2550.0,
463
+ "min": 1500.0,
464
+ "max": 2550.0
465
+ },
466
+ {
467
+ "current": 2550.0,
468
+ "min": 1500.0,
469
+ "max": 2550.0
470
+ },
471
+ {
472
+ "current": 2550.0,
473
+ "min": 1500.0,
474
+ "max": 2550.0
475
+ },
476
+ {
477
+ "current": 2550.0,
478
+ "min": 1500.0,
479
+ "max": 2550.0
480
+ },
481
+ {
482
+ "current": 2550.0,
483
+ "min": 1500.0,
484
+ "max": 2550.0
485
+ },
486
+ {
487
+ "current": 2550.0,
488
+ "min": 1500.0,
489
+ "max": 2550.0
490
+ },
491
+ {
492
+ "current": 2550.0,
493
+ "min": 1500.0,
494
+ "max": 2550.0
495
+ },
496
+ {
497
+ "current": 2550.0,
498
+ "min": 1500.0,
499
+ "max": 2550.0
500
+ },
501
+ {
502
+ "current": 2550.0,
503
+ "min": 1500.0,
504
+ "max": 2550.0
505
+ },
506
+ {
507
+ "current": 2550.0,
508
+ "min": 1500.0,
509
+ "max": 2550.0
510
+ },
511
+ {
512
+ "current": 2550.0,
513
+ "min": 1500.0,
514
+ "max": 2550.0
515
+ },
516
+ {
517
+ "current": 2550.0,
518
+ "min": 1500.0,
519
+ "max": 2550.0
520
+ },
521
+ {
522
+ "current": 2550.0,
523
+ "min": 1500.0,
524
+ "max": 2550.0
525
+ },
526
+ {
527
+ "current": 2550.0,
528
+ "min": 1500.0,
529
+ "max": 2550.0
530
+ },
531
+ {
532
+ "current": 2550.0,
533
+ "min": 1500.0,
534
+ "max": 2550.0
535
+ },
536
+ {
537
+ "current": 2550.0,
538
+ "min": 1500.0,
539
+ "max": 2550.0
540
+ },
541
+ {
542
+ "current": 2550.0,
543
+ "min": 1500.0,
544
+ "max": 2550.0
545
+ },
546
+ {
547
+ "current": 2550.0,
548
+ "min": 1500.0,
549
+ "max": 2550.0
550
+ },
551
+ {
552
+ "current": 2550.0,
553
+ "min": 1500.0,
554
+ "max": 2550.0
555
+ },
556
+ {
557
+ "current": 2550.0,
558
+ "min": 1500.0,
559
+ "max": 2550.0
560
+ },
561
+ {
562
+ "current": 2550.0,
563
+ "min": 1500.0,
564
+ "max": 2550.0
565
+ },
566
+ {
567
+ "current": 2550.0,
568
+ "min": 1500.0,
569
+ "max": 2550.0
570
+ },
571
+ {
572
+ "current": 2550.0,
573
+ "min": 1500.0,
574
+ "max": 2550.0
575
+ },
576
+ {
577
+ "current": 2550.0,
578
+ "min": 1500.0,
579
+ "max": 2550.0
580
+ },
581
+ {
582
+ "current": 2550.0,
583
+ "min": 1500.0,
584
+ "max": 2550.0
585
+ },
586
+ {
587
+ "current": 2550.0,
588
+ "min": 1500.0,
589
+ "max": 2550.0
590
+ },
591
+ {
592
+ "current": 2550.0,
593
+ "min": 1500.0,
594
+ "max": 2550.0
595
+ },
596
+ {
597
+ "current": 2550.0,
598
+ "min": 1500.0,
599
+ "max": 2550.0
600
+ },
601
+ {
602
+ "current": 2550.0,
603
+ "min": 1500.0,
604
+ "max": 2550.0
605
+ },
606
+ {
607
+ "current": 2550.0,
608
+ "min": 1500.0,
609
+ "max": 2550.0
610
+ },
611
+ {
612
+ "current": 2550.0,
613
+ "min": 1500.0,
614
+ "max": 2550.0
615
+ },
616
+ {
617
+ "current": 2550.0,
618
+ "min": 1500.0,
619
+ "max": 2550.0
620
+ },
621
+ {
622
+ "current": 2550.0,
623
+ "min": 1500.0,
624
+ "max": 2550.0
625
+ },
626
+ {
627
+ "current": 2550.0,
628
+ "min": 1500.0,
629
+ "max": 2550.0
630
+ },
631
+ {
632
+ "current": 2550.0,
633
+ "min": 1500.0,
634
+ "max": 2550.0
635
+ },
636
+ {
637
+ "current": 2550.0,
638
+ "min": 1500.0,
639
+ "max": 2550.0
640
+ },
641
+ {
642
+ "current": 2550.0,
643
+ "min": 1500.0,
644
+ "max": 2550.0
645
+ },
646
+ {
647
+ "current": 2550.0,
648
+ "min": 1500.0,
649
+ "max": 2550.0
650
+ },
651
+ {
652
+ "current": 2550.0,
653
+ "min": 1500.0,
654
+ "max": 2550.0
655
+ },
656
+ {
657
+ "current": 2550.0,
658
+ "min": 1500.0,
659
+ "max": 2550.0
660
+ },
661
+ {
662
+ "current": 2550.0,
663
+ "min": 1500.0,
664
+ "max": 2550.0
665
+ },
666
+ {
667
+ "current": 2550.0,
668
+ "min": 1500.0,
669
+ "max": 2550.0
670
+ },
671
+ {
672
+ "current": 2550.0,
673
+ "min": 1500.0,
674
+ "max": 2550.0
675
+ },
676
+ {
677
+ "current": 2550.0,
678
+ "min": 1500.0,
679
+ "max": 2550.0
680
+ },
681
+ {
682
+ "current": 2550.0,
683
+ "min": 1500.0,
684
+ "max": 2550.0
685
+ },
686
+ {
687
+ "current": 2550.0,
688
+ "min": 1500.0,
689
+ "max": 2550.0
690
+ },
691
+ {
692
+ "current": 2550.0,
693
+ "min": 1500.0,
694
+ "max": 2550.0
695
+ },
696
+ {
697
+ "current": 2550.0,
698
+ "min": 1500.0,
699
+ "max": 2550.0
700
+ },
701
+ {
702
+ "current": 2550.0,
703
+ "min": 1500.0,
704
+ "max": 2550.0
705
+ },
706
+ {
707
+ "current": 2550.0,
708
+ "min": 1500.0,
709
+ "max": 2550.0
710
+ },
711
+ {
712
+ "current": 2550.0,
713
+ "min": 1500.0,
714
+ "max": 2550.0
715
+ },
716
+ {
717
+ "current": 2550.0,
718
+ "min": 1500.0,
719
+ "max": 2550.0
720
+ },
721
+ {
722
+ "current": 2550.0,
723
+ "min": 1500.0,
724
+ "max": 2550.0
725
+ },
726
+ {
727
+ "current": 2550.0,
728
+ "min": 1500.0,
729
+ "max": 2550.0
730
+ },
731
+ {
732
+ "current": 2550.0,
733
+ "min": 1500.0,
734
+ "max": 2550.0
735
+ },
736
+ {
737
+ "current": 2550.0,
738
+ "min": 1500.0,
739
+ "max": 2550.0
740
+ },
741
+ {
742
+ "current": 3171.005,
743
+ "min": 1500.0,
744
+ "max": 2550.0
745
+ },
746
+ {
747
+ "current": 2550.0,
748
+ "min": 1500.0,
749
+ "max": 2550.0
750
+ },
751
+ {
752
+ "current": 3250.155,
753
+ "min": 1500.0,
754
+ "max": 2550.0
755
+ },
756
+ {
757
+ "current": 2550.0,
758
+ "min": 1500.0,
759
+ "max": 2550.0
760
+ },
761
+ {
762
+ "current": 2550.0,
763
+ "min": 1500.0,
764
+ "max": 2550.0
765
+ },
766
+ {
767
+ "current": 2550.0,
768
+ "min": 1500.0,
769
+ "max": 2550.0
770
+ },
771
+ {
772
+ "current": 2550.0,
773
+ "min": 1500.0,
774
+ "max": 2550.0
775
+ },
776
+ {
777
+ "current": 2550.0,
778
+ "min": 1500.0,
779
+ "max": 2550.0
780
+ },
781
+ {
782
+ "current": 3249.692,
783
+ "min": 1500.0,
784
+ "max": 2550.0
785
+ },
786
+ {
787
+ "current": 2550.0,
788
+ "min": 1500.0,
789
+ "max": 2550.0
790
+ },
791
+ {
792
+ "current": 2550.0,
793
+ "min": 1500.0,
794
+ "max": 2550.0
795
+ },
796
+ {
797
+ "current": 2550.0,
798
+ "min": 1500.0,
799
+ "max": 2550.0
800
+ },
801
+ {
802
+ "current": 2550.0,
803
+ "min": 1500.0,
804
+ "max": 2550.0
805
+ },
806
+ {
807
+ "current": 2550.0,
808
+ "min": 1500.0,
809
+ "max": 2550.0
810
+ },
811
+ {
812
+ "current": 2550.0,
813
+ "min": 1500.0,
814
+ "max": 2550.0
815
+ },
816
+ {
817
+ "current": 2550.0,
818
+ "min": 1500.0,
819
+ "max": 2550.0
820
+ },
821
+ {
822
+ "current": 2550.0,
823
+ "min": 1500.0,
824
+ "max": 2550.0
825
+ },
826
+ {
827
+ "current": 2550.0,
828
+ "min": 1500.0,
829
+ "max": 2550.0
830
+ },
831
+ {
832
+ "current": 2550.0,
833
+ "min": 1500.0,
834
+ "max": 2550.0
835
+ },
836
+ {
837
+ "current": 2550.0,
838
+ "min": 1500.0,
839
+ "max": 2550.0
840
+ },
841
+ {
842
+ "current": 2550.0,
843
+ "min": 1500.0,
844
+ "max": 2550.0
845
+ },
846
+ {
847
+ "current": 2550.0,
848
+ "min": 1500.0,
849
+ "max": 2550.0
850
+ },
851
+ {
852
+ "current": 3249.559,
853
+ "min": 1500.0,
854
+ "max": 2550.0
855
+ },
856
+ {
857
+ "current": 2550.0,
858
+ "min": 1500.0,
859
+ "max": 2550.0
860
+ },
861
+ {
862
+ "current": 2550.0,
863
+ "min": 1500.0,
864
+ "max": 2550.0
865
+ },
866
+ {
867
+ "current": 2550.0,
868
+ "min": 1500.0,
869
+ "max": 2550.0
870
+ },
871
+ {
872
+ "current": 2550.0,
873
+ "min": 1500.0,
874
+ "max": 2550.0
875
+ },
876
+ {
877
+ "current": 2550.0,
878
+ "min": 1500.0,
879
+ "max": 2550.0
880
+ },
881
+ {
882
+ "current": 2550.0,
883
+ "min": 1500.0,
884
+ "max": 2550.0
885
+ },
886
+ {
887
+ "current": 2550.0,
888
+ "min": 1500.0,
889
+ "max": 2550.0
890
+ },
891
+ {
892
+ "current": 2550.0,
893
+ "min": 1500.0,
894
+ "max": 2550.0
895
+ },
896
+ {
897
+ "current": 2550.0,
898
+ "min": 1500.0,
899
+ "max": 2550.0
900
+ },
901
+ {
902
+ "current": 2550.0,
903
+ "min": 1500.0,
904
+ "max": 2550.0
905
+ },
906
+ {
907
+ "current": 2550.0,
908
+ "min": 1500.0,
909
+ "max": 2550.0
910
+ },
911
+ {
912
+ "current": 2550.0,
913
+ "min": 1500.0,
914
+ "max": 2550.0
915
+ },
916
+ {
917
+ "current": 2550.0,
918
+ "min": 1500.0,
919
+ "max": 2550.0
920
+ },
921
+ {
922
+ "current": 2550.0,
923
+ "min": 1500.0,
924
+ "max": 2550.0
925
+ },
926
+ {
927
+ "current": 2550.0,
928
+ "min": 1500.0,
929
+ "max": 2550.0
930
+ },
931
+ {
932
+ "current": 2550.0,
933
+ "min": 1500.0,
934
+ "max": 2550.0
935
+ },
936
+ {
937
+ "current": 2550.0,
938
+ "min": 1500.0,
939
+ "max": 2550.0
940
+ },
941
+ {
942
+ "current": 2550.0,
943
+ "min": 1500.0,
944
+ "max": 2550.0
945
+ },
946
+ {
947
+ "current": 2550.0,
948
+ "min": 1500.0,
949
+ "max": 2550.0
950
+ },
951
+ {
952
+ "current": 2550.0,
953
+ "min": 1500.0,
954
+ "max": 2550.0
955
+ },
956
+ {
957
+ "current": 2550.0,
958
+ "min": 1500.0,
959
+ "max": 2550.0
960
+ },
961
+ {
962
+ "current": 2550.0,
963
+ "min": 1500.0,
964
+ "max": 2550.0
965
+ },
966
+ {
967
+ "current": 2550.0,
968
+ "min": 1500.0,
969
+ "max": 2550.0
970
+ },
971
+ {
972
+ "current": 2550.0,
973
+ "min": 1500.0,
974
+ "max": 2550.0
975
+ },
976
+ {
977
+ "current": 2550.0,
978
+ "min": 1500.0,
979
+ "max": 2550.0
980
+ },
981
+ {
982
+ "current": 2550.0,
983
+ "min": 1500.0,
984
+ "max": 2550.0
985
+ },
986
+ {
987
+ "current": 2550.0,
988
+ "min": 1500.0,
989
+ "max": 2550.0
990
+ },
991
+ {
992
+ "current": 2550.0,
993
+ "min": 1500.0,
994
+ "max": 2550.0
995
+ },
996
+ {
997
+ "current": 2550.0,
998
+ "min": 1500.0,
999
+ "max": 2550.0
1000
+ },
1001
+ {
1002
+ "current": 2550.0,
1003
+ "min": 1500.0,
1004
+ "max": 2550.0
1005
+ },
1006
+ {
1007
+ "current": 2550.0,
1008
+ "min": 1500.0,
1009
+ "max": 2550.0
1010
+ },
1011
+ {
1012
+ "current": 2550.0,
1013
+ "min": 1500.0,
1014
+ "max": 2550.0
1015
+ },
1016
+ {
1017
+ "current": 2550.0,
1018
+ "min": 1500.0,
1019
+ "max": 2550.0
1020
+ },
1021
+ {
1022
+ "current": 2550.0,
1023
+ "min": 1500.0,
1024
+ "max": 2550.0
1025
+ },
1026
+ {
1027
+ "current": 2550.0,
1028
+ "min": 1500.0,
1029
+ "max": 2550.0
1030
+ },
1031
+ {
1032
+ "current": 2550.0,
1033
+ "min": 1500.0,
1034
+ "max": 2550.0
1035
+ },
1036
+ {
1037
+ "current": 2550.0,
1038
+ "min": 1500.0,
1039
+ "max": 2550.0
1040
+ },
1041
+ {
1042
+ "current": 2550.0,
1043
+ "min": 1500.0,
1044
+ "max": 2550.0
1045
+ },
1046
+ {
1047
+ "current": 2550.0,
1048
+ "min": 1500.0,
1049
+ "max": 2550.0
1050
+ },
1051
+ {
1052
+ "current": 2550.0,
1053
+ "min": 1500.0,
1054
+ "max": 2550.0
1055
+ },
1056
+ {
1057
+ "current": 2550.0,
1058
+ "min": 1500.0,
1059
+ "max": 2550.0
1060
+ },
1061
+ {
1062
+ "current": 2550.0,
1063
+ "min": 1500.0,
1064
+ "max": 2550.0
1065
+ },
1066
+ {
1067
+ "current": 2550.0,
1068
+ "min": 1500.0,
1069
+ "max": 2550.0
1070
+ },
1071
+ {
1072
+ "current": 2550.0,
1073
+ "min": 1500.0,
1074
+ "max": 2550.0
1075
+ },
1076
+ {
1077
+ "current": 2550.0,
1078
+ "min": 1500.0,
1079
+ "max": 2550.0
1080
+ },
1081
+ {
1082
+ "current": 2550.0,
1083
+ "min": 1500.0,
1084
+ "max": 2550.0
1085
+ },
1086
+ {
1087
+ "current": 2550.0,
1088
+ "min": 1500.0,
1089
+ "max": 2550.0
1090
+ },
1091
+ {
1092
+ "current": 2550.0,
1093
+ "min": 1500.0,
1094
+ "max": 2550.0
1095
+ },
1096
+ {
1097
+ "current": 2550.0,
1098
+ "min": 1500.0,
1099
+ "max": 2550.0
1100
+ },
1101
+ {
1102
+ "current": 2550.0,
1103
+ "min": 1500.0,
1104
+ "max": 2550.0
1105
+ },
1106
+ {
1107
+ "current": 2550.0,
1108
+ "min": 1500.0,
1109
+ "max": 2550.0
1110
+ },
1111
+ {
1112
+ "current": 2550.0,
1113
+ "min": 1500.0,
1114
+ "max": 2550.0
1115
+ },
1116
+ {
1117
+ "current": 2550.0,
1118
+ "min": 1500.0,
1119
+ "max": 2550.0
1120
+ },
1121
+ {
1122
+ "current": 2550.0,
1123
+ "min": 1500.0,
1124
+ "max": 2550.0
1125
+ },
1126
+ {
1127
+ "current": 2550.0,
1128
+ "min": 1500.0,
1129
+ "max": 2550.0
1130
+ },
1131
+ {
1132
+ "current": 2550.0,
1133
+ "min": 1500.0,
1134
+ "max": 2550.0
1135
+ },
1136
+ {
1137
+ "current": 2550.0,
1138
+ "min": 1500.0,
1139
+ "max": 2550.0
1140
+ },
1141
+ {
1142
+ "current": 2550.0,
1143
+ "min": 1500.0,
1144
+ "max": 2550.0
1145
+ },
1146
+ {
1147
+ "current": 2550.0,
1148
+ "min": 1500.0,
1149
+ "max": 2550.0
1150
+ },
1151
+ {
1152
+ "current": 2550.0,
1153
+ "min": 1500.0,
1154
+ "max": 2550.0
1155
+ },
1156
+ {
1157
+ "current": 2550.0,
1158
+ "min": 1500.0,
1159
+ "max": 2550.0
1160
+ },
1161
+ {
1162
+ "current": 2550.0,
1163
+ "min": 1500.0,
1164
+ "max": 2550.0
1165
+ },
1166
+ {
1167
+ "current": 2550.0,
1168
+ "min": 1500.0,
1169
+ "max": 2550.0
1170
+ },
1171
+ {
1172
+ "current": 2550.0,
1173
+ "min": 1500.0,
1174
+ "max": 2550.0
1175
+ },
1176
+ {
1177
+ "current": 2550.0,
1178
+ "min": 1500.0,
1179
+ "max": 2550.0
1180
+ },
1181
+ {
1182
+ "current": 2550.0,
1183
+ "min": 1500.0,
1184
+ "max": 2550.0
1185
+ },
1186
+ {
1187
+ "current": 2550.0,
1188
+ "min": 1500.0,
1189
+ "max": 2550.0
1190
+ },
1191
+ {
1192
+ "current": 2550.0,
1193
+ "min": 1500.0,
1194
+ "max": 2550.0
1195
+ },
1196
+ {
1197
+ "current": 2550.0,
1198
+ "min": 1500.0,
1199
+ "max": 2550.0
1200
+ },
1201
+ {
1202
+ "current": 2550.0,
1203
+ "min": 1500.0,
1204
+ "max": 2550.0
1205
+ },
1206
+ {
1207
+ "current": 2550.0,
1208
+ "min": 1500.0,
1209
+ "max": 2550.0
1210
+ },
1211
+ {
1212
+ "current": 2550.0,
1213
+ "min": 1500.0,
1214
+ "max": 2550.0
1215
+ },
1216
+ {
1217
+ "current": 2550.0,
1218
+ "min": 1500.0,
1219
+ "max": 2550.0
1220
+ },
1221
+ {
1222
+ "current": 2550.0,
1223
+ "min": 1500.0,
1224
+ "max": 2550.0
1225
+ },
1226
+ {
1227
+ "current": 2550.0,
1228
+ "min": 1500.0,
1229
+ "max": 2550.0
1230
+ },
1231
+ {
1232
+ "current": 2550.0,
1233
+ "min": 1500.0,
1234
+ "max": 2550.0
1235
+ },
1236
+ {
1237
+ "current": 2550.0,
1238
+ "min": 1500.0,
1239
+ "max": 2550.0
1240
+ },
1241
+ {
1242
+ "current": 2550.0,
1243
+ "min": 1500.0,
1244
+ "max": 2550.0
1245
+ },
1246
+ {
1247
+ "current": 2550.0,
1248
+ "min": 1500.0,
1249
+ "max": 2550.0
1250
+ },
1251
+ {
1252
+ "current": 2550.0,
1253
+ "min": 1500.0,
1254
+ "max": 2550.0
1255
+ },
1256
+ {
1257
+ "current": 2550.0,
1258
+ "min": 1500.0,
1259
+ "max": 2550.0
1260
+ },
1261
+ {
1262
+ "current": 2550.0,
1263
+ "min": 1500.0,
1264
+ "max": 2550.0
1265
+ },
1266
+ {
1267
+ "current": 2550.0,
1268
+ "min": 1500.0,
1269
+ "max": 2550.0
1270
+ },
1271
+ {
1272
+ "current": 2550.0,
1273
+ "min": 1500.0,
1274
+ "max": 2550.0
1275
+ },
1276
+ {
1277
+ "current": 2550.0,
1278
+ "min": 1500.0,
1279
+ "max": 2550.0
1280
+ },
1281
+ {
1282
+ "current": 2550.0,
1283
+ "min": 1500.0,
1284
+ "max": 2550.0
1285
+ },
1286
+ {
1287
+ "current": 2550.0,
1288
+ "min": 1500.0,
1289
+ "max": 2550.0
1290
+ },
1291
+ {
1292
+ "current": 2550.0,
1293
+ "min": 1500.0,
1294
+ "max": 2550.0
1295
+ },
1296
+ {
1297
+ "current": 2550.0,
1298
+ "min": 1500.0,
1299
+ "max": 2550.0
1300
+ },
1301
+ {
1302
+ "current": 2550.0,
1303
+ "min": 1500.0,
1304
+ "max": 2550.0
1305
+ },
1306
+ {
1307
+ "current": 2550.0,
1308
+ "min": 1500.0,
1309
+ "max": 2550.0
1310
+ },
1311
+ {
1312
+ "current": 2550.0,
1313
+ "min": 1500.0,
1314
+ "max": 2550.0
1315
+ }
1316
+ ],
1317
+ "disk": {
1318
+ "/": {
1319
+ "total": 242.1086883544922,
1320
+ "used": 86.46047592163086
1321
+ }
1322
+ },
1323
+ "gpu": "NVIDIA A100-SXM4-80GB",
1324
+ "gpu_count": 8,
1325
+ "gpu_devices": [
1326
+ {
1327
+ "name": "NVIDIA A100-SXM4-80GB",
1328
+ "memory_total": 85899345920
1329
+ },
1330
+ {
1331
+ "name": "NVIDIA A100-SXM4-80GB",
1332
+ "memory_total": 85899345920
1333
+ },
1334
+ {
1335
+ "name": "NVIDIA A100-SXM4-80GB",
1336
+ "memory_total": 85899345920
1337
+ },
1338
+ {
1339
+ "name": "NVIDIA A100-SXM4-80GB",
1340
+ "memory_total": 85899345920
1341
+ },
1342
+ {
1343
+ "name": "NVIDIA A100-SXM4-80GB",
1344
+ "memory_total": 85899345920
1345
+ },
1346
+ {
1347
+ "name": "NVIDIA A100-SXM4-80GB",
1348
+ "memory_total": 85899345920
1349
+ },
1350
+ {
1351
+ "name": "NVIDIA A100-SXM4-80GB",
1352
+ "memory_total": 85899345920
1353
+ },
1354
+ {
1355
+ "name": "NVIDIA A100-SXM4-80GB",
1356
+ "memory_total": 85899345920
1357
+ }
1358
+ ],
1359
+ "memory": {
1360
+ "total": 2015.6942749023438
1361
+ }
1362
+ }
stage1_qformer2_128/wandb/run-20240313_070332-tkilnh6o/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"Align/Step": 2180, "Align/Loss": 3.234865188598633, "Align/Loss (Raw)": 3.2307181358337402, "Align/Learning Rate": 5.510720031964844e-10, "Align/Step Time": 1.6244877763092518, "_timestamp": 1710316925.9726973, "_runtime": 3513.4836943149567, "_step": 2180, "_wandb": {"runtime": 3550}}
stage1_qformer2_128/wandb/run-20240313_070332-tkilnh6o/logs/debug-internal.log ADDED
The diff for this file is too large to render. See raw diff
 
stage1_qformer2_128/wandb/run-20240313_070332-tkilnh6o/logs/debug.log ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-03-13 07:03:32,477 INFO MainThread:865933 [wandb_setup.py:_flush():76] Current SDK version is 0.16.3
2
+ 2024-03-13 07:03:32,477 INFO MainThread:865933 [wandb_setup.py:_flush():76] Configure stats pid to 865933
3
+ 2024-03-13 07:03:32,477 INFO MainThread:865933 [wandb_setup.py:_flush():76] Loading settings from /home/lilei/.config/wandb/settings
4
+ 2024-03-13 07:03:32,477 INFO MainThread:865933 [wandb_setup.py:_flush():76] Loading settings from /home/lilei/prismatic-vlms/wandb/settings
5
+ 2024-03-13 07:03:32,477 INFO MainThread:865933 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
6
+ 2024-03-13 07:03:32,477 INFO MainThread:865933 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
7
+ 2024-03-13 07:03:32,477 INFO MainThread:865933 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': 'scripts/pretrain.py', 'program_abspath': '/home/lilei/prismatic-vlms/scripts/pretrain.py', 'program': '/home/lilei/prismatic-vlms/scripts/pretrain.py'}
8
+ 2024-03-13 07:03:32,477 INFO MainThread:865933 [wandb_init.py:_log_setup():526] Logging user logs to runs/stage1_qformer2_128/wandb/run-20240313_070332-tkilnh6o/logs/debug.log
9
+ 2024-03-13 07:03:32,477 INFO MainThread:865933 [wandb_init.py:_log_setup():527] Logging internal logs to runs/stage1_qformer2_128/wandb/run-20240313_070332-tkilnh6o/logs/debug-internal.log
10
+ 2024-03-13 07:03:32,477 INFO MainThread:865933 [wandb_init.py:init():566] calling init triggers
11
+ 2024-03-13 07:03:32,477 INFO MainThread:865933 [wandb_init.py:init():573] wandb.init called with sweep_config: {}
12
+ config: {'model': {'type': 'one-stage+7b', 'model_id': 'one-stage+7b_qformer2_128', 'arch_specifier': 'qformer2_128', 'vision_backbone_id': 'clip-vit-l-336px', 'llm_backbone_id': 'vicuna-v15-7b', 'image_resize_strategy': 'letterbox', 'llm_max_length': 2048, 'align_epochs': 1, 'align_max_steps': None, 'align_global_batch_size': 256, 'align_per_device_batch_size': 16, 'align_learning_rate': 0.001, 'align_weight_decay': 0.0, 'align_max_grad_norm': 1.0, 'align_lr_scheduler_type': 'linear-warmup+cosine-decay', 'align_warmup_ratio': 0.03, 'align_train_strategy': 'fsdp-shard-grad-op', 'finetune_epochs': 1, 'finetune_max_steps': None, 'finetune_global_batch_size': 128, 'finetune_per_device_batch_size': 16, 'finetune_learning_rate': 2e-05, 'finetune_weight_decay': 0.1, 'finetune_max_grad_norm': 1.0, 'finetune_lr_scheduler_type': 'linear-warmup+cosine-decay', 'finetune_warmup_ratio': 0.03, 'finetune_train_strategy': 'fsdp-full-shard', 'enable_gradient_checkpointing': True, 'enable_mixed_precision_training': True, 'reduce_in_full_precision': False}, 'dataset': {'type': 'llava-v15', 'dataset_id': 'llava-v15', 'align_stage_components': ['download/llava-laion-cc-sbu-558k/chat.json', 'download/llava-laion-cc-sbu-558k'], 'finetune_stage_components': ['download/llava-v1.5-instruct/llava_v1_5_mix665k.json', 'download/llava-v1.5-instruct'], 'dataset_root_dir': 'data'}, 'stage': 'align', 'pretrained_checkpoint': None, 'run_id': 'stage1_qformer2_128', 'run_root_dir': 'runs', 'seed': 7, 'hf_token': '.hf_token', 'trackers': ['jsonl', 'wandb'], 'wandb_project': 'hf-vlms', 'wandb_entity': 'lilei_stones', 'max_length': 4096}
13
+ 2024-03-13 07:03:32,477 INFO MainThread:865933 [wandb_init.py:init():616] starting backend
14
+ 2024-03-13 07:03:32,477 INFO MainThread:865933 [wandb_init.py:init():620] setting up manager
15
+ 2024-03-13 07:03:32,483 INFO MainThread:865933 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
16
+ 2024-03-13 07:03:32,488 INFO MainThread:865933 [wandb_init.py:init():628] backend started and connected
17
+ 2024-03-13 07:03:32,500 INFO MainThread:865933 [wandb_init.py:init():720] updated telemetry
18
+ 2024-03-13 07:03:32,536 INFO MainThread:865933 [wandb_init.py:init():753] communicating run to backend with 90.0 second timeout
19
+ 2024-03-13 07:03:33,072 INFO MainThread:865933 [wandb_run.py:_on_init():2262] communicating current version
20
+ 2024-03-13 07:03:33,118 INFO MainThread:865933 [wandb_run.py:_on_init():2271] got version response upgrade_message: "wandb version 0.16.4 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
21
+
22
+ 2024-03-13 07:03:33,118 INFO MainThread:865933 [wandb_init.py:init():804] starting run threads in backend
23
+ 2024-03-13 07:03:39,026 INFO MainThread:865933 [wandb_run.py:_console_start():2241] atexit reg
24
+ 2024-03-13 07:03:39,027 INFO MainThread:865933 [wandb_run.py:_redirect():2096] redirect: wrap_raw
25
+ 2024-03-13 07:03:39,027 INFO MainThread:865933 [wandb_run.py:_redirect():2161] Wrapping output streams.
26
+ 2024-03-13 07:03:39,027 INFO MainThread:865933 [wandb_run.py:_redirect():2186] Redirects installed.
27
+ 2024-03-13 07:03:39,028 INFO MainThread:865933 [wandb_init.py:init():847] run started, returning control to user process
28
+ 2024-03-13 08:02:43,581 INFO MainThread:865933 [wandb_run.py:_finish():1970] finishing run lilei_stones/hf-vlms/tkilnh6o
29
+ 2024-03-13 08:02:43,582 INFO MainThread:865933 [wandb_run.py:_atexit_cleanup():2210] got exitcode: 0
30
+ 2024-03-13 08:02:43,582 INFO MainThread:865933 [wandb_run.py:_restore():2193] restore
31
+ 2024-03-13 08:02:43,582 INFO MainThread:865933 [wandb_run.py:_restore():2199] restore done
32
+ 2024-03-13 08:02:51,643 INFO MainThread:865933 [wandb_run.py:_footer_history_summary_info():3866] rendering history
33
+ 2024-03-13 08:02:51,644 INFO MainThread:865933 [wandb_run.py:_footer_history_summary_info():3898] rendering summary
34
+ 2024-03-13 08:02:51,654 INFO MainThread:865933 [wandb_run.py:_footer_sync_info():3825] logging synced files
stage1_qformer2_128/wandb/run-20240313_070332-tkilnh6o/run-tkilnh6o.wandb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3ed617d9beec590563fe526d4ad71b0bbd244eac491b537325b491e961a34644
3
+ size 2532113
stage1_qformer2_256/checkpoints/latest-checkpoint.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2fdb8b9007a62c917dfd14cc345910d2325d5d5a1468c834eb32858e371b0b4b
3
+ size 1868636336
stage1_qformer2_256/checkpoints/step-002180-epoch-00-loss=2.5032.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2fdb8b9007a62c917dfd14cc345910d2325d5d5a1468c834eb32858e371b0b4b
3
+ size 1868636336
stage1_qformer2_256/config.json ADDED
@@ -0,0 +1,60 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset": {
3
+ "align_stage_components": [
4
+ "download/llava-laion-cc-sbu-558k/chat.json",
5
+ "download/llava-laion-cc-sbu-558k"
6
+ ],
7
+ "dataset_id": "llava-v15",
8
+ "dataset_root_dir": "data",
9
+ "finetune_stage_components": [
10
+ "download/llava-v1.5-instruct/llava_v1_5_mix665k.json",
11
+ "download/llava-v1.5-instruct"
12
+ ],
13
+ "type": "llava-v15"
14
+ },
15
+ "hf_token": ".hf_token",
16
+ "max_length": 4096,
17
+ "model": {
18
+ "align_epochs": 1,
19
+ "align_global_batch_size": 256,
20
+ "align_learning_rate": 0.001,
21
+ "align_lr_scheduler_type": "linear-warmup+cosine-decay",
22
+ "align_max_grad_norm": 1.0,
23
+ "align_max_steps": null,
24
+ "align_per_device_batch_size": 16,
25
+ "align_train_strategy": "fsdp-shard-grad-op",
26
+ "align_warmup_ratio": 0.03,
27
+ "align_weight_decay": 0.0,
28
+ "arch_specifier": "qformer2_256",
29
+ "enable_gradient_checkpointing": true,
30
+ "enable_mixed_precision_training": true,
31
+ "finetune_epochs": 1,
32
+ "finetune_global_batch_size": 128,
33
+ "finetune_learning_rate": 2e-05,
34
+ "finetune_lr_scheduler_type": "linear-warmup+cosine-decay",
35
+ "finetune_max_grad_norm": 1.0,
36
+ "finetune_max_steps": null,
37
+ "finetune_per_device_batch_size": 16,
38
+ "finetune_train_strategy": "fsdp-full-shard",
39
+ "finetune_warmup_ratio": 0.03,
40
+ "finetune_weight_decay": 0.1,
41
+ "image_resize_strategy": "letterbox",
42
+ "llm_backbone_id": "vicuna-v15-7b",
43
+ "llm_max_length": 2048,
44
+ "model_id": "one-stage+7b_qformer2_256",
45
+ "reduce_in_full_precision": false,
46
+ "type": "one-stage+7b",
47
+ "vision_backbone_id": "clip-vit-l-336px"
48
+ },
49
+ "pretrained_checkpoint": null,
50
+ "run_id": "stage1_qformer2_256",
51
+ "run_root_dir": "runs",
52
+ "seed": 7,
53
+ "stage": "align",
54
+ "trackers": [
55
+ "jsonl",
56
+ "wandb"
57
+ ],
58
+ "wandb_entity": "lilei_stones",
59
+ "wandb_project": "hf-vlms"
60
+ }
stage1_qformer2_256/config.yaml ADDED
@@ -0,0 +1,53 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ dataset:
2
+ align_stage_components:
3
+ - download/llava-laion-cc-sbu-558k/chat.json
4
+ - download/llava-laion-cc-sbu-558k
5
+ dataset_id: llava-v15
6
+ dataset_root_dir: data
7
+ finetune_stage_components:
8
+ - download/llava-v1.5-instruct/llava_v1_5_mix665k.json
9
+ - download/llava-v1.5-instruct
10
+ type: llava-v15
11
+ hf_token: .hf_token
12
+ max_length: 4096
13
+ model:
14
+ align_epochs: 1
15
+ align_global_batch_size: 256
16
+ align_learning_rate: 0.001
17
+ align_lr_scheduler_type: linear-warmup+cosine-decay
18
+ align_max_grad_norm: 1.0
19
+ align_max_steps: null
20
+ align_per_device_batch_size: 16
21
+ align_train_strategy: fsdp-shard-grad-op
22
+ align_warmup_ratio: 0.03
23
+ align_weight_decay: 0.0
24
+ arch_specifier: qformer2_256
25
+ enable_gradient_checkpointing: true
26
+ enable_mixed_precision_training: true
27
+ finetune_epochs: 1
28
+ finetune_global_batch_size: 128
29
+ finetune_learning_rate: 2.0e-05
30
+ finetune_lr_scheduler_type: linear-warmup+cosine-decay
31
+ finetune_max_grad_norm: 1.0
32
+ finetune_max_steps: null
33
+ finetune_per_device_batch_size: 16
34
+ finetune_train_strategy: fsdp-full-shard
35
+ finetune_warmup_ratio: 0.03
36
+ finetune_weight_decay: 0.1
37
+ image_resize_strategy: letterbox
38
+ llm_backbone_id: vicuna-v15-7b
39
+ llm_max_length: 2048
40
+ model_id: one-stage+7b_qformer2_256
41
+ reduce_in_full_precision: false
42
+ type: one-stage+7b
43
+ vision_backbone_id: clip-vit-l-336px
44
+ pretrained_checkpoint: null
45
+ run_id: stage1_qformer2_256
46
+ run_root_dir: runs
47
+ seed: 7
48
+ stage: align
49
+ trackers:
50
+ - jsonl
51
+ - wandb
52
+ wandb_entity: lilei_stones
53
+ wandb_project: hf-vlms
stage1_qformer2_256/run-metrics.jsonl ADDED
@@ -0,0 +1 @@
 
 
1
+ {"hparams": {"dataset": {"align_stage_components": ["download/llava-laion-cc-sbu-558k/chat.json", "download/llava-laion-cc-sbu-558k"], "dataset_id": "llava-v15", "dataset_root_dir": "data", "finetune_stage_components": ["download/llava-v1.5-instruct/llava_v1_5_mix665k.json", "download/llava-v1.5-instruct"], "type": "llava-v15"}, "hf_token": ".hf_token", "max_length": 4096, "model": {"align_epochs": 1, "align_global_batch_size": 256, "align_learning_rate": 0.001, "align_lr_scheduler_type": "linear-warmup+cosine-decay", "align_max_grad_norm": 1.0, "align_max_steps": null, "align_per_device_batch_size": 16, "align_train_strategy": "fsdp-shard-grad-op", "align_warmup_ratio": 0.03, "align_weight_decay": 0.0, "arch_specifier": "qformer2_256", "enable_gradient_checkpointing": true, "enable_mixed_precision_training": true, "finetune_epochs": 1, "finetune_global_batch_size": 128, "finetune_learning_rate": 2e-05, "finetune_lr_scheduler_type": "linear-warmup+cosine-decay", "finetune_max_grad_norm": 1.0, "finetune_max_steps": null, "finetune_per_device_batch_size": 16, "finetune_train_strategy": "fsdp-full-shard", "finetune_warmup_ratio": 0.03, "finetune_weight_decay": 0.1, "image_resize_strategy": "letterbox", "llm_backbone_id": "vicuna-v15-7b", "llm_max_length": 2048, "model_id": "one-stage+7b_qformer2_256", "reduce_in_full_precision": false, "type": "one-stage+7b", "vision_backbone_id": "clip-vit-l-336px"}, "pretrained_checkpoint": null, "run_id": "stage1_qformer2_256", "run_root_dir": "runs", "seed": 7, "stage": "align", "trackers": ["jsonl", "wandb"], "wandb_entity": "lilei_stones", "wandb_project": "hf-vlms"}, "run_id": "stage1_qformer2_256"}
stage1_qformer2_256/stage1_qformer2_256.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
stage1_qformer2_256/wandb/debug-internal.log ADDED
@@ -0,0 +1 @@
 
 
1
+ run-20240313_070329-pog6cg3f/logs/debug-internal.log
stage1_qformer2_256/wandb/debug.log ADDED
@@ -0,0 +1 @@
 
 
1
+ run-20240313_070329-pog6cg3f/logs/debug.log
stage1_qformer2_256/wandb/latest-run ADDED
@@ -0,0 +1 @@
 
 
1
+ run-20240313_070329-pog6cg3f
stage1_qformer2_256/wandb/run-20240313_070329-pog6cg3f/files/conda-environment.yaml ADDED
@@ -0,0 +1,194 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ name: pvlm
2
+ channels:
3
+ - pytorch
4
+ - nvidia
5
+ - defaults
6
+ dependencies:
7
+ - _libgcc_mutex=0.1=main
8
+ - _openmp_mutex=5.1=1_gnu
9
+ - blas=1.0=mkl
10
+ - bzip2=1.0.8=h7b6447c_0
11
+ - ca-certificates=2023.12.12=h06a4308_0
12
+ - certifi=2024.2.2=py310h06a4308_0
13
+ - charset-normalizer=2.0.4=pyhd3eb1b0_0
14
+ - cuda-cudart=11.8.89=0
15
+ - cuda-cupti=11.8.87=0
16
+ - cuda-libraries=11.8.0=0
17
+ - cuda-nvrtc=11.8.89=0
18
+ - cuda-nvtx=11.8.86=0
19
+ - cuda-runtime=11.8.0=0
20
+ - ffmpeg=4.3=hf484d3e_0
21
+ - filelock=3.13.1=py310h06a4308_0
22
+ - freetype=2.12.1=h4a9f257_0
23
+ - gmp=6.2.1=h295c915_3
24
+ - gmpy2=2.1.2=py310heeb90bb_0
25
+ - gnutls=3.6.15=he1e5248_0
26
+ - idna=3.4=py310h06a4308_0
27
+ - intel-openmp=2023.1.0=hdb19cb5_46306
28
+ - jinja2=3.1.3=py310h06a4308_0
29
+ - jpeg=9e=h5eee18b_1
30
+ - lame=3.100=h7b6447c_0
31
+ - lcms2=2.12=h3be6417_0
32
+ - ld_impl_linux-64=2.38=h1181459_1
33
+ - lerc=3.0=h295c915_0
34
+ - libcublas=11.11.3.6=0
35
+ - libcufft=10.9.0.58=0
36
+ - libcufile=1.8.1.2=0
37
+ - libcurand=10.3.4.107=0
38
+ - libcusolver=11.4.1.48=0
39
+ - libcusparse=11.7.5.86=0
40
+ - libdeflate=1.17=h5eee18b_1
41
+ - libffi=3.4.4=h6a678d5_0
42
+ - libgcc-ng=11.2.0=h1234567_1
43
+ - libgomp=11.2.0=h1234567_1
44
+ - libiconv=1.16=h7f8727e_2
45
+ - libidn2=2.3.4=h5eee18b_0
46
+ - libjpeg-turbo=2.0.0=h9bf148f_0
47
+ - libnpp=11.8.0.86=0
48
+ - libnvjpeg=11.9.0.86=0
49
+ - libpng=1.6.39=h5eee18b_0
50
+ - libstdcxx-ng=11.2.0=h1234567_1
51
+ - libtasn1=4.19.0=h5eee18b_0
52
+ - libtiff=4.5.1=h6a678d5_0
53
+ - libunistring=0.9.10=h27cfd23_0
54
+ - libuuid=1.41.5=h5eee18b_0
55
+ - libwebp-base=1.3.2=h5eee18b_0
56
+ - llvm-openmp=14.0.6=h9e868ea_0
57
+ - lz4-c=1.9.4=h6a678d5_0
58
+ - markupsafe=2.1.3=py310h5eee18b_0
59
+ - mkl=2023.1.0=h213fc3f_46344
60
+ - mkl-service=2.4.0=py310h5eee18b_1
61
+ - mkl_fft=1.3.8=py310h5eee18b_0
62
+ - mkl_random=1.2.4=py310hdb19cb5_0
63
+ - mpc=1.1.0=h10f8cd9_1
64
+ - mpfr=4.0.2=hb69a4c5_1
65
+ - mpmath=1.3.0=py310h06a4308_0
66
+ - ncurses=6.4=h6a678d5_0
67
+ - nettle=3.7.3=hbbd107a_1
68
+ - networkx=3.1=py310h06a4308_0
69
+ - numpy=1.26.4=py310h5f9d8c6_0
70
+ - numpy-base=1.26.4=py310hb5e798b_0
71
+ - openh264=2.1.1=h4ff587b_0
72
+ - openjpeg=2.4.0=h3ad879b_0
73
+ - openssl=3.0.13=h7f8727e_0
74
+ - pillow=10.2.0=py310h5eee18b_0
75
+ - pip=23.3.1=py310h06a4308_0
76
+ - python=3.10.13=h955ad1f_0
77
+ - pytorch=2.2.1=py3.10_cuda11.8_cudnn8.7.0_0
78
+ - pytorch-cuda=11.8=h7e8668a_5
79
+ - pytorch-mutex=1.0=cuda
80
+ - pyyaml=6.0.1=py310h5eee18b_0
81
+ - readline=8.2=h5eee18b_0
82
+ - requests=2.31.0=py310h06a4308_1
83
+ - setuptools=68.2.2=py310h06a4308_0
84
+ - sqlite=3.41.2=h5eee18b_0
85
+ - sympy=1.12=py310h06a4308_0
86
+ - tbb=2021.8.0=hdb19cb5_0
87
+ - tk=8.6.12=h1ccaba5_0
88
+ - torchaudio=2.2.1=py310_cu118
89
+ - torchtriton=2.2.0=py310
90
+ - torchvision=0.17.1=py310_cu118
91
+ - typing_extensions=4.9.0=py310h06a4308_1
92
+ - urllib3=2.1.0=py310h06a4308_0
93
+ - wheel=0.41.2=py310h06a4308_0
94
+ - xz=5.4.5=h5eee18b_0
95
+ - yaml=0.2.5=h7b6447c_0
96
+ - zlib=1.2.13=h5eee18b_0
97
+ - zstd=1.5.5=hc292b87_0
98
+ - pip:
99
+ - accelerate==0.27.2
100
+ - aiohttp==3.9.3
101
+ - aiosignal==1.3.1
102
+ - annotated-types==0.6.0
103
+ - appdirs==1.4.4
104
+ - async-timeout==4.0.3
105
+ - attrs==23.2.0
106
+ - beautifulsoup4==4.12.3
107
+ - blis==0.7.11
108
+ - catalogue==2.0.10
109
+ - click==8.1.7
110
+ - cloudpathlib==0.16.0
111
+ - confection==0.1.4
112
+ - contourpy==1.2.0
113
+ - cycler==0.12.1
114
+ - cymem==2.0.8
115
+ - datasets==2.18.0
116
+ - deepspeed==0.13.4
117
+ - dill==0.3.8
118
+ - docker-pycreds==0.4.0
119
+ - draccus==0.7.1
120
+ - einops==0.7.0
121
+ - flash-attn==2.5.5
122
+ - fonttools==4.49.0
123
+ - frozenlist==1.4.1
124
+ - fsspec==2024.2.0
125
+ - gdown==5.1.0
126
+ - gitdb==4.0.11
127
+ - gitpython==3.1.42
128
+ - hjson==3.1.0
129
+ - huggingface-hub==0.20.3
130
+ - install==1.3.5
131
+ - jsonlines==4.0.0
132
+ - kiwisolver==1.4.5
133
+ - langcodes==3.3.0
134
+ - markdown-it-py==3.0.0
135
+ - matplotlib==3.8.3
136
+ - mdurl==0.1.2
137
+ - mergedeep==1.3.4
138
+ - multidict==6.0.5
139
+ - multiprocess==0.70.16
140
+ - murmurhash==1.0.10
141
+ - mypy-extensions==1.0.0
142
+ - ninja==1.11.1.1
143
+ - packaging==23.2
144
+ - pandas==2.2.1
145
+ - preshed==3.0.9
146
+ - prismatic==0.0.1
147
+ - protobuf==4.25.3
148
+ - psutil==5.9.8
149
+ - py-cpuinfo==9.0.0
150
+ - pyarrow==15.0.0
151
+ - pyarrow-hotfix==0.6
152
+ - pycocoevalcap==1.2
153
+ - pycocotools==2.0.7
154
+ - pydantic==2.6.3
155
+ - pydantic-core==2.16.3
156
+ - pygments==2.17.2
157
+ - pynvml==11.5.0
158
+ - pyparsing==3.1.1
159
+ - pysocks==1.7.1
160
+ - pythainlp==5.0.1
161
+ - python-dateutil==2.9.0.post0
162
+ - pytz==2024.1
163
+ - pyyaml-include==1.3.2
164
+ - regex==2023.12.25
165
+ - rich==13.7.0
166
+ - safetensors==0.4.2
167
+ - sentencepiece==0.2.0
168
+ - sentry-sdk==1.40.5
169
+ - setproctitle==1.3.3
170
+ - six==1.16.0
171
+ - smart-open==6.4.0
172
+ - smmap==5.0.1
173
+ - soupsieve==2.5
174
+ - spacy==3.7.4
175
+ - spacy-legacy==3.0.12
176
+ - spacy-loggers==1.0.5
177
+ - srsly==2.4.8
178
+ - sudachidict-core==20240109
179
+ - sudachipy==0.6.8
180
+ - thinc==8.2.3
181
+ - timm==0.9.16
182
+ - tokenizers==0.15.2
183
+ - toml==0.10.2
184
+ - tqdm==4.66.2
185
+ - transformers==4.38.1
186
+ - typer==0.9.0
187
+ - typing-inspect==0.9.0
188
+ - tzdata==2024.1
189
+ - wandb==0.16.3
190
+ - wasabi==1.1.2
191
+ - weasel==0.3.4
192
+ - xxhash==3.4.1
193
+ - yarl==1.9.4
194
+ prefix: /home/lilei/anaconda3/envs/pvlm
stage1_qformer2_256/wandb/run-20240313_070329-pog6cg3f/files/config.yaml ADDED
@@ -0,0 +1,119 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wandb_version: 1
2
+
3
+ model:
4
+ desc: null
5
+ value:
6
+ type: one-stage+7b
7
+ model_id: one-stage+7b_qformer2_256
8
+ arch_specifier: qformer2_256
9
+ vision_backbone_id: clip-vit-l-336px
10
+ llm_backbone_id: vicuna-v15-7b
11
+ image_resize_strategy: letterbox
12
+ llm_max_length: 2048
13
+ align_epochs: 1
14
+ align_max_steps: null
15
+ align_global_batch_size: 256
16
+ align_per_device_batch_size: 16
17
+ align_learning_rate: 0.001
18
+ align_weight_decay: 0.0
19
+ align_max_grad_norm: 1.0
20
+ align_lr_scheduler_type: linear-warmup+cosine-decay
21
+ align_warmup_ratio: 0.03
22
+ align_train_strategy: fsdp-shard-grad-op
23
+ finetune_epochs: 1
24
+ finetune_max_steps: null
25
+ finetune_global_batch_size: 128
26
+ finetune_per_device_batch_size: 16
27
+ finetune_learning_rate: 2.0e-05
28
+ finetune_weight_decay: 0.1
29
+ finetune_max_grad_norm: 1.0
30
+ finetune_lr_scheduler_type: linear-warmup+cosine-decay
31
+ finetune_warmup_ratio: 0.03
32
+ finetune_train_strategy: fsdp-full-shard
33
+ enable_gradient_checkpointing: true
34
+ enable_mixed_precision_training: true
35
+ reduce_in_full_precision: false
36
+ dataset:
37
+ desc: null
38
+ value:
39
+ type: llava-v15
40
+ dataset_id: llava-v15
41
+ align_stage_components:
42
+ - download/llava-laion-cc-sbu-558k/chat.json
43
+ - download/llava-laion-cc-sbu-558k
44
+ finetune_stage_components:
45
+ - download/llava-v1.5-instruct/llava_v1_5_mix665k.json
46
+ - download/llava-v1.5-instruct
47
+ dataset_root_dir: data
48
+ stage:
49
+ desc: null
50
+ value: align
51
+ pretrained_checkpoint:
52
+ desc: null
53
+ value: null
54
+ run_id:
55
+ desc: null
56
+ value: stage1_qformer2_256
57
+ run_root_dir:
58
+ desc: null
59
+ value: runs
60
+ seed:
61
+ desc: null
62
+ value: 7
63
+ hf_token:
64
+ desc: null
65
+ value: .hf_token
66
+ trackers:
67
+ desc: null
68
+ value:
69
+ - jsonl
70
+ - wandb
71
+ wandb_project:
72
+ desc: null
73
+ value: hf-vlms
74
+ wandb_entity:
75
+ desc: null
76
+ value: lilei_stones
77
+ max_length:
78
+ desc: null
79
+ value: 4096
80
+ _wandb:
81
+ desc: null
82
+ value:
83
+ python_version: 3.10.13
84
+ cli_version: 0.16.3
85
+ framework: huggingface
86
+ huggingface_version: 4.38.1
87
+ is_jupyter_run: false
88
+ is_kaggle_kernel: false
89
+ start_time: 1710313409.826913
90
+ t:
91
+ 1:
92
+ - 1
93
+ - 11
94
+ - 41
95
+ - 49
96
+ - 51
97
+ - 55
98
+ - 63
99
+ - 71
100
+ 2:
101
+ - 1
102
+ - 11
103
+ - 41
104
+ - 49
105
+ - 51
106
+ - 55
107
+ - 63
108
+ - 71
109
+ 3:
110
+ - 2
111
+ - 13
112
+ - 16
113
+ - 23
114
+ 4: 3.10.13
115
+ 5: 0.16.3
116
+ 6: 4.38.1
117
+ 8:
118
+ - 5
119
+ 13: linux-x86_64
stage1_qformer2_256/wandb/run-20240313_070329-pog6cg3f/files/output.log ADDED
@@ -0,0 +1,2180 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ 03/13 [07:03:36] INFO  | >> [*] Starting Training Loop ]8;id=72103;file:///home/lilei/prismatic-vlms/scripts/pretrain.py\pretrain.py]8;;\:]8;id=880770;file:///home/lilei/prismatic-vlms/scripts/pretrain.py#226\226]8;;\
3
+
4
+
5
+
6
+
7
+
8
+
9
+
10
+
11
+
12
+
13
+
14
+
15
+
16
+
17
+
18
+
19
+
20
+
21
+
22
+
23
+
24
+
25
+
26
+
27
+
28
+
29
+
30
+
31
+
32
+
33
+
34
+
35
+
36
+
37
+
38
+
39
+
40
+
41
+
42
+
43
+
44
+
45
+
46
+
47
+
48
+
49
+
50
+
51
+
52
+
53
+
54
+
55
+
56
+
57
+
58
+
59
+
60
+
61
+
62
+
63
+
64
+
65
+
66
+
67
+
68
+
69
+
70
+
71
+
72
+
73
+
74
+
75
+
76
+
77
+
78
+
79
+
80
+
81
+
82
+
83
+
84
+
85
+
86
+
87
+
88
+
89
+
90
+
91
+
92
+
93
+
94
+
95
+
96
+
97
+
98
+
99
+
100
+
101
+
102
+
103
+
104
+
105
+
106
+
107
+
108
+
109
+
110
+
111
+
112
+
113
+
114
+
115
+
116
+
117
+
118
+
119
+
120
+
121
+
122
+
123
+
124
+
125
+
126
+
127
+
128
+
129
+
130
+
131
+
132
+
133
+
134
+
135
+
136
+
137
+
138
+
139
+
140
+
141
+
142
+
143
+
144
+
145
+
146
+
147
+
148
+
149
+
150
+
151
+
152
+
153
+
154
+
155
+
156
+
157
+
158
+
159
+
160
+
161
+
162
+
163
+
164
+
165
+
166
+
167
+
168
+
169
+
170
+
171
+
172
+
173
+
174
+
175
+
176
+
177
+
178
+
179
+
180
+
181
+
182
+
183
+
184
+
185
+
186
+
187
+
188
+
189
+
190
+
191
+
192
+
193
+
194
+
195
+
196
+
197
+
198
+
199
+
200
+
201
+
202
+
203
+
204
+
205
+
206
+
207
+
208
+
209
+
210
+
211
+
212
+
213
+
214
+
215
+
216
+
217
+
218
+
219
+
220
+
221
+
222
+
223
+
224
+
225
+
226
+
227
+
228
+
229
+
230
+
231
+
232
+
233
+
234
+
235
+
236
+
237
+
238
+
239
+
240
+
241
+
242
+
243
+
244
+
245
+
246
+
247
+
248
+
249
+
250
+
251
+
252
+
253
+
254
+
255
+
256
+
257
+
258
+
259
+
260
+
261
+
262
+
263
+
264
+
265
+
266
+
267
+
268
+
269
+
270
+
271
+
272
+
273
+
274
+
275
+
276
+
277
+
278
+
279
+
280
+
281
+
282
+
283
+
284
+
285
+
286
+
287
+
288
+
289
+
290
+
291
+
292
+
293
+
294
+
295
+
296
+
297
+
298
+
299
+
300
+
301
+
302
+
303
+
304
+
305
+
306
+
307
+
308
+
309
+
310
+
311
+
312
+
313
+
314
+
315
+
316
+
317
+
318
+
319
+
320
+
321
+
322
+
323
+
324
+
325
+
326
+
327
+
328
+
329
+
330
+
331
+
332
+
333
+
334
+
335
+
336
+
337
+
338
+
339
+
340
+
341
+
342
+
343
+
344
+
345
+
346
+
347
+
348
+
349
+
350
+
351
+
352
+
353
+
354
+
355
+
356
+
357
+
358
+
359
+
360
+
361
+
362
+
363
+
364
+
365
+
366
+
367
+
368
+
369
+
370
+
371
+
372
+
373
+
374
+
375
+
376
+
377
+
378
+
379
+
380
+
381
+
382
+
383
+
384
+
385
+
386
+
387
+
388
+
389
+
390
+
391
+
392
+
393
+
394
+
395
+
396
+
397
+
398
+
399
+
400
+
401
+
402
+
403
+
404
+
405
+
406
+
407
+
408
+
409
+
410
+
411
+
412
+
413
+
414
+
415
+
416
+
417
+
418
+
419
+
420
+
421
+
422
+
423
+
424
+
425
+
426
+
427
+
428
+
429
+
430
+
431
+
432
+
433
+
434
+
435
+
436
+
437
+
438
+
439
+
440
+
441
+
442
+
443
+
444
+
445
+
446
+
447
+
448
+
449
+
450
+
451
+
452
+
453
+
454
+
455
+
456
+
457
+
458
+
459
+
460
+
461
+
462
+
463
+
464
+
465
+
466
+
467
+
468
+
469
+
470
+
471
+
472
+
473
+
474
+
475
+
476
+
477
+
478
+
479
+
480
+
481
+
482
+
483
+
484
+
485
+
486
+
487
+
488
+
489
+
490
+
491
+
492
+
493
+
494
+
495
+
496
+
497
+
498
+
499
+
500
+
501
+
502
+
503
+
504
+
505
+
506
+
507
+
508
+
509
+
510
+
511
+
512
+
513
+
514
+
515
+
516
+
517
+
518
+
519
+
520
+
521
+
522
+
523
+
524
+
525
+
526
+
527
+
528
+
529
+
530
+
531
+
532
+
533
+
534
+
535
+
536
+
537
+
538
+
539
+
540
+
541
+
542
+
543
+
544
+
545
+
546
+
547
+
548
+
549
+
550
+
551
+
552
+
553
+
554
+
555
+
556
+
557
+
558
+
559
+
560
+
561
+
562
+
563
+
564
+
565
+
566
+
567
+
568
+
569
+
570
+
571
+
572
+
573
+
574
+
575
+
576
+
577
+
578
+
579
+
580
+
581
+
582
+
583
+
584
+
585
+
586
+
587
+
588
+
589
+
590
+
591
+
592
+
593
+
594
+
595
+
596
+
597
+
598
+
599
+
600
+
601
+
602
+
603
+
604
+
605
+
606
+
607
+
608
+
609
+
610
+
611
+
612
+
613
+
614
+
615
+
616
+
617
+
618
+
619
+
620
+
621
+
622
+
623
+
624
+
625
+
626
+
627
+
628
+
629
+
630
+
631
+
632
+
633
+
634
+
635
+
636
+
637
+
638
+
639
+
640
+
641
+
642
+
643
+
644
+
645
+
646
+
647
+
648
+
649
+
650
+
651
+
652
+
653
+
654
+
655
+
656
+
657
+
658
+
659
+
660
+
661
+
662
+
663
+
664
+
665
+
666
+
667
+
668
+
669
+
670
+
671
+
672
+
673
+
674
+
675
+
676
+
677
+
678
+
679
+
680
+
681
+
682
+
683
+
684
+
685
+
686
+
687
+
688
+
689
+
690
+
691
+
692
+
693
+
694
+
695
+
696
+
697
+
698
+
699
+
700
+
701
+
702
+
703
+
704
+
705
+
706
+
707
+
708
+
709
+
710
+
711
+
712
+
713
+
714
+
715
+
716
+
717
+
718
+
719
+
720
+
721
+
722
+
723
+
724
+
725
+
726
+
727
+
728
+
729
+
730
+
731
+
732
+
733
+
734
+
735
+
736
+
737
+
738
+
739
+
740
+
741
+
742
+
743
+
744
+
745
+
746
+
747
+
748
+
749
+
750
+
751
+
752
+
753
+
754
+
755
+
756
+
757
+
758
+
759
+
760
+
761
+
762
+
763
+
764
+
765
+
766
+
767
+
768
+
769
+
770
+
771
+
772
+
773
+
774
+
775
+
776
+
777
+
778
+
779
+
780
+
781
+
782
+
783
+
784
+
785
+
786
+
787
+
788
+
789
+
790
+
791
+
792
+
793
+
794
+
795
+
796
+
797
+
798
+
799
+
800
+
801
+
802
+
803
+
804
+
805
+
806
+
807
+
808
+
809
+
810
+
811
+
812
+
813
+
814
+
815
+
816
+
817
+
818
+
819
+
820
+
821
+
822
+
823
+
824
+
825
+
826
+
827
+
828
+
829
+
830
+
831
+
832
+
833
+
834
+
835
+
836
+
837
+
838
+
839
+
840
+
841
+
842
+
843
+
844
+
845
+
846
+
847
+
848
+
849
+
850
+
851
+
852
+
853
+
854
+
855
+
856
+
857
+
858
+
859
+
860
+
861
+
862
+
863
+
864
+
865
+
866
+
867
+
868
+
869
+
870
+
871
+
872
+
873
+
874
+
875
+
876
+
877
+
878
+
879
+
880
+
881
+
882
+
883
+
884
+
885
+
886
+
887
+
888
+
889
+
890
+
891
+
892
+
893
+
894
+
895
+
896
+
897
+
898
+
899
+
900
+
901
+
902
+
903
+
904
+
905
+
906
+
907
+
908
+
909
+
910
+
911
+
912
+
913
+
914
+
915
+
916
+
917
+
918
+
919
+
920
+
921
+
922
+
923
+
924
+
925
+
926
+
927
+
928
+
929
+
930
+
931
+
932
+
933
+
934
+
935
+
936
+
937
+
938
+
939
+
940
+
941
+
942
+
943
+
944
+
945
+
946
+
947
+
948
+
949
+
950
+
951
+
952
+
953
+
954
+
955
+
956
+
957
+
958
+
959
+
960
+
961
+
962
+
963
+
964
+
965
+
966
+
967
+
968
+
969
+
970
+
971
+
972
+
973
+
974
+
975
+
976
+
977
+
978
+
979
+
980
+
981
+
982
+
983
+
984
+
985
+
986
+
987
+
988
+
989
+
990
+
991
+
992
+
993
+
994
+
995
+
996
+
997
+
998
+
999
+
1000
+
1001
+
1002
+
1003
+
1004
+
1005
+
1006
+
1007
+
1008
+
1009
+
1010
+
1011
+
1012
+
1013
+
1014
+
1015
+
1016
+
1017
+
1018
+
1019
+
1020
+
1021
+
1022
+
1023
+
1024
+
1025
+
1026
+
1027
+
1028
+
1029
+
1030
+
1031
+
1032
+
1033
+
1034
+
1035
+
1036
+
1037
+
1038
+
1039
+
1040
+
1041
+
1042
+
1043
+
1044
+
1045
+
1046
+
1047
+
1048
+
1049
+
1050
+
1051
+
1052
+
1053
+
1054
+
1055
+
1056
+
1057
+
1058
+
1059
+
1060
+
1061
+
1062
+
1063
+
1064
+
1065
+
1066
+
1067
+
1068
+
1069
+
1070
+
1071
+
1072
+
1073
+
1074
+
1075
+
1076
+
1077
+
1078
+
1079
+
1080
+
1081
+
1082
+
1083
+
1084
+
1085
+
1086
+
1087
+
1088
+
1089
+
1090
+
1091
+
1092
+
1093
+
1094
+
1095
+
1096
+
1097
+
1098
+
1099
+
1100
+
1101
+
1102
+
1103
+
1104
+
1105
+
1106
+
1107
+
1108
+
1109
+
1110
+
1111
+
1112
+
1113
+
1114
+
1115
+
1116
+
1117
+
1118
+
1119
+
1120
+
1121
+
1122
+
1123
+
1124
+
1125
+
1126
+
1127
+
1128
+
1129
+
1130
+
1131
+
1132
+
1133
+
1134
+
1135
+
1136
+
1137
+
1138
+
1139
+
1140
+
1141
+
1142
+
1143
+
1144
+
1145
+
1146
+
1147
+
1148
+
1149
+
1150
+
1151
+
1152
+
1153
+
1154
+
1155
+
1156
+
1157
+
1158
+
1159
+
1160
+
1161
+
1162
+
1163
+
1164
+
1165
+
1166
+
1167
+
1168
+
1169
+
1170
+
1171
+
1172
+
1173
+
1174
+
1175
+
1176
+
1177
+
1178
+
1179
+
1180
+
1181
+
1182
+
1183
+
1184
+
1185
+
1186
+
1187
+
1188
+
1189
+
1190
+
1191
+
1192
+
1193
+
1194
+
1195
+
1196
+
1197
+
1198
+
1199
+
1200
+
1201
+
1202
+
1203
+
1204
+
1205
+
1206
+
1207
+
1208
+
1209
+
1210
+
1211
+
1212
+
1213
+
1214
+
1215
+
1216
+
1217
+
1218
+
1219
+
1220
+
1221
+
1222
+
1223
+
1224
+
1225
+
1226
+
1227
+
1228
+
1229
+
1230
+
1231
+
1232
+
1233
+
1234
+
1235
+
1236
+
1237
+
1238
+
1239
+
1240
+
1241
+
1242
+
1243
+
1244
+
1245
+
1246
+
1247
+
1248
+
1249
+
1250
+
1251
+
1252
+
1253
+
1254
+
1255
+
1256
+
1257
+
1258
+
1259
+
1260
+
1261
+
1262
+
1263
+
1264
+
1265
+
1266
+
1267
+
1268
+
1269
+
1270
+
1271
+
1272
+
1273
+
1274
+
1275
+
1276
+
1277
+
1278
+
1279
+
1280
+
1281
+
1282
+
1283
+
1284
+
1285
+
1286
+
1287
+
1288
+
1289
+
1290
+
1291
+
1292
+
1293
+
1294
+
1295
+
1296
+
1297
+
1298
+
1299
+
1300
+
1301
+
1302
+
1303
+
1304
+
1305
+
1306
+
1307
+
1308
+
1309
+
1310
+
1311
+
1312
+
1313
+
1314
+
1315
+
1316
+
1317
+
1318
+
1319
+
1320
+
1321
+
1322
+
1323
+
1324
+
1325
+
1326
+
1327
+
1328
+
1329
+
1330
+
1331
+
1332
+
1333
+
1334
+
1335
+
1336
+
1337
+
1338
+
1339
+
1340
+
1341
+
1342
+
1343
+
1344
+
1345
+
1346
+
1347
+
1348
+
1349
+
1350
+
1351
+
1352
+
1353
+
1354
+
1355
+
1356
+
1357
+
1358
+
1359
+
1360
+
1361
+
1362
+
1363
+
1364
+
1365
+
1366
+
1367
+
1368
+
1369
+
1370
+
1371
+
1372
+
1373
+
1374
+
1375
+
1376
+
1377
+
1378
+
1379
+
1380
+
1381
+
1382
+
1383
+
1384
+
1385
+
1386
+
1387
+
1388
+
1389
+
1390
+
1391
+
1392
+
1393
+
1394
+
1395
+
1396
+
1397
+
1398
+
1399
+
1400
+
1401
+
1402
+
1403
+
1404
+
1405
+
1406
+
1407
+
1408
+
1409
+
1410
+
1411
+
1412
+
1413
+
1414
+
1415
+
1416
+
1417
+
1418
+
1419
+
1420
+
1421
+
1422
+
1423
+
1424
+
1425
+
1426
+
1427
+
1428
+
1429
+
1430
+
1431
+
1432
+
1433
+
1434
+
1435
+
1436
+
1437
+
1438
+
1439
+
1440
+
1441
+
1442
+
1443
+
1444
+
1445
+
1446
+
1447
+
1448
+
1449
+
1450
+
1451
+
1452
+
1453
+
1454
+
1455
+
1456
+
1457
+
1458
+
1459
+
1460
+
1461
+
1462
+
1463
+
1464
+
1465
+
1466
+
1467
+
1468
+
1469
+
1470
+
1471
+
1472
+
1473
+
1474
+
1475
+
1476
+
1477
+
1478
+
1479
+
1480
+
1481
+
1482
+
1483
+
1484
+
1485
+
1486
+
1487
+
1488
+
1489
+
1490
+
1491
+
1492
+
1493
+
1494
+
1495
+
1496
+
1497
+
1498
+
1499
+
1500
+
1501
+
1502
+
1503
+
1504
+
1505
+
1506
+
1507
+
1508
+
1509
+
1510
+
1511
+
1512
+
1513
+
1514
+
1515
+
1516
+
1517
+
1518
+
1519
+
1520
+
1521
+
1522
+
1523
+
1524
+
1525
+
1526
+
1527
+
1528
+
1529
+
1530
+
1531
+
1532
+
1533
+
1534
+
1535
+
1536
+
1537
+
1538
+
1539
+
1540
+
1541
+
1542
+
1543
+
1544
+
1545
+
1546
+
1547
+
1548
+
1549
+
1550
+
1551
+
1552
+
1553
+
1554
+
1555
+
1556
+
1557
+
1558
+
1559
+
1560
+
1561
+
1562
+
1563
+
1564
+
1565
+
1566
+
1567
+
1568
+
1569
+
1570
+
1571
+
1572
+
1573
+
1574
+
1575
+
1576
+
1577
+
1578
+
1579
+
1580
+
1581
+
1582
+
1583
+
1584
+
1585
+
1586
+
1587
+
1588
+
1589
+
1590
+
1591
+
1592
+
1593
+
1594
+
1595
+
1596
+
1597
+
1598
+
1599
+
1600
+
1601
+
1602
+
1603
+
1604
+
1605
+
1606
+
1607
+
1608
+
1609
+
1610
+
1611
+
1612
+
1613
+
1614
+
1615
+
1616
+
1617
+
1618
+
1619
+
1620
+
1621
+
1622
+
1623
+
1624
+
1625
+
1626
+
1627
+
1628
+
1629
+
1630
+
1631
+
1632
+
1633
+
1634
+
1635
+
1636
+
1637
+
1638
+
1639
+
1640
+
1641
+
1642
+
1643
+
1644
+
1645
+
1646
+
1647
+
1648
+
1649
+
1650
+
1651
+
1652
+
1653
+
1654
+
1655
+
1656
+
1657
+
1658
+
1659
+
1660
+
1661
+
1662
+
1663
+
1664
+
1665
+
1666
+
1667
+
1668
+
1669
+
1670
+
1671
+
1672
+
1673
+
1674
+
1675
+
1676
+
1677
+
1678
+
1679
+
1680
+
1681
+
1682
+
1683
+
1684
+
1685
+
1686
+
1687
+
1688
+
1689
+
1690
+
1691
+
1692
+
1693
+
1694
+
1695
+
1696
+
1697
+
1698
+
1699
+
1700
+
1701
+
1702
+
1703
+
1704
+
1705
+
1706
+
1707
+
1708
+
1709
+
1710
+
1711
+
1712
+
1713
+
1714
+
1715
+
1716
+
1717
+
1718
+
1719
+
1720
+
1721
+
1722
+
1723
+
1724
+
1725
+
1726
+
1727
+
1728
+
1729
+
1730
+
1731
+
1732
+
1733
+
1734
+
1735
+
1736
+
1737
+
1738
+
1739
+
1740
+
1741
+
1742
+
1743
+
1744
+
1745
+
1746
+
1747
+
1748
+
1749
+
1750
+
1751
+
1752
+
1753
+
1754
+
1755
+
1756
+
1757
+
1758
+
1759
+
1760
+
1761
+
1762
+
1763
+
1764
+
1765
+
1766
+
1767
+
1768
+
1769
+
1770
+
1771
+
1772
+
1773
+
1774
+
1775
+
1776
+
1777
+
1778
+
1779
+
1780
+
1781
+
1782
+
1783
+
1784
+
1785
+
1786
+
1787
+
1788
+
1789
+
1790
+
1791
+
1792
+
1793
+
1794
+
1795
+
1796
+
1797
+
1798
+
1799
+
1800
+
1801
+
1802
+
1803
+
1804
+
1805
+
1806
+
1807
+
1808
+
1809
+
1810
+
1811
+
1812
+
1813
+
1814
+
1815
+
1816
+
1817
+
1818
+
1819
+
1820
+
1821
+
1822
+
1823
+
1824
+
1825
+
1826
+
1827
+
1828
+
1829
+
1830
+
1831
+
1832
+
1833
+
1834
+
1835
+
1836
+
1837
+
1838
+
1839
+
1840
+
1841
+
1842
+
1843
+
1844
+
1845
+
1846
+
1847
+
1848
+
1849
+
1850
+
1851
+
1852
+
1853
+
1854
+
1855
+
1856
+
1857
+
1858
+
1859
+
1860
+
1861
+
1862
+
1863
+
1864
+
1865
+
1866
+
1867
+
1868
+
1869
+
1870
+
1871
+
1872
+
1873
+
1874
+
1875
+
1876
+
1877
+
1878
+
1879
+
1880
+
1881
+
1882
+
1883
+
1884
+
1885
+
1886
+
1887
+
1888
+
1889
+
1890
+
1891
+
1892
+
1893
+
1894
+
1895
+
1896
+
1897
+
1898
+
1899
+
1900
+
1901
+
1902
+
1903
+
1904
+
1905
+
1906
+
1907
+
1908
+
1909
+
1910
+
1911
+
1912
+
1913
+
1914
+
1915
+
1916
+
1917
+
1918
+
1919
+
1920
+
1921
+
1922
+
1923
+
1924
+
1925
+
1926
+
1927
+
1928
+
1929
+
1930
+
1931
+
1932
+
1933
+
1934
+
1935
+
1936
+
1937
+
1938
+
1939
+
1940
+
1941
+
1942
+
1943
+
1944
+
1945
+
1946
+
1947
+
1948
+
1949
+
1950
+
1951
+
1952
+
1953
+
1954
+
1955
+
1956
+
1957
+
1958
+
1959
+
1960
+
1961
+
1962
+
1963
+
1964
+
1965
+
1966
+
1967
+
1968
+
1969
+
1970
+
1971
+
1972
+
1973
+
1974
+
1975
+
1976
+
1977
+
1978
+
1979
+
1980
+
1981
+
1982
+
1983
+
1984
+
1985
+
1986
+
1987
+
1988
+
1989
+
1990
+
1991
+
1992
+
1993
+
1994
+
1995
+
1996
+
1997
+
1998
+
1999
+
2000
+
2001
+
2002
+
2003
+
2004
+
2005
+
2006
+
2007
+
2008
+
2009
+
2010
+
2011
+
2012
+
2013
+
2014
+
2015
+
2016
+
2017
+
2018
+
2019
+
2020
+
2021
+
2022
+
2023
+
2024
+
2025
+
2026
+
2027
+
2028
+
2029
+
2030
+
2031
+
2032
+
2033
+
2034
+
2035
+
2036
+
2037
+
2038
+
2039
+
2040
+
2041
+
2042
+
2043
+
2044
+
2045
+
2046
+
2047
+
2048
+
2049
+
2050
+
2051
+
2052
+
2053
+
2054
+
2055
+
2056
+
2057
+
2058
+
2059
+
2060
+
2061
+
2062
+
2063
+
2064
+
2065
+
2066
+
2067
+
2068
+
2069
+
2070
+
2071
+
2072
+
2073
+
2074
+
2075
+
2076
+
2077
+
2078
+
2079
+
2080
+
2081
+
2082
+
2083
+
2084
+
2085
+
2086
+
2087
+
2088
+
2089
+
2090
+
2091
+
2092
+
2093
+
2094
+
2095
+
2096
+
2097
+
2098
+
2099
+
2100
+
2101
+
2102
+
2103
+
2104
+
2105
+
2106
+
2107
+
2108
+
2109
+
2110
+
2111
+
2112
+
2113
+
2114
+
2115
+
2116
+
2117
+
2118
+
2119
+
2120
+
2121
+
2122
+
2123
+
2124
+
2125
+
2126
+
2127
+
2128
+
2129
+
2130
+
2131
+
2132
+
2133
+
2134
+
2135
+
2136
+
2137
+
2138
+
2139
+
2140
+
2141
+
2142
+
2143
+
2144
+
2145
+
2146
+
2147
+
2148
+
2149
+
2150
+
2151
+
2152
+
2153
+
2154
+
2155
+
2156
+
2157
+
2158
+
2159
+
2160
+
2161
+
2162
+
2163
+
2164
+
2165
+
2166
+
2167
+
2168
+
2169
+
2170
+
2171
+
2172
+
2173
+
2174
+
2175
+
2176
+
2177
+
2178
+ =>> [Global Step] 002180 =>> LR :: 0.000000 -- Loss :: 3.2451: 100%|█| 2180/2180 [1:33:12<00:00, 2.62s
2179
+
2180
+ 03/13 [08:37:25] INFO  | >> [*] Done with Training =>> Finalizing Metrics ]8;id=696414;file:///home/lilei/prismatic-vlms/scripts/pretrain.py\pretrain.py]8;;\:]8;id=68157;file:///home/lilei/prismatic-vlms/scripts/pretrain.py#230\230]8;;\
stage1_qformer2_256/wandb/run-20240313_070329-pog6cg3f/files/requirements.txt ADDED
@@ -0,0 +1,122 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ accelerate==0.27.2
2
+ aiohttp==3.9.3
3
+ aiosignal==1.3.1
4
+ annotated-types==0.6.0
5
+ appdirs==1.4.4
6
+ async-timeout==4.0.3
7
+ attrs==23.2.0
8
+ beautifulsoup4==4.12.3
9
+ blis==0.7.11
10
+ catalogue==2.0.10
11
+ certifi==2024.2.2
12
+ charset-normalizer==2.0.4
13
+ click==8.1.7
14
+ cloudpathlib==0.16.0
15
+ confection==0.1.4
16
+ contourpy==1.2.0
17
+ cycler==0.12.1
18
+ cymem==2.0.8
19
+ datasets==2.18.0
20
+ deepspeed==0.13.4
21
+ dill==0.3.8
22
+ docker-pycreds==0.4.0
23
+ draccus==0.7.1
24
+ einops==0.7.0
25
+ filelock==3.13.1
26
+ flash-attn==2.5.5
27
+ fonttools==4.49.0
28
+ frozenlist==1.4.1
29
+ fsspec==2024.2.0
30
+ gdown==5.1.0
31
+ gitdb==4.0.11
32
+ gitpython==3.1.42
33
+ gmpy2==2.1.2
34
+ hjson==3.1.0
35
+ huggingface-hub==0.20.3
36
+ idna==3.4
37
+ install==1.3.5
38
+ jinja2==3.1.3
39
+ jsonlines==4.0.0
40
+ kiwisolver==1.4.5
41
+ langcodes==3.3.0
42
+ markdown-it-py==3.0.0
43
+ markupsafe==2.1.3
44
+ matplotlib==3.8.3
45
+ mdurl==0.1.2
46
+ mergedeep==1.3.4
47
+ mkl-fft==1.3.8
48
+ mkl-random==1.2.4
49
+ mkl-service==2.4.0
50
+ mpmath==1.3.0
51
+ multidict==6.0.5
52
+ multiprocess==0.70.16
53
+ murmurhash==1.0.10
54
+ mypy-extensions==1.0.0
55
+ networkx==3.1
56
+ ninja==1.11.1.1
57
+ numpy==1.26.4
58
+ ocifs==1.3.1
59
+ packaging==23.2
60
+ pandas==2.2.1
61
+ pillow==10.2.0
62
+ pip==23.3.1
63
+ preshed==3.0.9
64
+ prismatic==0.0.1
65
+ protobuf==4.25.3
66
+ psutil==5.9.8
67
+ py-cpuinfo==9.0.0
68
+ pyarrow-hotfix==0.6
69
+ pyarrow==15.0.0
70
+ pycocoevalcap==1.2
71
+ pycocotools==2.0.7
72
+ pydantic-core==2.16.3
73
+ pydantic==2.6.3
74
+ pygments==2.17.2
75
+ pynvml==11.5.0
76
+ pyparsing==3.1.1
77
+ pysocks==1.7.1
78
+ pythainlp==5.0.1
79
+ python-dateutil==2.9.0.post0
80
+ pytz==2024.1
81
+ pyyaml-include==1.3.2
82
+ pyyaml==6.0.1
83
+ regex==2023.12.25
84
+ requests==2.31.0
85
+ rich==13.7.0
86
+ safetensors==0.4.2
87
+ sentencepiece==0.2.0
88
+ sentry-sdk==1.40.5
89
+ setproctitle==1.3.3
90
+ setuptools==68.2.2
91
+ six==1.16.0
92
+ smart-open==6.4.0
93
+ smmap==5.0.1
94
+ soupsieve==2.5
95
+ spacy-legacy==3.0.12
96
+ spacy-loggers==1.0.5
97
+ spacy==3.7.4
98
+ srsly==2.4.8
99
+ sudachidict-core==20240109
100
+ sudachipy==0.6.8
101
+ sympy==1.12
102
+ thinc==8.2.3
103
+ timm==0.9.16
104
+ tokenizers==0.15.2
105
+ toml==0.10.2
106
+ torch==2.2.1
107
+ torchaudio==2.2.1
108
+ torchvision==0.17.1
109
+ tqdm==4.66.2
110
+ transformers==4.38.1
111
+ triton==2.2.0
112
+ typer==0.9.0
113
+ typing-extensions==4.9.0
114
+ typing-inspect==0.9.0
115
+ tzdata==2024.1
116
+ urllib3==2.1.0
117
+ wandb==0.16.3
118
+ wasabi==1.1.2
119
+ weasel==0.3.4
120
+ wheel==0.41.2
121
+ xxhash==3.4.1
122
+ yarl==1.9.4
stage1_qformer2_256/wandb/run-20240313_070329-pog6cg3f/files/wandb-metadata.json ADDED
@@ -0,0 +1,1362 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.15.0-1027-oracle-x86_64-with-glibc2.31",
3
+ "python": "3.10.13",
4
+ "heartbeatAt": "2024-03-13T07:03:30.597779",
5
+ "startedAt": "2024-03-13T07:03:29.769600",
6
+ "docker": null,
7
+ "cuda": null,
8
+ "args": [
9
+ "--model.type",
10
+ "one-stage+7b",
11
+ "--stage",
12
+ "align",
13
+ "--model.model_id",
14
+ "one-stage+7b_qformer2_256",
15
+ "--run_id",
16
+ "stage1_qformer2_256",
17
+ "--model.arch_specifier",
18
+ "qformer2_256"
19
+ ],
20
+ "state": "running",
21
+ "program": "/home/lilei/prismatic-vlms/scripts/pretrain.py",
22
+ "codePathLocal": "scripts/pretrain.py",
23
+ "codePath": "scripts/pretrain.py",
24
+ "git": {
25
+ "remote": "https://github.com/TRI-ML/prismatic-vlms.git",
26
+ "commit": "e87c4d04915bf8fad0d90f2ae79454df778ee2d3"
27
+ },
28
+ "email": "tobiaslee@qq.com",
29
+ "root": "/home/lilei/prismatic-vlms",
30
+ "host": "compute-permanent-node-91",
31
+ "username": "lilei",
32
+ "executable": "/home/lilei/anaconda3/envs/pvlm/bin/python",
33
+ "cpu_count": 128,
34
+ "cpu_count_logical": 255,
35
+ "cpu_freq": {
36
+ "current": 2567.1000549019604,
37
+ "min": 1500.0,
38
+ "max": 2550.0
39
+ },
40
+ "cpu_freq_per_core": [
41
+ {
42
+ "current": 2550.0,
43
+ "min": 1500.0,
44
+ "max": 2550.0
45
+ },
46
+ {
47
+ "current": 2550.0,
48
+ "min": 1500.0,
49
+ "max": 2550.0
50
+ },
51
+ {
52
+ "current": 2550.0,
53
+ "min": 1500.0,
54
+ "max": 2550.0
55
+ },
56
+ {
57
+ "current": 2550.0,
58
+ "min": 1500.0,
59
+ "max": 2550.0
60
+ },
61
+ {
62
+ "current": 2550.0,
63
+ "min": 1500.0,
64
+ "max": 2550.0
65
+ },
66
+ {
67
+ "current": 2550.0,
68
+ "min": 1500.0,
69
+ "max": 2550.0
70
+ },
71
+ {
72
+ "current": 2550.0,
73
+ "min": 1500.0,
74
+ "max": 2550.0
75
+ },
76
+ {
77
+ "current": 2550.0,
78
+ "min": 1500.0,
79
+ "max": 2550.0
80
+ },
81
+ {
82
+ "current": 2550.0,
83
+ "min": 1500.0,
84
+ "max": 2550.0
85
+ },
86
+ {
87
+ "current": 2643.771,
88
+ "min": 1500.0,
89
+ "max": 2550.0
90
+ },
91
+ {
92
+ "current": 2550.0,
93
+ "min": 1500.0,
94
+ "max": 2550.0
95
+ },
96
+ {
97
+ "current": 2550.0,
98
+ "min": 1500.0,
99
+ "max": 2550.0
100
+ },
101
+ {
102
+ "current": 2550.0,
103
+ "min": 1500.0,
104
+ "max": 2550.0
105
+ },
106
+ {
107
+ "current": 2550.0,
108
+ "min": 1500.0,
109
+ "max": 2550.0
110
+ },
111
+ {
112
+ "current": 2550.0,
113
+ "min": 1500.0,
114
+ "max": 2550.0
115
+ },
116
+ {
117
+ "current": 2550.0,
118
+ "min": 1500.0,
119
+ "max": 2550.0
120
+ },
121
+ {
122
+ "current": 3249.468,
123
+ "min": 1500.0,
124
+ "max": 2550.0
125
+ },
126
+ {
127
+ "current": 2550.0,
128
+ "min": 1500.0,
129
+ "max": 2550.0
130
+ },
131
+ {
132
+ "current": 2550.0,
133
+ "min": 1500.0,
134
+ "max": 2550.0
135
+ },
136
+ {
137
+ "current": 2550.0,
138
+ "min": 1500.0,
139
+ "max": 2550.0
140
+ },
141
+ {
142
+ "current": 2550.0,
143
+ "min": 1500.0,
144
+ "max": 2550.0
145
+ },
146
+ {
147
+ "current": 2550.0,
148
+ "min": 1500.0,
149
+ "max": 2550.0
150
+ },
151
+ {
152
+ "current": 2550.0,
153
+ "min": 1500.0,
154
+ "max": 2550.0
155
+ },
156
+ {
157
+ "current": 2550.0,
158
+ "min": 1500.0,
159
+ "max": 2550.0
160
+ },
161
+ {
162
+ "current": 2550.0,
163
+ "min": 1500.0,
164
+ "max": 2550.0
165
+ },
166
+ {
167
+ "current": 2550.0,
168
+ "min": 1500.0,
169
+ "max": 2550.0
170
+ },
171
+ {
172
+ "current": 2550.0,
173
+ "min": 1500.0,
174
+ "max": 2550.0
175
+ },
176
+ {
177
+ "current": 2550.0,
178
+ "min": 1500.0,
179
+ "max": 2550.0
180
+ },
181
+ {
182
+ "current": 2550.0,
183
+ "min": 1500.0,
184
+ "max": 2550.0
185
+ },
186
+ {
187
+ "current": 2550.0,
188
+ "min": 1500.0,
189
+ "max": 2550.0
190
+ },
191
+ {
192
+ "current": 2550.0,
193
+ "min": 1500.0,
194
+ "max": 2550.0
195
+ },
196
+ {
197
+ "current": 2550.0,
198
+ "min": 1500.0,
199
+ "max": 2550.0
200
+ },
201
+ {
202
+ "current": 3249.321,
203
+ "min": 1500.0,
204
+ "max": 2550.0
205
+ },
206
+ {
207
+ "current": 2550.0,
208
+ "min": 1500.0,
209
+ "max": 2550.0
210
+ },
211
+ {
212
+ "current": 2550.0,
213
+ "min": 1500.0,
214
+ "max": 2550.0
215
+ },
216
+ {
217
+ "current": 2550.0,
218
+ "min": 1500.0,
219
+ "max": 2550.0
220
+ },
221
+ {
222
+ "current": 2550.0,
223
+ "min": 1500.0,
224
+ "max": 2550.0
225
+ },
226
+ {
227
+ "current": 2550.0,
228
+ "min": 1500.0,
229
+ "max": 2550.0
230
+ },
231
+ {
232
+ "current": 2550.0,
233
+ "min": 1500.0,
234
+ "max": 2550.0
235
+ },
236
+ {
237
+ "current": 2550.0,
238
+ "min": 1500.0,
239
+ "max": 2550.0
240
+ },
241
+ {
242
+ "current": 2550.0,
243
+ "min": 1500.0,
244
+ "max": 2550.0
245
+ },
246
+ {
247
+ "current": 2550.0,
248
+ "min": 1500.0,
249
+ "max": 2550.0
250
+ },
251
+ {
252
+ "current": 2550.0,
253
+ "min": 1500.0,
254
+ "max": 2550.0
255
+ },
256
+ {
257
+ "current": 2550.0,
258
+ "min": 1500.0,
259
+ "max": 2550.0
260
+ },
261
+ {
262
+ "current": 2550.0,
263
+ "min": 1500.0,
264
+ "max": 2550.0
265
+ },
266
+ {
267
+ "current": 2550.0,
268
+ "min": 1500.0,
269
+ "max": 2550.0
270
+ },
271
+ {
272
+ "current": 2550.0,
273
+ "min": 1500.0,
274
+ "max": 2550.0
275
+ },
276
+ {
277
+ "current": 2550.0,
278
+ "min": 1500.0,
279
+ "max": 2550.0
280
+ },
281
+ {
282
+ "current": 3248.96,
283
+ "min": 1500.0,
284
+ "max": 2550.0
285
+ },
286
+ {
287
+ "current": 2550.0,
288
+ "min": 1500.0,
289
+ "max": 2550.0
290
+ },
291
+ {
292
+ "current": 2550.0,
293
+ "min": 1500.0,
294
+ "max": 2550.0
295
+ },
296
+ {
297
+ "current": 2550.0,
298
+ "min": 1500.0,
299
+ "max": 2550.0
300
+ },
301
+ {
302
+ "current": 2550.0,
303
+ "min": 1500.0,
304
+ "max": 2550.0
305
+ },
306
+ {
307
+ "current": 2550.0,
308
+ "min": 1500.0,
309
+ "max": 2550.0
310
+ },
311
+ {
312
+ "current": 2550.0,
313
+ "min": 1500.0,
314
+ "max": 2550.0
315
+ },
316
+ {
317
+ "current": 2550.0,
318
+ "min": 1500.0,
319
+ "max": 2550.0
320
+ },
321
+ {
322
+ "current": 2550.0,
323
+ "min": 1500.0,
324
+ "max": 2550.0
325
+ },
326
+ {
327
+ "current": 2550.0,
328
+ "min": 1500.0,
329
+ "max": 2550.0
330
+ },
331
+ {
332
+ "current": 2550.0,
333
+ "min": 1500.0,
334
+ "max": 2550.0
335
+ },
336
+ {
337
+ "current": 2550.0,
338
+ "min": 1500.0,
339
+ "max": 2550.0
340
+ },
341
+ {
342
+ "current": 2550.0,
343
+ "min": 1500.0,
344
+ "max": 2550.0
345
+ },
346
+ {
347
+ "current": 2550.0,
348
+ "min": 1500.0,
349
+ "max": 2550.0
350
+ },
351
+ {
352
+ "current": 2550.0,
353
+ "min": 1500.0,
354
+ "max": 2550.0
355
+ },
356
+ {
357
+ "current": 2550.0,
358
+ "min": 1500.0,
359
+ "max": 2550.0
360
+ },
361
+ {
362
+ "current": 2550.0,
363
+ "min": 1500.0,
364
+ "max": 2550.0
365
+ },
366
+ {
367
+ "current": 2550.0,
368
+ "min": 1500.0,
369
+ "max": 2550.0
370
+ },
371
+ {
372
+ "current": 2550.0,
373
+ "min": 1500.0,
374
+ "max": 2550.0
375
+ },
376
+ {
377
+ "current": 2550.0,
378
+ "min": 1500.0,
379
+ "max": 2550.0
380
+ },
381
+ {
382
+ "current": 2550.0,
383
+ "min": 1500.0,
384
+ "max": 2550.0
385
+ },
386
+ {
387
+ "current": 2550.0,
388
+ "min": 1500.0,
389
+ "max": 2550.0
390
+ },
391
+ {
392
+ "current": 2550.0,
393
+ "min": 1500.0,
394
+ "max": 2550.0
395
+ },
396
+ {
397
+ "current": 2550.0,
398
+ "min": 1500.0,
399
+ "max": 2550.0
400
+ },
401
+ {
402
+ "current": 2550.0,
403
+ "min": 1500.0,
404
+ "max": 2550.0
405
+ },
406
+ {
407
+ "current": 2550.0,
408
+ "min": 1500.0,
409
+ "max": 2550.0
410
+ },
411
+ {
412
+ "current": 2550.0,
413
+ "min": 1500.0,
414
+ "max": 2550.0
415
+ },
416
+ {
417
+ "current": 2550.0,
418
+ "min": 1500.0,
419
+ "max": 2550.0
420
+ },
421
+ {
422
+ "current": 2550.0,
423
+ "min": 1500.0,
424
+ "max": 2550.0
425
+ },
426
+ {
427
+ "current": 2550.0,
428
+ "min": 1500.0,
429
+ "max": 2550.0
430
+ },
431
+ {
432
+ "current": 2550.0,
433
+ "min": 1500.0,
434
+ "max": 2550.0
435
+ },
436
+ {
437
+ "current": 2550.0,
438
+ "min": 1500.0,
439
+ "max": 2550.0
440
+ },
441
+ {
442
+ "current": 2550.0,
443
+ "min": 1500.0,
444
+ "max": 2550.0
445
+ },
446
+ {
447
+ "current": 2550.0,
448
+ "min": 1500.0,
449
+ "max": 2550.0
450
+ },
451
+ {
452
+ "current": 2550.0,
453
+ "min": 1500.0,
454
+ "max": 2550.0
455
+ },
456
+ {
457
+ "current": 2550.0,
458
+ "min": 1500.0,
459
+ "max": 2550.0
460
+ },
461
+ {
462
+ "current": 2550.0,
463
+ "min": 1500.0,
464
+ "max": 2550.0
465
+ },
466
+ {
467
+ "current": 2550.0,
468
+ "min": 1500.0,
469
+ "max": 2550.0
470
+ },
471
+ {
472
+ "current": 2550.0,
473
+ "min": 1500.0,
474
+ "max": 2550.0
475
+ },
476
+ {
477
+ "current": 2550.0,
478
+ "min": 1500.0,
479
+ "max": 2550.0
480
+ },
481
+ {
482
+ "current": 2550.0,
483
+ "min": 1500.0,
484
+ "max": 2550.0
485
+ },
486
+ {
487
+ "current": 2550.0,
488
+ "min": 1500.0,
489
+ "max": 2550.0
490
+ },
491
+ {
492
+ "current": 2550.0,
493
+ "min": 1500.0,
494
+ "max": 2550.0
495
+ },
496
+ {
497
+ "current": 2550.0,
498
+ "min": 1500.0,
499
+ "max": 2550.0
500
+ },
501
+ {
502
+ "current": 2550.0,
503
+ "min": 1500.0,
504
+ "max": 2550.0
505
+ },
506
+ {
507
+ "current": 2550.0,
508
+ "min": 1500.0,
509
+ "max": 2550.0
510
+ },
511
+ {
512
+ "current": 2550.0,
513
+ "min": 1500.0,
514
+ "max": 2550.0
515
+ },
516
+ {
517
+ "current": 2550.0,
518
+ "min": 1500.0,
519
+ "max": 2550.0
520
+ },
521
+ {
522
+ "current": 2550.0,
523
+ "min": 1500.0,
524
+ "max": 2550.0
525
+ },
526
+ {
527
+ "current": 2550.0,
528
+ "min": 1500.0,
529
+ "max": 2550.0
530
+ },
531
+ {
532
+ "current": 2550.0,
533
+ "min": 1500.0,
534
+ "max": 2550.0
535
+ },
536
+ {
537
+ "current": 2550.0,
538
+ "min": 1500.0,
539
+ "max": 2550.0
540
+ },
541
+ {
542
+ "current": 2550.0,
543
+ "min": 1500.0,
544
+ "max": 2550.0
545
+ },
546
+ {
547
+ "current": 2550.0,
548
+ "min": 1500.0,
549
+ "max": 2550.0
550
+ },
551
+ {
552
+ "current": 2550.0,
553
+ "min": 1500.0,
554
+ "max": 2550.0
555
+ },
556
+ {
557
+ "current": 2550.0,
558
+ "min": 1500.0,
559
+ "max": 2550.0
560
+ },
561
+ {
562
+ "current": 2550.0,
563
+ "min": 1500.0,
564
+ "max": 2550.0
565
+ },
566
+ {
567
+ "current": 2550.0,
568
+ "min": 1500.0,
569
+ "max": 2550.0
570
+ },
571
+ {
572
+ "current": 2550.0,
573
+ "min": 1500.0,
574
+ "max": 2550.0
575
+ },
576
+ {
577
+ "current": 2550.0,
578
+ "min": 1500.0,
579
+ "max": 2550.0
580
+ },
581
+ {
582
+ "current": 2550.0,
583
+ "min": 1500.0,
584
+ "max": 2550.0
585
+ },
586
+ {
587
+ "current": 2550.0,
588
+ "min": 1500.0,
589
+ "max": 2550.0
590
+ },
591
+ {
592
+ "current": 2550.0,
593
+ "min": 1500.0,
594
+ "max": 2550.0
595
+ },
596
+ {
597
+ "current": 2550.0,
598
+ "min": 1500.0,
599
+ "max": 2550.0
600
+ },
601
+ {
602
+ "current": 2550.0,
603
+ "min": 1500.0,
604
+ "max": 2550.0
605
+ },
606
+ {
607
+ "current": 2550.0,
608
+ "min": 1500.0,
609
+ "max": 2550.0
610
+ },
611
+ {
612
+ "current": 2550.0,
613
+ "min": 1500.0,
614
+ "max": 2550.0
615
+ },
616
+ {
617
+ "current": 2550.0,
618
+ "min": 1500.0,
619
+ "max": 2550.0
620
+ },
621
+ {
622
+ "current": 2550.0,
623
+ "min": 1500.0,
624
+ "max": 2550.0
625
+ },
626
+ {
627
+ "current": 2550.0,
628
+ "min": 1500.0,
629
+ "max": 2550.0
630
+ },
631
+ {
632
+ "current": 2550.0,
633
+ "min": 1500.0,
634
+ "max": 2550.0
635
+ },
636
+ {
637
+ "current": 2550.0,
638
+ "min": 1500.0,
639
+ "max": 2550.0
640
+ },
641
+ {
642
+ "current": 2550.0,
643
+ "min": 1500.0,
644
+ "max": 2550.0
645
+ },
646
+ {
647
+ "current": 2550.0,
648
+ "min": 1500.0,
649
+ "max": 2550.0
650
+ },
651
+ {
652
+ "current": 2550.0,
653
+ "min": 1500.0,
654
+ "max": 2550.0
655
+ },
656
+ {
657
+ "current": 2550.0,
658
+ "min": 1500.0,
659
+ "max": 2550.0
660
+ },
661
+ {
662
+ "current": 2550.0,
663
+ "min": 1500.0,
664
+ "max": 2550.0
665
+ },
666
+ {
667
+ "current": 2550.0,
668
+ "min": 1500.0,
669
+ "max": 2550.0
670
+ },
671
+ {
672
+ "current": 2550.0,
673
+ "min": 1500.0,
674
+ "max": 2550.0
675
+ },
676
+ {
677
+ "current": 2550.0,
678
+ "min": 1500.0,
679
+ "max": 2550.0
680
+ },
681
+ {
682
+ "current": 2550.0,
683
+ "min": 1500.0,
684
+ "max": 2550.0
685
+ },
686
+ {
687
+ "current": 2550.0,
688
+ "min": 1500.0,
689
+ "max": 2550.0
690
+ },
691
+ {
692
+ "current": 2550.0,
693
+ "min": 1500.0,
694
+ "max": 2550.0
695
+ },
696
+ {
697
+ "current": 2550.0,
698
+ "min": 1500.0,
699
+ "max": 2550.0
700
+ },
701
+ {
702
+ "current": 3248.901,
703
+ "min": 1500.0,
704
+ "max": 2550.0
705
+ },
706
+ {
707
+ "current": 2550.0,
708
+ "min": 1500.0,
709
+ "max": 2550.0
710
+ },
711
+ {
712
+ "current": 2550.0,
713
+ "min": 1500.0,
714
+ "max": 2550.0
715
+ },
716
+ {
717
+ "current": 2550.0,
718
+ "min": 1500.0,
719
+ "max": 2550.0
720
+ },
721
+ {
722
+ "current": 2550.0,
723
+ "min": 1500.0,
724
+ "max": 2550.0
725
+ },
726
+ {
727
+ "current": 2550.0,
728
+ "min": 1500.0,
729
+ "max": 2550.0
730
+ },
731
+ {
732
+ "current": 2550.0,
733
+ "min": 1500.0,
734
+ "max": 2550.0
735
+ },
736
+ {
737
+ "current": 3249.789,
738
+ "min": 1500.0,
739
+ "max": 2550.0
740
+ },
741
+ {
742
+ "current": 2550.0,
743
+ "min": 1500.0,
744
+ "max": 2550.0
745
+ },
746
+ {
747
+ "current": 2550.0,
748
+ "min": 1500.0,
749
+ "max": 2550.0
750
+ },
751
+ {
752
+ "current": 2550.0,
753
+ "min": 1500.0,
754
+ "max": 2550.0
755
+ },
756
+ {
757
+ "current": 2550.0,
758
+ "min": 1500.0,
759
+ "max": 2550.0
760
+ },
761
+ {
762
+ "current": 2550.0,
763
+ "min": 1500.0,
764
+ "max": 2550.0
765
+ },
766
+ {
767
+ "current": 2550.0,
768
+ "min": 1500.0,
769
+ "max": 2550.0
770
+ },
771
+ {
772
+ "current": 2550.0,
773
+ "min": 1500.0,
774
+ "max": 2550.0
775
+ },
776
+ {
777
+ "current": 2550.0,
778
+ "min": 1500.0,
779
+ "max": 2550.0
780
+ },
781
+ {
782
+ "current": 2550.0,
783
+ "min": 1500.0,
784
+ "max": 2550.0
785
+ },
786
+ {
787
+ "current": 2550.0,
788
+ "min": 1500.0,
789
+ "max": 2550.0
790
+ },
791
+ {
792
+ "current": 2550.0,
793
+ "min": 1500.0,
794
+ "max": 2550.0
795
+ },
796
+ {
797
+ "current": 2550.0,
798
+ "min": 1500.0,
799
+ "max": 2550.0
800
+ },
801
+ {
802
+ "current": 2550.0,
803
+ "min": 1500.0,
804
+ "max": 2550.0
805
+ },
806
+ {
807
+ "current": 2550.0,
808
+ "min": 1500.0,
809
+ "max": 2550.0
810
+ },
811
+ {
812
+ "current": 2550.0,
813
+ "min": 1500.0,
814
+ "max": 2550.0
815
+ },
816
+ {
817
+ "current": 3249.09,
818
+ "min": 1500.0,
819
+ "max": 2550.0
820
+ },
821
+ {
822
+ "current": 2550.0,
823
+ "min": 1500.0,
824
+ "max": 2550.0
825
+ },
826
+ {
827
+ "current": 3249.758,
828
+ "min": 1500.0,
829
+ "max": 2550.0
830
+ },
831
+ {
832
+ "current": 2550.0,
833
+ "min": 1500.0,
834
+ "max": 2550.0
835
+ },
836
+ {
837
+ "current": 2550.0,
838
+ "min": 1500.0,
839
+ "max": 2550.0
840
+ },
841
+ {
842
+ "current": 2550.0,
843
+ "min": 1500.0,
844
+ "max": 2550.0
845
+ },
846
+ {
847
+ "current": 2550.0,
848
+ "min": 1500.0,
849
+ "max": 2550.0
850
+ },
851
+ {
852
+ "current": 2550.0,
853
+ "min": 1500.0,
854
+ "max": 2550.0
855
+ },
856
+ {
857
+ "current": 2550.0,
858
+ "min": 1500.0,
859
+ "max": 2550.0
860
+ },
861
+ {
862
+ "current": 2550.0,
863
+ "min": 1500.0,
864
+ "max": 2550.0
865
+ },
866
+ {
867
+ "current": 2550.0,
868
+ "min": 1500.0,
869
+ "max": 2550.0
870
+ },
871
+ {
872
+ "current": 2550.0,
873
+ "min": 1500.0,
874
+ "max": 2550.0
875
+ },
876
+ {
877
+ "current": 2550.0,
878
+ "min": 1500.0,
879
+ "max": 2550.0
880
+ },
881
+ {
882
+ "current": 2550.0,
883
+ "min": 1500.0,
884
+ "max": 2550.0
885
+ },
886
+ {
887
+ "current": 2550.0,
888
+ "min": 1500.0,
889
+ "max": 2550.0
890
+ },
891
+ {
892
+ "current": 2550.0,
893
+ "min": 1500.0,
894
+ "max": 2550.0
895
+ },
896
+ {
897
+ "current": 2550.0,
898
+ "min": 1500.0,
899
+ "max": 2550.0
900
+ },
901
+ {
902
+ "current": 2550.0,
903
+ "min": 1500.0,
904
+ "max": 2550.0
905
+ },
906
+ {
907
+ "current": 2550.0,
908
+ "min": 1500.0,
909
+ "max": 2550.0
910
+ },
911
+ {
912
+ "current": 2550.0,
913
+ "min": 1500.0,
914
+ "max": 2550.0
915
+ },
916
+ {
917
+ "current": 2550.0,
918
+ "min": 1500.0,
919
+ "max": 2550.0
920
+ },
921
+ {
922
+ "current": 2550.0,
923
+ "min": 1500.0,
924
+ "max": 2550.0
925
+ },
926
+ {
927
+ "current": 2550.0,
928
+ "min": 1500.0,
929
+ "max": 2550.0
930
+ },
931
+ {
932
+ "current": 2550.0,
933
+ "min": 1500.0,
934
+ "max": 2550.0
935
+ },
936
+ {
937
+ "current": 2550.0,
938
+ "min": 1500.0,
939
+ "max": 2550.0
940
+ },
941
+ {
942
+ "current": 2550.0,
943
+ "min": 1500.0,
944
+ "max": 2550.0
945
+ },
946
+ {
947
+ "current": 2550.0,
948
+ "min": 1500.0,
949
+ "max": 2550.0
950
+ },
951
+ {
952
+ "current": 2550.0,
953
+ "min": 1500.0,
954
+ "max": 2550.0
955
+ },
956
+ {
957
+ "current": 2550.0,
958
+ "min": 1500.0,
959
+ "max": 2550.0
960
+ },
961
+ {
962
+ "current": 2550.0,
963
+ "min": 1500.0,
964
+ "max": 2550.0
965
+ },
966
+ {
967
+ "current": 2550.0,
968
+ "min": 1500.0,
969
+ "max": 2550.0
970
+ },
971
+ {
972
+ "current": 2550.0,
973
+ "min": 1500.0,
974
+ "max": 2550.0
975
+ },
976
+ {
977
+ "current": 2550.0,
978
+ "min": 1500.0,
979
+ "max": 2550.0
980
+ },
981
+ {
982
+ "current": 2550.0,
983
+ "min": 1500.0,
984
+ "max": 2550.0
985
+ },
986
+ {
987
+ "current": 3249.146,
988
+ "min": 1500.0,
989
+ "max": 2550.0
990
+ },
991
+ {
992
+ "current": 2550.0,
993
+ "min": 1500.0,
994
+ "max": 2550.0
995
+ },
996
+ {
997
+ "current": 2550.0,
998
+ "min": 1500.0,
999
+ "max": 2550.0
1000
+ },
1001
+ {
1002
+ "current": 2550.0,
1003
+ "min": 1500.0,
1004
+ "max": 2550.0
1005
+ },
1006
+ {
1007
+ "current": 2550.0,
1008
+ "min": 1500.0,
1009
+ "max": 2550.0
1010
+ },
1011
+ {
1012
+ "current": 2550.0,
1013
+ "min": 1500.0,
1014
+ "max": 2550.0
1015
+ },
1016
+ {
1017
+ "current": 2550.0,
1018
+ "min": 1500.0,
1019
+ "max": 2550.0
1020
+ },
1021
+ {
1022
+ "current": 2550.0,
1023
+ "min": 1500.0,
1024
+ "max": 2550.0
1025
+ },
1026
+ {
1027
+ "current": 2550.0,
1028
+ "min": 1500.0,
1029
+ "max": 2550.0
1030
+ },
1031
+ {
1032
+ "current": 2550.0,
1033
+ "min": 1500.0,
1034
+ "max": 2550.0
1035
+ },
1036
+ {
1037
+ "current": 2550.0,
1038
+ "min": 1500.0,
1039
+ "max": 2550.0
1040
+ },
1041
+ {
1042
+ "current": 2550.0,
1043
+ "min": 1500.0,
1044
+ "max": 2550.0
1045
+ },
1046
+ {
1047
+ "current": 2550.0,
1048
+ "min": 1500.0,
1049
+ "max": 2550.0
1050
+ },
1051
+ {
1052
+ "current": 2550.0,
1053
+ "min": 1500.0,
1054
+ "max": 2550.0
1055
+ },
1056
+ {
1057
+ "current": 2550.0,
1058
+ "min": 1500.0,
1059
+ "max": 2550.0
1060
+ },
1061
+ {
1062
+ "current": 2550.0,
1063
+ "min": 1500.0,
1064
+ "max": 2550.0
1065
+ },
1066
+ {
1067
+ "current": 2550.0,
1068
+ "min": 1500.0,
1069
+ "max": 2550.0
1070
+ },
1071
+ {
1072
+ "current": 2550.0,
1073
+ "min": 1500.0,
1074
+ "max": 2550.0
1075
+ },
1076
+ {
1077
+ "current": 2550.0,
1078
+ "min": 1500.0,
1079
+ "max": 2550.0
1080
+ },
1081
+ {
1082
+ "current": 2550.0,
1083
+ "min": 1500.0,
1084
+ "max": 2550.0
1085
+ },
1086
+ {
1087
+ "current": 2550.0,
1088
+ "min": 1500.0,
1089
+ "max": 2550.0
1090
+ },
1091
+ {
1092
+ "current": 2550.0,
1093
+ "min": 1500.0,
1094
+ "max": 2550.0
1095
+ },
1096
+ {
1097
+ "current": 2550.0,
1098
+ "min": 1500.0,
1099
+ "max": 2550.0
1100
+ },
1101
+ {
1102
+ "current": 2550.0,
1103
+ "min": 1500.0,
1104
+ "max": 2550.0
1105
+ },
1106
+ {
1107
+ "current": 2550.0,
1108
+ "min": 1500.0,
1109
+ "max": 2550.0
1110
+ },
1111
+ {
1112
+ "current": 2550.0,
1113
+ "min": 1500.0,
1114
+ "max": 2550.0
1115
+ },
1116
+ {
1117
+ "current": 2550.0,
1118
+ "min": 1500.0,
1119
+ "max": 2550.0
1120
+ },
1121
+ {
1122
+ "current": 2550.0,
1123
+ "min": 1500.0,
1124
+ "max": 2550.0
1125
+ },
1126
+ {
1127
+ "current": 2550.0,
1128
+ "min": 1500.0,
1129
+ "max": 2550.0
1130
+ },
1131
+ {
1132
+ "current": 2550.0,
1133
+ "min": 1500.0,
1134
+ "max": 2550.0
1135
+ },
1136
+ {
1137
+ "current": 2550.0,
1138
+ "min": 1500.0,
1139
+ "max": 2550.0
1140
+ },
1141
+ {
1142
+ "current": 2550.0,
1143
+ "min": 1500.0,
1144
+ "max": 2550.0
1145
+ },
1146
+ {
1147
+ "current": 2550.0,
1148
+ "min": 1500.0,
1149
+ "max": 2550.0
1150
+ },
1151
+ {
1152
+ "current": 2550.0,
1153
+ "min": 1500.0,
1154
+ "max": 2550.0
1155
+ },
1156
+ {
1157
+ "current": 2550.0,
1158
+ "min": 1500.0,
1159
+ "max": 2550.0
1160
+ },
1161
+ {
1162
+ "current": 2550.0,
1163
+ "min": 1500.0,
1164
+ "max": 2550.0
1165
+ },
1166
+ {
1167
+ "current": 2550.0,
1168
+ "min": 1500.0,
1169
+ "max": 2550.0
1170
+ },
1171
+ {
1172
+ "current": 2550.0,
1173
+ "min": 1500.0,
1174
+ "max": 2550.0
1175
+ },
1176
+ {
1177
+ "current": 2550.0,
1178
+ "min": 1500.0,
1179
+ "max": 2550.0
1180
+ },
1181
+ {
1182
+ "current": 2550.0,
1183
+ "min": 1500.0,
1184
+ "max": 2550.0
1185
+ },
1186
+ {
1187
+ "current": 2550.0,
1188
+ "min": 1500.0,
1189
+ "max": 2550.0
1190
+ },
1191
+ {
1192
+ "current": 2550.0,
1193
+ "min": 1500.0,
1194
+ "max": 2550.0
1195
+ },
1196
+ {
1197
+ "current": 2550.0,
1198
+ "min": 1500.0,
1199
+ "max": 2550.0
1200
+ },
1201
+ {
1202
+ "current": 2550.0,
1203
+ "min": 1500.0,
1204
+ "max": 2550.0
1205
+ },
1206
+ {
1207
+ "current": 2550.0,
1208
+ "min": 1500.0,
1209
+ "max": 2550.0
1210
+ },
1211
+ {
1212
+ "current": 2550.0,
1213
+ "min": 1500.0,
1214
+ "max": 2550.0
1215
+ },
1216
+ {
1217
+ "current": 2550.0,
1218
+ "min": 1500.0,
1219
+ "max": 2550.0
1220
+ },
1221
+ {
1222
+ "current": 2550.0,
1223
+ "min": 1500.0,
1224
+ "max": 2550.0
1225
+ },
1226
+ {
1227
+ "current": 2550.0,
1228
+ "min": 1500.0,
1229
+ "max": 2550.0
1230
+ },
1231
+ {
1232
+ "current": 2550.0,
1233
+ "min": 1500.0,
1234
+ "max": 2550.0
1235
+ },
1236
+ {
1237
+ "current": 2550.0,
1238
+ "min": 1500.0,
1239
+ "max": 2550.0
1240
+ },
1241
+ {
1242
+ "current": 2550.0,
1243
+ "min": 1500.0,
1244
+ "max": 2550.0
1245
+ },
1246
+ {
1247
+ "current": 2550.0,
1248
+ "min": 1500.0,
1249
+ "max": 2550.0
1250
+ },
1251
+ {
1252
+ "current": 2550.0,
1253
+ "min": 1500.0,
1254
+ "max": 2550.0
1255
+ },
1256
+ {
1257
+ "current": 2550.0,
1258
+ "min": 1500.0,
1259
+ "max": 2550.0
1260
+ },
1261
+ {
1262
+ "current": 2550.0,
1263
+ "min": 1500.0,
1264
+ "max": 2550.0
1265
+ },
1266
+ {
1267
+ "current": 2550.0,
1268
+ "min": 1500.0,
1269
+ "max": 2550.0
1270
+ },
1271
+ {
1272
+ "current": 2550.0,
1273
+ "min": 1500.0,
1274
+ "max": 2550.0
1275
+ },
1276
+ {
1277
+ "current": 2550.0,
1278
+ "min": 1500.0,
1279
+ "max": 2550.0
1280
+ },
1281
+ {
1282
+ "current": 2550.0,
1283
+ "min": 1500.0,
1284
+ "max": 2550.0
1285
+ },
1286
+ {
1287
+ "current": 2550.0,
1288
+ "min": 1500.0,
1289
+ "max": 2550.0
1290
+ },
1291
+ {
1292
+ "current": 2550.0,
1293
+ "min": 1500.0,
1294
+ "max": 2550.0
1295
+ },
1296
+ {
1297
+ "current": 2550.0,
1298
+ "min": 1500.0,
1299
+ "max": 2550.0
1300
+ },
1301
+ {
1302
+ "current": 2550.0,
1303
+ "min": 1500.0,
1304
+ "max": 2550.0
1305
+ },
1306
+ {
1307
+ "current": 2550.0,
1308
+ "min": 1500.0,
1309
+ "max": 2550.0
1310
+ },
1311
+ {
1312
+ "current": 2550.0,
1313
+ "min": 1500.0,
1314
+ "max": 2550.0
1315
+ }
1316
+ ],
1317
+ "disk": {
1318
+ "/": {
1319
+ "total": 242.1086883544922,
1320
+ "used": 64.85347366333008
1321
+ }
1322
+ },
1323
+ "gpu": "NVIDIA A100-SXM4-80GB",
1324
+ "gpu_count": 8,
1325
+ "gpu_devices": [
1326
+ {
1327
+ "name": "NVIDIA A100-SXM4-80GB",
1328
+ "memory_total": 85899345920
1329
+ },
1330
+ {
1331
+ "name": "NVIDIA A100-SXM4-80GB",
1332
+ "memory_total": 85899345920
1333
+ },
1334
+ {
1335
+ "name": "NVIDIA A100-SXM4-80GB",
1336
+ "memory_total": 85899345920
1337
+ },
1338
+ {
1339
+ "name": "NVIDIA A100-SXM4-80GB",
1340
+ "memory_total": 85899345920
1341
+ },
1342
+ {
1343
+ "name": "NVIDIA A100-SXM4-80GB",
1344
+ "memory_total": 85899345920
1345
+ },
1346
+ {
1347
+ "name": "NVIDIA A100-SXM4-80GB",
1348
+ "memory_total": 85899345920
1349
+ },
1350
+ {
1351
+ "name": "NVIDIA A100-SXM4-80GB",
1352
+ "memory_total": 85899345920
1353
+ },
1354
+ {
1355
+ "name": "NVIDIA A100-SXM4-80GB",
1356
+ "memory_total": 85899345920
1357
+ }
1358
+ ],
1359
+ "memory": {
1360
+ "total": 2015.6942749023438
1361
+ }
1362
+ }
stage1_qformer2_256/wandb/run-20240313_070329-pog6cg3f/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"Align/Step": 2180, "Align/Loss": 3.2451138496398926, "Align/Loss (Raw)": 3.248356580734253, "Align/Learning Rate": 5.510720031964844e-10, "Align/Step Time": 2.565823743119836, "_timestamp": 1710319009.287915, "_runtime": 5599.461001873016, "_step": 2180, "_wandb": {"runtime": 5634}}
stage1_qformer2_256/wandb/run-20240313_070329-pog6cg3f/logs/debug-internal.log ADDED
The diff for this file is too large to render. See raw diff
 
stage1_qformer2_256/wandb/run-20240313_070329-pog6cg3f/logs/debug.log ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-03-13 07:03:29,813 INFO MainThread:4087609 [wandb_setup.py:_flush():76] Current SDK version is 0.16.3
2
+ 2024-03-13 07:03:29,813 INFO MainThread:4087609 [wandb_setup.py:_flush():76] Configure stats pid to 4087609
3
+ 2024-03-13 07:03:29,813 INFO MainThread:4087609 [wandb_setup.py:_flush():76] Loading settings from /home/lilei/.config/wandb/settings
4
+ 2024-03-13 07:03:29,813 INFO MainThread:4087609 [wandb_setup.py:_flush():76] Loading settings from /home/lilei/prismatic-vlms/wandb/settings
5
+ 2024-03-13 07:03:29,813 INFO MainThread:4087609 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
6
+ 2024-03-13 07:03:29,813 INFO MainThread:4087609 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
7
+ 2024-03-13 07:03:29,813 INFO MainThread:4087609 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': 'scripts/pretrain.py', 'program_abspath': '/home/lilei/prismatic-vlms/scripts/pretrain.py', 'program': '/home/lilei/prismatic-vlms/scripts/pretrain.py'}
8
+ 2024-03-13 07:03:29,813 INFO MainThread:4087609 [wandb_init.py:_log_setup():526] Logging user logs to runs/stage1_qformer2_256/wandb/run-20240313_070329-pog6cg3f/logs/debug.log
9
+ 2024-03-13 07:03:29,813 INFO MainThread:4087609 [wandb_init.py:_log_setup():527] Logging internal logs to runs/stage1_qformer2_256/wandb/run-20240313_070329-pog6cg3f/logs/debug-internal.log
10
+ 2024-03-13 07:03:29,813 INFO MainThread:4087609 [wandb_init.py:init():566] calling init triggers
11
+ 2024-03-13 07:03:29,813 INFO MainThread:4087609 [wandb_init.py:init():573] wandb.init called with sweep_config: {}
12
+ config: {'model': {'type': 'one-stage+7b', 'model_id': 'one-stage+7b_qformer2_256', 'arch_specifier': 'qformer2_256', 'vision_backbone_id': 'clip-vit-l-336px', 'llm_backbone_id': 'vicuna-v15-7b', 'image_resize_strategy': 'letterbox', 'llm_max_length': 2048, 'align_epochs': 1, 'align_max_steps': None, 'align_global_batch_size': 256, 'align_per_device_batch_size': 16, 'align_learning_rate': 0.001, 'align_weight_decay': 0.0, 'align_max_grad_norm': 1.0, 'align_lr_scheduler_type': 'linear-warmup+cosine-decay', 'align_warmup_ratio': 0.03, 'align_train_strategy': 'fsdp-shard-grad-op', 'finetune_epochs': 1, 'finetune_max_steps': None, 'finetune_global_batch_size': 128, 'finetune_per_device_batch_size': 16, 'finetune_learning_rate': 2e-05, 'finetune_weight_decay': 0.1, 'finetune_max_grad_norm': 1.0, 'finetune_lr_scheduler_type': 'linear-warmup+cosine-decay', 'finetune_warmup_ratio': 0.03, 'finetune_train_strategy': 'fsdp-full-shard', 'enable_gradient_checkpointing': True, 'enable_mixed_precision_training': True, 'reduce_in_full_precision': False}, 'dataset': {'type': 'llava-v15', 'dataset_id': 'llava-v15', 'align_stage_components': ['download/llava-laion-cc-sbu-558k/chat.json', 'download/llava-laion-cc-sbu-558k'], 'finetune_stage_components': ['download/llava-v1.5-instruct/llava_v1_5_mix665k.json', 'download/llava-v1.5-instruct'], 'dataset_root_dir': 'data'}, 'stage': 'align', 'pretrained_checkpoint': None, 'run_id': 'stage1_qformer2_256', 'run_root_dir': 'runs', 'seed': 7, 'hf_token': '.hf_token', 'trackers': ['jsonl', 'wandb'], 'wandb_project': 'hf-vlms', 'wandb_entity': 'lilei_stones', 'max_length': 4096}
13
+ 2024-03-13 07:03:29,814 INFO MainThread:4087609 [wandb_init.py:init():616] starting backend
14
+ 2024-03-13 07:03:29,814 INFO MainThread:4087609 [wandb_init.py:init():620] setting up manager
15
+ 2024-03-13 07:03:29,820 INFO MainThread:4087609 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
16
+ 2024-03-13 07:03:29,826 INFO MainThread:4087609 [wandb_init.py:init():628] backend started and connected
17
+ 2024-03-13 07:03:29,838 INFO MainThread:4087609 [wandb_init.py:init():720] updated telemetry
18
+ 2024-03-13 07:03:29,872 INFO MainThread:4087609 [wandb_init.py:init():753] communicating run to backend with 90.0 second timeout
19
+ 2024-03-13 07:03:30,392 INFO MainThread:4087609 [wandb_run.py:_on_init():2262] communicating current version
20
+ 2024-03-13 07:03:30,435 INFO MainThread:4087609 [wandb_run.py:_on_init():2271] got version response upgrade_message: "wandb version 0.16.4 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
21
+
22
+ 2024-03-13 07:03:30,435 INFO MainThread:4087609 [wandb_init.py:init():804] starting run threads in backend
23
+ 2024-03-13 07:03:36,601 INFO MainThread:4087609 [wandb_run.py:_console_start():2241] atexit reg
24
+ 2024-03-13 07:03:36,601 INFO MainThread:4087609 [wandb_run.py:_redirect():2096] redirect: wrap_raw
25
+ 2024-03-13 07:03:36,601 INFO MainThread:4087609 [wandb_run.py:_redirect():2161] Wrapping output streams.
26
+ 2024-03-13 07:03:36,601 INFO MainThread:4087609 [wandb_run.py:_redirect():2186] Redirects installed.
27
+ 2024-03-13 07:03:36,602 INFO MainThread:4087609 [wandb_init.py:init():847] run started, returning control to user process
28
+ 2024-03-13 08:37:25,479 INFO MainThread:4087609 [wandb_run.py:_finish():1970] finishing run lilei_stones/hf-vlms/pog6cg3f
29
+ 2024-03-13 08:37:25,479 INFO MainThread:4087609 [wandb_run.py:_atexit_cleanup():2210] got exitcode: 0
30
+ 2024-03-13 08:37:25,479 INFO MainThread:4087609 [wandb_run.py:_restore():2193] restore
31
+ 2024-03-13 08:37:25,479 INFO MainThread:4087609 [wandb_run.py:_restore():2199] restore done
32
+ 2024-03-13 08:37:33,373 INFO MainThread:4087609 [wandb_run.py:_footer_history_summary_info():3866] rendering history
33
+ 2024-03-13 08:37:33,374 INFO MainThread:4087609 [wandb_run.py:_footer_history_summary_info():3898] rendering summary
34
+ 2024-03-13 08:37:33,383 INFO MainThread:4087609 [wandb_run.py:_footer_sync_info():3825] logging synced files
stage1_qformer2_256/wandb/run-20240313_070329-pog6cg3f/run-pog6cg3f.wandb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7476af5d166e0fec7bf04be617697089fc77890c6d27ff2288cf0744537c0edf
3
+ size 3183875
stage1_qformer2_32/checkpoints/latest-checkpoint.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b6ced7009dc5fc88089d52f6b21ceee3dd1078bff0fe1ac3272954cc8e4a1dc1
3
+ size 1864966320
stage1_qformer2_32/checkpoints/step-002180-epoch-00-loss=2.4617.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b6ced7009dc5fc88089d52f6b21ceee3dd1078bff0fe1ac3272954cc8e4a1dc1
3
+ size 1864966320
stage1_qformer2_32/config.json ADDED
@@ -0,0 +1,60 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset": {
3
+ "align_stage_components": [
4
+ "download/llava-laion-cc-sbu-558k/chat.json",
5
+ "download/llava-laion-cc-sbu-558k"
6
+ ],
7
+ "dataset_id": "llava-v15",
8
+ "dataset_root_dir": "data",
9
+ "finetune_stage_components": [
10
+ "download/llava-v1.5-instruct/llava_v1_5_mix665k.json",
11
+ "download/llava-v1.5-instruct"
12
+ ],
13
+ "type": "llava-v15"
14
+ },
15
+ "hf_token": ".hf_token",
16
+ "max_length": 4096,
17
+ "model": {
18
+ "align_epochs": 1,
19
+ "align_global_batch_size": 256,
20
+ "align_learning_rate": 0.001,
21
+ "align_lr_scheduler_type": "linear-warmup+cosine-decay",
22
+ "align_max_grad_norm": 1.0,
23
+ "align_max_steps": null,
24
+ "align_per_device_batch_size": 16,
25
+ "align_train_strategy": "fsdp-shard-grad-op",
26
+ "align_warmup_ratio": 0.03,
27
+ "align_weight_decay": 0.0,
28
+ "arch_specifier": "qformer2_32",
29
+ "enable_gradient_checkpointing": true,
30
+ "enable_mixed_precision_training": true,
31
+ "finetune_epochs": 1,
32
+ "finetune_global_batch_size": 128,
33
+ "finetune_learning_rate": 2e-05,
34
+ "finetune_lr_scheduler_type": "linear-warmup+cosine-decay",
35
+ "finetune_max_grad_norm": 1.0,
36
+ "finetune_max_steps": null,
37
+ "finetune_per_device_batch_size": 16,
38
+ "finetune_train_strategy": "fsdp-full-shard",
39
+ "finetune_warmup_ratio": 0.03,
40
+ "finetune_weight_decay": 0.1,
41
+ "image_resize_strategy": "letterbox",
42
+ "llm_backbone_id": "vicuna-v15-7b",
43
+ "llm_max_length": 2048,
44
+ "model_id": "one-stage+7b_qformer2_32",
45
+ "reduce_in_full_precision": false,
46
+ "type": "one-stage+7b",
47
+ "vision_backbone_id": "clip-vit-l-336px"
48
+ },
49
+ "pretrained_checkpoint": null,
50
+ "run_id": "stage1_qformer2_32",
51
+ "run_root_dir": "runs",
52
+ "seed": 7,
53
+ "stage": "align",
54
+ "trackers": [
55
+ "jsonl",
56
+ "wandb"
57
+ ],
58
+ "wandb_entity": "lilei_stones",
59
+ "wandb_project": "hf-vlms"
60
+ }
stage1_qformer2_32/config.yaml ADDED
@@ -0,0 +1,53 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ dataset:
2
+ align_stage_components:
3
+ - download/llava-laion-cc-sbu-558k/chat.json
4
+ - download/llava-laion-cc-sbu-558k
5
+ dataset_id: llava-v15
6
+ dataset_root_dir: data
7
+ finetune_stage_components:
8
+ - download/llava-v1.5-instruct/llava_v1_5_mix665k.json
9
+ - download/llava-v1.5-instruct
10
+ type: llava-v15
11
+ hf_token: .hf_token
12
+ max_length: 4096
13
+ model:
14
+ align_epochs: 1
15
+ align_global_batch_size: 256
16
+ align_learning_rate: 0.001
17
+ align_lr_scheduler_type: linear-warmup+cosine-decay
18
+ align_max_grad_norm: 1.0
19
+ align_max_steps: null
20
+ align_per_device_batch_size: 16
21
+ align_train_strategy: fsdp-shard-grad-op
22
+ align_warmup_ratio: 0.03
23
+ align_weight_decay: 0.0
24
+ arch_specifier: qformer2_32
25
+ enable_gradient_checkpointing: true
26
+ enable_mixed_precision_training: true
27
+ finetune_epochs: 1
28
+ finetune_global_batch_size: 128
29
+ finetune_learning_rate: 2.0e-05
30
+ finetune_lr_scheduler_type: linear-warmup+cosine-decay
31
+ finetune_max_grad_norm: 1.0
32
+ finetune_max_steps: null
33
+ finetune_per_device_batch_size: 16
34
+ finetune_train_strategy: fsdp-full-shard
35
+ finetune_warmup_ratio: 0.03
36
+ finetune_weight_decay: 0.1
37
+ image_resize_strategy: letterbox
38
+ llm_backbone_id: vicuna-v15-7b
39
+ llm_max_length: 2048
40
+ model_id: one-stage+7b_qformer2_32
41
+ reduce_in_full_precision: false
42
+ type: one-stage+7b
43
+ vision_backbone_id: clip-vit-l-336px
44
+ pretrained_checkpoint: null
45
+ run_id: stage1_qformer2_32
46
+ run_root_dir: runs
47
+ seed: 7
48
+ stage: align
49
+ trackers:
50
+ - jsonl
51
+ - wandb
52
+ wandb_entity: lilei_stones
53
+ wandb_project: hf-vlms
stage1_qformer2_32/run-metrics.jsonl ADDED
@@ -0,0 +1 @@
 
 
1
+ {"hparams": {"dataset": {"align_stage_components": ["download/llava-laion-cc-sbu-558k/chat.json", "download/llava-laion-cc-sbu-558k"], "dataset_id": "llava-v15", "dataset_root_dir": "data", "finetune_stage_components": ["download/llava-v1.5-instruct/llava_v1_5_mix665k.json", "download/llava-v1.5-instruct"], "type": "llava-v15"}, "hf_token": ".hf_token", "max_length": 4096, "model": {"align_epochs": 1, "align_global_batch_size": 256, "align_learning_rate": 0.001, "align_lr_scheduler_type": "linear-warmup+cosine-decay", "align_max_grad_norm": 1.0, "align_max_steps": null, "align_per_device_batch_size": 16, "align_train_strategy": "fsdp-shard-grad-op", "align_warmup_ratio": 0.03, "align_weight_decay": 0.0, "arch_specifier": "qformer2_32", "enable_gradient_checkpointing": true, "enable_mixed_precision_training": true, "finetune_epochs": 1, "finetune_global_batch_size": 128, "finetune_learning_rate": 2e-05, "finetune_lr_scheduler_type": "linear-warmup+cosine-decay", "finetune_max_grad_norm": 1.0, "finetune_max_steps": null, "finetune_per_device_batch_size": 16, "finetune_train_strategy": "fsdp-full-shard", "finetune_warmup_ratio": 0.03, "finetune_weight_decay": 0.1, "image_resize_strategy": "letterbox", "llm_backbone_id": "vicuna-v15-7b", "llm_max_length": 2048, "model_id": "one-stage+7b_qformer2_32", "reduce_in_full_precision": false, "type": "one-stage+7b", "vision_backbone_id": "clip-vit-l-336px"}, "pretrained_checkpoint": null, "run_id": "stage1_qformer2_32", "run_root_dir": "runs", "seed": 7, "stage": "align", "trackers": ["jsonl", "wandb"], "wandb_entity": "lilei_stones", "wandb_project": "hf-vlms"}, "run_id": "stage1_qformer2_32"}
stage1_qformer2_32/stage1_qformer2_32.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
stage1_qformer2_32/wandb/debug-internal.log ADDED
@@ -0,0 +1 @@
 
 
1
+ run-20240313_070204-8nxg74lb/logs/debug-internal.log
stage1_qformer2_32/wandb/debug.log ADDED
@@ -0,0 +1 @@
 
 
1
+ run-20240313_070204-8nxg74lb/logs/debug.log
stage1_qformer2_32/wandb/latest-run ADDED
@@ -0,0 +1 @@
 
 
1
+ run-20240313_070204-8nxg74lb
stage1_qformer2_32/wandb/run-20240313_070204-8nxg74lb/files/conda-environment.yaml ADDED
@@ -0,0 +1,194 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ name: pvlm
2
+ channels:
3
+ - pytorch
4
+ - nvidia
5
+ - defaults
6
+ dependencies:
7
+ - _libgcc_mutex=0.1=main
8
+ - _openmp_mutex=5.1=1_gnu
9
+ - blas=1.0=mkl
10
+ - bzip2=1.0.8=h7b6447c_0
11
+ - ca-certificates=2023.12.12=h06a4308_0
12
+ - certifi=2024.2.2=py310h06a4308_0
13
+ - charset-normalizer=2.0.4=pyhd3eb1b0_0
14
+ - cuda-cudart=11.8.89=0
15
+ - cuda-cupti=11.8.87=0
16
+ - cuda-libraries=11.8.0=0
17
+ - cuda-nvrtc=11.8.89=0
18
+ - cuda-nvtx=11.8.86=0
19
+ - cuda-runtime=11.8.0=0
20
+ - ffmpeg=4.3=hf484d3e_0
21
+ - filelock=3.13.1=py310h06a4308_0
22
+ - freetype=2.12.1=h4a9f257_0
23
+ - gmp=6.2.1=h295c915_3
24
+ - gmpy2=2.1.2=py310heeb90bb_0
25
+ - gnutls=3.6.15=he1e5248_0
26
+ - idna=3.4=py310h06a4308_0
27
+ - intel-openmp=2023.1.0=hdb19cb5_46306
28
+ - jinja2=3.1.3=py310h06a4308_0
29
+ - jpeg=9e=h5eee18b_1
30
+ - lame=3.100=h7b6447c_0
31
+ - lcms2=2.12=h3be6417_0
32
+ - ld_impl_linux-64=2.38=h1181459_1
33
+ - lerc=3.0=h295c915_0
34
+ - libcublas=11.11.3.6=0
35
+ - libcufft=10.9.0.58=0
36
+ - libcufile=1.8.1.2=0
37
+ - libcurand=10.3.4.107=0
38
+ - libcusolver=11.4.1.48=0
39
+ - libcusparse=11.7.5.86=0
40
+ - libdeflate=1.17=h5eee18b_1
41
+ - libffi=3.4.4=h6a678d5_0
42
+ - libgcc-ng=11.2.0=h1234567_1
43
+ - libgomp=11.2.0=h1234567_1
44
+ - libiconv=1.16=h7f8727e_2
45
+ - libidn2=2.3.4=h5eee18b_0
46
+ - libjpeg-turbo=2.0.0=h9bf148f_0
47
+ - libnpp=11.8.0.86=0
48
+ - libnvjpeg=11.9.0.86=0
49
+ - libpng=1.6.39=h5eee18b_0
50
+ - libstdcxx-ng=11.2.0=h1234567_1
51
+ - libtasn1=4.19.0=h5eee18b_0
52
+ - libtiff=4.5.1=h6a678d5_0
53
+ - libunistring=0.9.10=h27cfd23_0
54
+ - libuuid=1.41.5=h5eee18b_0
55
+ - libwebp-base=1.3.2=h5eee18b_0
56
+ - llvm-openmp=14.0.6=h9e868ea_0
57
+ - lz4-c=1.9.4=h6a678d5_0
58
+ - markupsafe=2.1.3=py310h5eee18b_0
59
+ - mkl=2023.1.0=h213fc3f_46344
60
+ - mkl-service=2.4.0=py310h5eee18b_1
61
+ - mkl_fft=1.3.8=py310h5eee18b_0
62
+ - mkl_random=1.2.4=py310hdb19cb5_0
63
+ - mpc=1.1.0=h10f8cd9_1
64
+ - mpfr=4.0.2=hb69a4c5_1
65
+ - mpmath=1.3.0=py310h06a4308_0
66
+ - ncurses=6.4=h6a678d5_0
67
+ - nettle=3.7.3=hbbd107a_1
68
+ - networkx=3.1=py310h06a4308_0
69
+ - numpy=1.26.4=py310h5f9d8c6_0
70
+ - numpy-base=1.26.4=py310hb5e798b_0
71
+ - openh264=2.1.1=h4ff587b_0
72
+ - openjpeg=2.4.0=h3ad879b_0
73
+ - openssl=3.0.13=h7f8727e_0
74
+ - pillow=10.2.0=py310h5eee18b_0
75
+ - pip=23.3.1=py310h06a4308_0
76
+ - python=3.10.13=h955ad1f_0
77
+ - pytorch=2.2.1=py3.10_cuda11.8_cudnn8.7.0_0
78
+ - pytorch-cuda=11.8=h7e8668a_5
79
+ - pytorch-mutex=1.0=cuda
80
+ - pyyaml=6.0.1=py310h5eee18b_0
81
+ - readline=8.2=h5eee18b_0
82
+ - requests=2.31.0=py310h06a4308_1
83
+ - setuptools=68.2.2=py310h06a4308_0
84
+ - sqlite=3.41.2=h5eee18b_0
85
+ - sympy=1.12=py310h06a4308_0
86
+ - tbb=2021.8.0=hdb19cb5_0
87
+ - tk=8.6.12=h1ccaba5_0
88
+ - torchaudio=2.2.1=py310_cu118
89
+ - torchtriton=2.2.0=py310
90
+ - torchvision=0.17.1=py310_cu118
91
+ - typing_extensions=4.9.0=py310h06a4308_1
92
+ - urllib3=2.1.0=py310h06a4308_0
93
+ - wheel=0.41.2=py310h06a4308_0
94
+ - xz=5.4.5=h5eee18b_0
95
+ - yaml=0.2.5=h7b6447c_0
96
+ - zlib=1.2.13=h5eee18b_0
97
+ - zstd=1.5.5=hc292b87_0
98
+ - pip:
99
+ - accelerate==0.27.2
100
+ - aiohttp==3.9.3
101
+ - aiosignal==1.3.1
102
+ - annotated-types==0.6.0
103
+ - appdirs==1.4.4
104
+ - async-timeout==4.0.3
105
+ - attrs==23.2.0
106
+ - beautifulsoup4==4.12.3
107
+ - blis==0.7.11
108
+ - catalogue==2.0.10
109
+ - click==8.1.7
110
+ - cloudpathlib==0.16.0
111
+ - confection==0.1.4
112
+ - contourpy==1.2.0
113
+ - cycler==0.12.1
114
+ - cymem==2.0.8
115
+ - datasets==2.18.0
116
+ - deepspeed==0.13.4
117
+ - dill==0.3.8
118
+ - docker-pycreds==0.4.0
119
+ - draccus==0.7.1
120
+ - einops==0.7.0
121
+ - flash-attn==2.5.5
122
+ - fonttools==4.49.0
123
+ - frozenlist==1.4.1
124
+ - fsspec==2024.2.0
125
+ - gdown==5.1.0
126
+ - gitdb==4.0.11
127
+ - gitpython==3.1.42
128
+ - hjson==3.1.0
129
+ - huggingface-hub==0.20.3
130
+ - install==1.3.5
131
+ - jsonlines==4.0.0
132
+ - kiwisolver==1.4.5
133
+ - langcodes==3.3.0
134
+ - markdown-it-py==3.0.0
135
+ - matplotlib==3.8.3
136
+ - mdurl==0.1.2
137
+ - mergedeep==1.3.4
138
+ - multidict==6.0.5
139
+ - multiprocess==0.70.16
140
+ - murmurhash==1.0.10
141
+ - mypy-extensions==1.0.0
142
+ - ninja==1.11.1.1
143
+ - packaging==23.2
144
+ - pandas==2.2.1
145
+ - preshed==3.0.9
146
+ - prismatic==0.0.1
147
+ - protobuf==4.25.3
148
+ - psutil==5.9.8
149
+ - py-cpuinfo==9.0.0
150
+ - pyarrow==15.0.0
151
+ - pyarrow-hotfix==0.6
152
+ - pycocoevalcap==1.2
153
+ - pycocotools==2.0.7
154
+ - pydantic==2.6.3
155
+ - pydantic-core==2.16.3
156
+ - pygments==2.17.2
157
+ - pynvml==11.5.0
158
+ - pyparsing==3.1.1
159
+ - pysocks==1.7.1
160
+ - pythainlp==5.0.1
161
+ - python-dateutil==2.9.0.post0
162
+ - pytz==2024.1
163
+ - pyyaml-include==1.3.2
164
+ - regex==2023.12.25
165
+ - rich==13.7.0
166
+ - safetensors==0.4.2
167
+ - sentencepiece==0.2.0
168
+ - sentry-sdk==1.40.5
169
+ - setproctitle==1.3.3
170
+ - six==1.16.0
171
+ - smart-open==6.4.0
172
+ - smmap==5.0.1
173
+ - soupsieve==2.5
174
+ - spacy==3.7.4
175
+ - spacy-legacy==3.0.12
176
+ - spacy-loggers==1.0.5
177
+ - srsly==2.4.8
178
+ - sudachidict-core==20240109
179
+ - sudachipy==0.6.8
180
+ - thinc==8.2.3
181
+ - timm==0.9.16
182
+ - tokenizers==0.15.2
183
+ - toml==0.10.2
184
+ - tqdm==4.66.2
185
+ - transformers==4.38.1
186
+ - typer==0.9.0
187
+ - typing-inspect==0.9.0
188
+ - tzdata==2024.1
189
+ - wandb==0.16.3
190
+ - wasabi==1.1.2
191
+ - weasel==0.3.4
192
+ - xxhash==3.4.1
193
+ - yarl==1.9.4
194
+ prefix: /home/lilei/anaconda3/envs/pvlm
stage1_qformer2_32/wandb/run-20240313_070204-8nxg74lb/files/config.yaml ADDED
@@ -0,0 +1,119 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wandb_version: 1
2
+
3
+ model:
4
+ desc: null
5
+ value:
6
+ type: one-stage+7b
7
+ model_id: one-stage+7b_qformer2_32
8
+ arch_specifier: qformer2_32
9
+ vision_backbone_id: clip-vit-l-336px
10
+ llm_backbone_id: vicuna-v15-7b
11
+ image_resize_strategy: letterbox
12
+ llm_max_length: 2048
13
+ align_epochs: 1
14
+ align_max_steps: null
15
+ align_global_batch_size: 256
16
+ align_per_device_batch_size: 16
17
+ align_learning_rate: 0.001
18
+ align_weight_decay: 0.0
19
+ align_max_grad_norm: 1.0
20
+ align_lr_scheduler_type: linear-warmup+cosine-decay
21
+ align_warmup_ratio: 0.03
22
+ align_train_strategy: fsdp-shard-grad-op
23
+ finetune_epochs: 1
24
+ finetune_max_steps: null
25
+ finetune_global_batch_size: 128
26
+ finetune_per_device_batch_size: 16
27
+ finetune_learning_rate: 2.0e-05
28
+ finetune_weight_decay: 0.1
29
+ finetune_max_grad_norm: 1.0
30
+ finetune_lr_scheduler_type: linear-warmup+cosine-decay
31
+ finetune_warmup_ratio: 0.03
32
+ finetune_train_strategy: fsdp-full-shard
33
+ enable_gradient_checkpointing: true
34
+ enable_mixed_precision_training: true
35
+ reduce_in_full_precision: false
36
+ dataset:
37
+ desc: null
38
+ value:
39
+ type: llava-v15
40
+ dataset_id: llava-v15
41
+ align_stage_components:
42
+ - download/llava-laion-cc-sbu-558k/chat.json
43
+ - download/llava-laion-cc-sbu-558k
44
+ finetune_stage_components:
45
+ - download/llava-v1.5-instruct/llava_v1_5_mix665k.json
46
+ - download/llava-v1.5-instruct
47
+ dataset_root_dir: data
48
+ stage:
49
+ desc: null
50
+ value: align
51
+ pretrained_checkpoint:
52
+ desc: null
53
+ value: null
54
+ run_id:
55
+ desc: null
56
+ value: stage1_qformer2_32
57
+ run_root_dir:
58
+ desc: null
59
+ value: runs
60
+ seed:
61
+ desc: null
62
+ value: 7
63
+ hf_token:
64
+ desc: null
65
+ value: .hf_token
66
+ trackers:
67
+ desc: null
68
+ value:
69
+ - jsonl
70
+ - wandb
71
+ wandb_project:
72
+ desc: null
73
+ value: hf-vlms
74
+ wandb_entity:
75
+ desc: null
76
+ value: lilei_stones
77
+ max_length:
78
+ desc: null
79
+ value: 4096
80
+ _wandb:
81
+ desc: null
82
+ value:
83
+ python_version: 3.10.13
84
+ cli_version: 0.16.3
85
+ framework: huggingface
86
+ huggingface_version: 4.38.1
87
+ is_jupyter_run: false
88
+ is_kaggle_kernel: false
89
+ start_time: 1710313324.485729
90
+ t:
91
+ 1:
92
+ - 1
93
+ - 11
94
+ - 41
95
+ - 49
96
+ - 51
97
+ - 55
98
+ - 63
99
+ - 71
100
+ 2:
101
+ - 1
102
+ - 11
103
+ - 41
104
+ - 49
105
+ - 51
106
+ - 55
107
+ - 63
108
+ - 71
109
+ 3:
110
+ - 2
111
+ - 13
112
+ - 16
113
+ - 23
114
+ 4: 3.10.13
115
+ 5: 0.16.3
116
+ 6: 4.38.1
117
+ 8:
118
+ - 5
119
+ 13: linux-x86_64
stage1_qformer2_32/wandb/run-20240313_070204-8nxg74lb/files/output.log ADDED
@@ -0,0 +1,1031 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ =>> [Global Step] 000000 =>> LR :: 0.000000: 0%| | 0/2180 [00:00<?, ?it/s]
3
+
4
+
5
+
6
+
7
+
8
+
9
+
10
+
11
+
12
+
13
+
14
+
15
+
16
+
17
+
18
+
19
+
20
+
21
+
22
+
23
+
24
+
25
+
26
+
27
+
28
+
29
+
30
+
31
+
32
+
33
+
34
+
35
+
36
+
37
+
38
+
39
+
40
+
41
+
42
+
43
+
44
+
45
+
46
+
47
+
48
+
49
+
50
+
51
+
52
+
53
+
54
+
55
+
56
+
57
+
58
+
59
+
60
+
61
+
62
+
63
+
64
+
65
+
66
+
67
+
68
+
69
+
70
+
71
+
72
+
73
+
74
+
75
+
76
+
77
+
78
+
79
+
80
+
81
+
82
+
83
+
84
+
85
+
86
+
87
+
88
+
89
+
90
+
91
+
92
+
93
+
94
+
95
+
96
+
97
+
98
+
99
+
100
+
101
+
102
+
103
+
104
+
105
+
106
+
107
+
108
+
109
+
110
+
111
+
112
+
113
+
114
+
115
+
116
+
117
+
118
+
119
+
120
+
121
+
122
+
123
+
124
+
125
+
126
+
127
+
128
+
129
+
130
+
131
+
132
+
133
+
134
+
135
+
136
+
137
+
138
+
139
+
140
+
141
+
142
+
143
+
144
+
145
+
146
+
147
+
148
+
149
+
150
+
151
+
152
+
153
+
154
+
155
+
156
+
157
+
158
+
159
+
160
+
161
+
162
+
163
+
164
+
165
+
166
+
167
+
168
+
169
+
170
+
171
+
172
+
173
+
174
+
175
+
176
+
177
+
178
+
179
+
180
+
181
+
182
+
183
+
184
+
185
+
186
+
187
+
188
+
189
+
190
+
191
+
192
+
193
+
194
+
195
+
196
+
197
+
198
+
199
+
200
+
201
+
202
+
203
+
204
+
205
+
206
+
207
+
208
+
209
+
210
+
211
+
212
+
213
+
214
+
215
+
216
+
217
+
218
+
219
+
220
+
221
+
222
+
223
+
224
+
225
+
226
+
227
+
228
+
229
+
230
+
231
+
232
+
233
+
234
+
235
+
236
+
237
+
238
+
239
+
240
+
241
+
242
+
243
+
244
+
245
+
246
+
247
+
248
+
249
+
250
+
251
+
252
+
253
+
254
+
255
+
256
+
257
+
258
+
259
+
260
+
261
+
262
+
263
+
264
+
265
+
266
+
267
+
268
+
269
+
270
+
271
+
272
+
273
+
274
+
275
+
276
+
277
+
278
+
279
+
280
+
281
+
282
+
283
+
284
+
285
+
286
+
287
+
288
+
289
+
290
+
291
+
292
+
293
+
294
+
295
+
296
+
297
+
298
+
299
+
300
+
301
+
302
+
303
+
304
+
305
+
306
+
307
+
308
+
309
+
310
+
311
+
312
+
313
+
314
+
315
+
316
+
317
+
318
+
319
+
320
+
321
+
322
+
323
+
324
+
325
+
326
+
327
+
328
+
329
+
330
+
331
+
332
+
333
+
334
+
335
+
336
+
337
+
338
+
339
+
340
+
341
+
342
+
343
+
344
+
345
+
346
+
347
+
348
+
349
+
350
+
351
+
352
+
353
+
354
+
355
+
356
+
357
+
358
+
359
+
360
+
361
+
362
+
363
+
364
+
365
+
366
+
367
+
368
+
369
+
370
+
371
+
372
+
373
+
374
+
375
+
376
+
377
+
378
+
379
+
380
+
381
+
382
+
383
+
384
+
385
+
386
+
387
+
388
+
389
+
390
+
391
+
392
+
393
+
394
+
395
+
396
+
397
+
398
+
399
+
400
+
401
+
402
+
403
+
404
+
405
+
406
+
407
+
408
+
409
+
410
+
411
+
412
+
413
+
414
+
415
+
416
+
417
+
418
+
419
+
420
+
421
+
422
+
423
+
424
+
425
+
426
+
427
+
428
+
429
+
430
+
431
+
432
+
433
+
434
+
435
+
436
+
437
+
438
+
439
+
440
+
441
+
442
+
443
+
444
+
445
+
446
+
447
+
448
+
449
+
450
+
451
+
452
+
453
+
454
+
455
+
456
+
457
+
458
+
459
+
460
+
461
+
462
+
463
+
464
+
465
+
466
+
467
+
468
+
469
+
470
+
471
+
472
+
473
+
474
+
475
+
476
+
477
+
478
+
479
+
480
+
481
+
482
+
483
+
484
+
485
+
486
+
487
+
488
+
489
+
490
+
491
+
492
+
493
+
494
+
495
+
496
+
497
+
498
+
499
+
500
+
501
+
502
+
503
+
504
+
505
+
506
+
507
+
508
+
509
+
510
+
511
+
512
+
513
+
514
+
515
+
516
+
517
+
518
+
519
+
520
+
521
+
522
+
523
+
524
+
525
+
526
+
527
+
528
+
529
+
530
+
531
+
532
+
533
+
534
+
535
+
536
+
537
+
538
+
539
+
540
+
541
+
542
+
543
+
544
+
545
+
546
+
547
+
548
+
549
+
550
+
551
+
552
+
553
+
554
+
555
+
556
+
557
+
558
+
559
+
560
+
561
+
562
+
563
+
564
+
565
+
566
+
567
+
568
+
569
+
570
+
571
+
572
+
573
+
574
+
575
+
576
+
577
+
578
+
579
+
580
+
581
+
582
+
583
+
584
+
585
+
586
+
587
+
588
+
589
+
590
+
591
+
592
+
593
+
594
+
595
+
596
+
597
+
598
+
599
+
600
+
601
+
602
+
603
+
604
+
605
+
606
+
607
+
608
+
609
+
610
+
611
+
612
+
613
+
614
+
615
+
616
+
617
+
618
+
619
+
620
+
621
+
622
+
623
+
624
+
625
+
626
+
627
+
628
+
629
+
630
+
631
+
632
+
633
+
634
+
635
+
636
+
637
+
638
+
639
+
640
+
641
+
642
+
643
+
644
+
645
+
646
+
647
+
648
+
649
+
650
+
651
+
652
+
653
+
654
+
655
+
656
+
657
+
658
+
659
+
660
+
661
+
662
+
663
+
664
+
665
+
666
+
667
+
668
+
669
+
670
+
671
+
672
+
673
+
674
+
675
+
676
+
677
+
678
+
679
+
680
+
681
+
682
+
683
+
684
+
685
+
686
+
687
+
688
+
689
+
690
+
691
+
692
+
693
+
694
+
695
+
696
+
697
+
698
+
699
+
700
+
701
+
702
+
703
+
704
+
705
+
706
+
707
+
708
+
709
+
710
+
711
+
712
+
713
+
714
+
715
+
716
+
717
+
718
+
719
+
720
+
721
+
722
+
723
+
724
+
725
+
726
+
727
+
728
+
729
+
730
+
731
+
732
+
733
+
734
+
735
+
736
+
737
+
738
+
739
+
740
+
741
+
742
+
743
+
744
+
745
+
746
+
747
+
748
+
749
+
750
+
751
+
752
+
753
+
754
+
755
+
756
+
757
+
758
+
759
+
760
+
761
+
762
+
763
+
764
+
765
+
766
+
767
+
768
+
769
+
770
+
771
+
772
+
773
+
774
+
775
+
776
+
777
+
778
+
779
+
780
+
781
+
782
+
783
+
784
+
785
+
786
+
787
+
788
+
789
+
790
+
791
+
792
+
793
+
794
+
795
+
796
+
797
+
798
+
799
+
800
+
801
+
802
+
803
+
804
+
805
+
806
+
807
+
808
+
809
+
810
+
811
+
812
+
813
+
814
+
815
+
816
+
817
+
818
+
819
+
820
+
821
+
822
+
823
+
824
+
825
+
826
+
827
+
828
+
829
+
830
+
831
+
832
+
833
+
834
+
835
+
836
+
837
+
838
+
839
+
840
+
841
+
842
+
843
+
844
+
845
+
846
+
847
+
848
+
849
+
850
+
851
+
852
+
853
+
854
+
855
+
856
+
857
+
858
+
859
+
860
+
861
+
862
+
863
+
864
+
865
+
866
+
867
+
868
+
869
+
870
+
871
+
872
+
873
+
874
+
875
+
876
+
877
+
878
+
879
+
880
+
881
+
882
+
883
+
884
+
885
+
886
+
887
+
888
+
889
+
890
+
891
+
892
+
893
+
894
+
895
+
896
+
897
+
898
+
899
+
900
+
901
+
902
+
903
+
904
+
905
+
906
+
907
+
908
+
909
+
910
+
911
+
912
+
913
+
914
+
915
+
916
+
917
+
918
+
919
+
920
+
921
+
922
+
923
+
924
+
925
+
926
+
927
+
928
+
929
+
930
+
931
+
932
+
933
+
934
+
935
+
936
+
937
+
938
+
939
+
940
+
941
+
942
+
943
+
944
+
945
+
946
+
947
+
948
+
949
+
950
+
951
+
952
+
953
+
954
+
955
+
956
+
957
+
958
+
959
+
960
+
961
+
962
+
963
+
964
+
965
+
966
+
967
+
968
+
969
+
970
+
971
+
972
+
973
+
974
+
975
+
976
+
977
+
978
+
979
+
980
+
981
+
982
+
983
+
984
+
985
+
986
+
987
+
988
+
989
+
990
+
991
+
992
+
993
+
994
+
995
+
996
+
997
+
998
+
999
+
1000
+
1001
+
1002
+
1003
+
1004
+
1005
+
1006
+
1007
+
1008
+
1009
+
1010
+
1011
+
1012
+
1013
+
1014
+
1015
+
1016
+
1017
+
1018
+
1019
+
1020
+
1021
+
1022
+
1023
+
1024
+
1025
+
1026
+
1027
+
1028
+
1029
+ =>> [Global Step] 002180 =>> LR :: 0.000000 -- Loss :: 3.2271: 100%|█| 2180/2180 [34:14<00:00, 1.09it/
1030
+
1031
+ 03/13 [07:37:05] INFO  | >> [*] Done with Training =>> Finalizing Metrics ]8;id=696414;file:///home/lilei/prismatic-vlms/scripts/pretrain.py\pretrain.py]8;;\:]8;id=68157;file:///home/lilei/prismatic-vlms/scripts/pretrain.py#230\230]8;;\
stage1_qformer2_32/wandb/run-20240313_070204-8nxg74lb/files/requirements.txt ADDED
@@ -0,0 +1,122 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ accelerate==0.27.2
2
+ aiohttp==3.9.3
3
+ aiosignal==1.3.1
4
+ annotated-types==0.6.0
5
+ appdirs==1.4.4
6
+ async-timeout==4.0.3
7
+ attrs==23.2.0
8
+ beautifulsoup4==4.12.3
9
+ blis==0.7.11
10
+ catalogue==2.0.10
11
+ certifi==2024.2.2
12
+ charset-normalizer==2.0.4
13
+ click==8.1.7
14
+ cloudpathlib==0.16.0
15
+ confection==0.1.4
16
+ contourpy==1.2.0
17
+ cycler==0.12.1
18
+ cymem==2.0.8
19
+ datasets==2.18.0
20
+ deepspeed==0.13.4
21
+ dill==0.3.8
22
+ docker-pycreds==0.4.0
23
+ draccus==0.7.1
24
+ einops==0.7.0
25
+ filelock==3.13.1
26
+ flash-attn==2.5.5
27
+ fonttools==4.49.0
28
+ frozenlist==1.4.1
29
+ fsspec==2024.2.0
30
+ gdown==5.1.0
31
+ gitdb==4.0.11
32
+ gitpython==3.1.42
33
+ gmpy2==2.1.2
34
+ hjson==3.1.0
35
+ huggingface-hub==0.20.3
36
+ idna==3.4
37
+ install==1.3.5
38
+ jinja2==3.1.3
39
+ jsonlines==4.0.0
40
+ kiwisolver==1.4.5
41
+ langcodes==3.3.0
42
+ markdown-it-py==3.0.0
43
+ markupsafe==2.1.3
44
+ matplotlib==3.8.3
45
+ mdurl==0.1.2
46
+ mergedeep==1.3.4
47
+ mkl-fft==1.3.8
48
+ mkl-random==1.2.4
49
+ mkl-service==2.4.0
50
+ mpmath==1.3.0
51
+ multidict==6.0.5
52
+ multiprocess==0.70.16
53
+ murmurhash==1.0.10
54
+ mypy-extensions==1.0.0
55
+ networkx==3.1
56
+ ninja==1.11.1.1
57
+ numpy==1.26.4
58
+ ocifs==1.3.1
59
+ packaging==23.2
60
+ pandas==2.2.1
61
+ pillow==10.2.0
62
+ pip==23.3.1
63
+ preshed==3.0.9
64
+ prismatic==0.0.1
65
+ protobuf==4.25.3
66
+ psutil==5.9.8
67
+ py-cpuinfo==9.0.0
68
+ pyarrow-hotfix==0.6
69
+ pyarrow==15.0.0
70
+ pycocoevalcap==1.2
71
+ pycocotools==2.0.7
72
+ pydantic-core==2.16.3
73
+ pydantic==2.6.3
74
+ pygments==2.17.2
75
+ pynvml==11.5.0
76
+ pyparsing==3.1.1
77
+ pysocks==1.7.1
78
+ pythainlp==5.0.1
79
+ python-dateutil==2.9.0.post0
80
+ pytz==2024.1
81
+ pyyaml-include==1.3.2
82
+ pyyaml==6.0.1
83
+ regex==2023.12.25
84
+ requests==2.31.0
85
+ rich==13.7.0
86
+ safetensors==0.4.2
87
+ sentencepiece==0.2.0
88
+ sentry-sdk==1.40.5
89
+ setproctitle==1.3.3
90
+ setuptools==68.2.2
91
+ six==1.16.0
92
+ smart-open==6.4.0
93
+ smmap==5.0.1
94
+ soupsieve==2.5
95
+ spacy-legacy==3.0.12
96
+ spacy-loggers==1.0.5
97
+ spacy==3.7.4
98
+ srsly==2.4.8
99
+ sudachidict-core==20240109
100
+ sudachipy==0.6.8
101
+ sympy==1.12
102
+ thinc==8.2.3
103
+ timm==0.9.16
104
+ tokenizers==0.15.2
105
+ toml==0.10.2
106
+ torch==2.2.1
107
+ torchaudio==2.2.1
108
+ torchvision==0.17.1
109
+ tqdm==4.66.2
110
+ transformers==4.38.1
111
+ triton==2.2.0
112
+ typer==0.9.0
113
+ typing-extensions==4.9.0
114
+ typing-inspect==0.9.0
115
+ tzdata==2024.1
116
+ urllib3==2.1.0
117
+ wandb==0.16.3
118
+ wasabi==1.1.2
119
+ weasel==0.3.4
120
+ wheel==0.41.2
121
+ xxhash==3.4.1
122
+ yarl==1.9.4
stage1_qformer2_32/wandb/run-20240313_070204-8nxg74lb/files/wandb-metadata.json ADDED
@@ -0,0 +1,1362 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.15.0-1027-oracle-x86_64-with-glibc2.31",
3
+ "python": "3.10.13",
4
+ "heartbeatAt": "2024-03-13T07:02:05.249307",
5
+ "startedAt": "2024-03-13T07:02:04.427744",
6
+ "docker": null,
7
+ "cuda": null,
8
+ "args": [
9
+ "--model.type",
10
+ "one-stage+7b",
11
+ "--stage",
12
+ "align",
13
+ "--model.model_id",
14
+ "one-stage+7b_qformer2_32",
15
+ "--run_id",
16
+ "stage1_qformer2_32",
17
+ "--model.arch_specifier",
18
+ "qformer2_32"
19
+ ],
20
+ "state": "running",
21
+ "program": "/home/lilei/prismatic-vlms/scripts/pretrain.py",
22
+ "codePathLocal": "scripts/pretrain.py",
23
+ "codePath": "scripts/pretrain.py",
24
+ "git": {
25
+ "remote": "https://github.com/TRI-ML/prismatic-vlms.git",
26
+ "commit": "e87c4d04915bf8fad0d90f2ae79454df778ee2d3"
27
+ },
28
+ "email": "tobiaslee@qq.com",
29
+ "root": "/home/lilei/prismatic-vlms",
30
+ "host": "compute-permanent-node-824",
31
+ "username": "lilei",
32
+ "executable": "/home/lilei/anaconda3/envs/pvlm/bin/python",
33
+ "cpu_count": 128,
34
+ "cpu_count_logical": 255,
35
+ "cpu_freq": {
36
+ "current": 2574.0054,
37
+ "min": 1500.0,
38
+ "max": 2550.0
39
+ },
40
+ "cpu_freq_per_core": [
41
+ {
42
+ "current": 2550.0,
43
+ "min": 1500.0,
44
+ "max": 2550.0
45
+ },
46
+ {
47
+ "current": 2550.0,
48
+ "min": 1500.0,
49
+ "max": 2550.0
50
+ },
51
+ {
52
+ "current": 2550.0,
53
+ "min": 1500.0,
54
+ "max": 2550.0
55
+ },
56
+ {
57
+ "current": 2550.0,
58
+ "min": 1500.0,
59
+ "max": 2550.0
60
+ },
61
+ {
62
+ "current": 2550.0,
63
+ "min": 1500.0,
64
+ "max": 2550.0
65
+ },
66
+ {
67
+ "current": 2550.0,
68
+ "min": 1500.0,
69
+ "max": 2550.0
70
+ },
71
+ {
72
+ "current": 2550.0,
73
+ "min": 1500.0,
74
+ "max": 2550.0
75
+ },
76
+ {
77
+ "current": 2550.0,
78
+ "min": 1500.0,
79
+ "max": 2550.0
80
+ },
81
+ {
82
+ "current": 2550.0,
83
+ "min": 1500.0,
84
+ "max": 2550.0
85
+ },
86
+ {
87
+ "current": 3249.805,
88
+ "min": 1500.0,
89
+ "max": 2550.0
90
+ },
91
+ {
92
+ "current": 2550.0,
93
+ "min": 1500.0,
94
+ "max": 2550.0
95
+ },
96
+ {
97
+ "current": 2550.0,
98
+ "min": 1500.0,
99
+ "max": 2550.0
100
+ },
101
+ {
102
+ "current": 2550.0,
103
+ "min": 1500.0,
104
+ "max": 2550.0
105
+ },
106
+ {
107
+ "current": 2550.0,
108
+ "min": 1500.0,
109
+ "max": 2550.0
110
+ },
111
+ {
112
+ "current": 2550.0,
113
+ "min": 1500.0,
114
+ "max": 2550.0
115
+ },
116
+ {
117
+ "current": 2550.0,
118
+ "min": 1500.0,
119
+ "max": 2550.0
120
+ },
121
+ {
122
+ "current": 3247.271,
123
+ "min": 1500.0,
124
+ "max": 2550.0
125
+ },
126
+ {
127
+ "current": 2550.0,
128
+ "min": 1500.0,
129
+ "max": 2550.0
130
+ },
131
+ {
132
+ "current": 2550.0,
133
+ "min": 1500.0,
134
+ "max": 2550.0
135
+ },
136
+ {
137
+ "current": 2550.0,
138
+ "min": 1500.0,
139
+ "max": 2550.0
140
+ },
141
+ {
142
+ "current": 2550.0,
143
+ "min": 1500.0,
144
+ "max": 2550.0
145
+ },
146
+ {
147
+ "current": 2550.0,
148
+ "min": 1500.0,
149
+ "max": 2550.0
150
+ },
151
+ {
152
+ "current": 2550.0,
153
+ "min": 1500.0,
154
+ "max": 2550.0
155
+ },
156
+ {
157
+ "current": 2550.0,
158
+ "min": 1500.0,
159
+ "max": 2550.0
160
+ },
161
+ {
162
+ "current": 2550.0,
163
+ "min": 1500.0,
164
+ "max": 2550.0
165
+ },
166
+ {
167
+ "current": 2550.0,
168
+ "min": 1500.0,
169
+ "max": 2550.0
170
+ },
171
+ {
172
+ "current": 2550.0,
173
+ "min": 1500.0,
174
+ "max": 2550.0
175
+ },
176
+ {
177
+ "current": 2550.0,
178
+ "min": 1500.0,
179
+ "max": 2550.0
180
+ },
181
+ {
182
+ "current": 2550.0,
183
+ "min": 1500.0,
184
+ "max": 2550.0
185
+ },
186
+ {
187
+ "current": 2550.0,
188
+ "min": 1500.0,
189
+ "max": 2550.0
190
+ },
191
+ {
192
+ "current": 2550.0,
193
+ "min": 1500.0,
194
+ "max": 2550.0
195
+ },
196
+ {
197
+ "current": 2550.0,
198
+ "min": 1500.0,
199
+ "max": 2550.0
200
+ },
201
+ {
202
+ "current": 2550.0,
203
+ "min": 1500.0,
204
+ "max": 2550.0
205
+ },
206
+ {
207
+ "current": 2550.0,
208
+ "min": 1500.0,
209
+ "max": 2550.0
210
+ },
211
+ {
212
+ "current": 2550.0,
213
+ "min": 1500.0,
214
+ "max": 2550.0
215
+ },
216
+ {
217
+ "current": 2550.0,
218
+ "min": 1500.0,
219
+ "max": 2550.0
220
+ },
221
+ {
222
+ "current": 2550.0,
223
+ "min": 1500.0,
224
+ "max": 2550.0
225
+ },
226
+ {
227
+ "current": 2550.0,
228
+ "min": 1500.0,
229
+ "max": 2550.0
230
+ },
231
+ {
232
+ "current": 2550.0,
233
+ "min": 1500.0,
234
+ "max": 2550.0
235
+ },
236
+ {
237
+ "current": 2550.0,
238
+ "min": 1500.0,
239
+ "max": 2550.0
240
+ },
241
+ {
242
+ "current": 2550.0,
243
+ "min": 1500.0,
244
+ "max": 2550.0
245
+ },
246
+ {
247
+ "current": 2550.0,
248
+ "min": 1500.0,
249
+ "max": 2550.0
250
+ },
251
+ {
252
+ "current": 2550.0,
253
+ "min": 1500.0,
254
+ "max": 2550.0
255
+ },
256
+ {
257
+ "current": 2550.0,
258
+ "min": 1500.0,
259
+ "max": 2550.0
260
+ },
261
+ {
262
+ "current": 2550.0,
263
+ "min": 1500.0,
264
+ "max": 2550.0
265
+ },
266
+ {
267
+ "current": 2550.0,
268
+ "min": 1500.0,
269
+ "max": 2550.0
270
+ },
271
+ {
272
+ "current": 2550.0,
273
+ "min": 1500.0,
274
+ "max": 2550.0
275
+ },
276
+ {
277
+ "current": 2550.0,
278
+ "min": 1500.0,
279
+ "max": 2550.0
280
+ },
281
+ {
282
+ "current": 2550.0,
283
+ "min": 1500.0,
284
+ "max": 2550.0
285
+ },
286
+ {
287
+ "current": 2550.0,
288
+ "min": 1500.0,
289
+ "max": 2550.0
290
+ },
291
+ {
292
+ "current": 2550.0,
293
+ "min": 1500.0,
294
+ "max": 2550.0
295
+ },
296
+ {
297
+ "current": 2550.0,
298
+ "min": 1500.0,
299
+ "max": 2550.0
300
+ },
301
+ {
302
+ "current": 2550.0,
303
+ "min": 1500.0,
304
+ "max": 2550.0
305
+ },
306
+ {
307
+ "current": 2550.0,
308
+ "min": 1500.0,
309
+ "max": 2550.0
310
+ },
311
+ {
312
+ "current": 2550.0,
313
+ "min": 1500.0,
314
+ "max": 2550.0
315
+ },
316
+ {
317
+ "current": 2550.0,
318
+ "min": 1500.0,
319
+ "max": 2550.0
320
+ },
321
+ {
322
+ "current": 2550.0,
323
+ "min": 1500.0,
324
+ "max": 2550.0
325
+ },
326
+ {
327
+ "current": 2550.0,
328
+ "min": 1500.0,
329
+ "max": 2550.0
330
+ },
331
+ {
332
+ "current": 2550.0,
333
+ "min": 1500.0,
334
+ "max": 2550.0
335
+ },
336
+ {
337
+ "current": 2550.0,
338
+ "min": 1500.0,
339
+ "max": 2550.0
340
+ },
341
+ {
342
+ "current": 2550.0,
343
+ "min": 1500.0,
344
+ "max": 2550.0
345
+ },
346
+ {
347
+ "current": 2550.0,
348
+ "min": 1500.0,
349
+ "max": 2550.0
350
+ },
351
+ {
352
+ "current": 3249.381,
353
+ "min": 1500.0,
354
+ "max": 2550.0
355
+ },
356
+ {
357
+ "current": 2550.0,
358
+ "min": 1500.0,
359
+ "max": 2550.0
360
+ },
361
+ {
362
+ "current": 2550.0,
363
+ "min": 1500.0,
364
+ "max": 2550.0
365
+ },
366
+ {
367
+ "current": 2550.0,
368
+ "min": 1500.0,
369
+ "max": 2550.0
370
+ },
371
+ {
372
+ "current": 2550.0,
373
+ "min": 1500.0,
374
+ "max": 2550.0
375
+ },
376
+ {
377
+ "current": 2550.0,
378
+ "min": 1500.0,
379
+ "max": 2550.0
380
+ },
381
+ {
382
+ "current": 2550.0,
383
+ "min": 1500.0,
384
+ "max": 2550.0
385
+ },
386
+ {
387
+ "current": 2550.0,
388
+ "min": 1500.0,
389
+ "max": 2550.0
390
+ },
391
+ {
392
+ "current": 2550.0,
393
+ "min": 1500.0,
394
+ "max": 2550.0
395
+ },
396
+ {
397
+ "current": 2550.0,
398
+ "min": 1500.0,
399
+ "max": 2550.0
400
+ },
401
+ {
402
+ "current": 2550.0,
403
+ "min": 1500.0,
404
+ "max": 2550.0
405
+ },
406
+ {
407
+ "current": 2550.0,
408
+ "min": 1500.0,
409
+ "max": 2550.0
410
+ },
411
+ {
412
+ "current": 2550.0,
413
+ "min": 1500.0,
414
+ "max": 2550.0
415
+ },
416
+ {
417
+ "current": 2550.0,
418
+ "min": 1500.0,
419
+ "max": 2550.0
420
+ },
421
+ {
422
+ "current": 2550.0,
423
+ "min": 1500.0,
424
+ "max": 2550.0
425
+ },
426
+ {
427
+ "current": 2550.0,
428
+ "min": 1500.0,
429
+ "max": 2550.0
430
+ },
431
+ {
432
+ "current": 2550.0,
433
+ "min": 1500.0,
434
+ "max": 2550.0
435
+ },
436
+ {
437
+ "current": 2550.0,
438
+ "min": 1500.0,
439
+ "max": 2550.0
440
+ },
441
+ {
442
+ "current": 2550.0,
443
+ "min": 1500.0,
444
+ "max": 2550.0
445
+ },
446
+ {
447
+ "current": 2550.0,
448
+ "min": 1500.0,
449
+ "max": 2550.0
450
+ },
451
+ {
452
+ "current": 2550.0,
453
+ "min": 1500.0,
454
+ "max": 2550.0
455
+ },
456
+ {
457
+ "current": 2550.0,
458
+ "min": 1500.0,
459
+ "max": 2550.0
460
+ },
461
+ {
462
+ "current": 2550.0,
463
+ "min": 1500.0,
464
+ "max": 2550.0
465
+ },
466
+ {
467
+ "current": 2550.0,
468
+ "min": 1500.0,
469
+ "max": 2550.0
470
+ },
471
+ {
472
+ "current": 2550.0,
473
+ "min": 1500.0,
474
+ "max": 2550.0
475
+ },
476
+ {
477
+ "current": 2550.0,
478
+ "min": 1500.0,
479
+ "max": 2550.0
480
+ },
481
+ {
482
+ "current": 2550.0,
483
+ "min": 1500.0,
484
+ "max": 2550.0
485
+ },
486
+ {
487
+ "current": 2550.0,
488
+ "min": 1500.0,
489
+ "max": 2550.0
490
+ },
491
+ {
492
+ "current": 2550.0,
493
+ "min": 1500.0,
494
+ "max": 2550.0
495
+ },
496
+ {
497
+ "current": 2550.0,
498
+ "min": 1500.0,
499
+ "max": 2550.0
500
+ },
501
+ {
502
+ "current": 2550.0,
503
+ "min": 1500.0,
504
+ "max": 2550.0
505
+ },
506
+ {
507
+ "current": 2550.0,
508
+ "min": 1500.0,
509
+ "max": 2550.0
510
+ },
511
+ {
512
+ "current": 2550.0,
513
+ "min": 1500.0,
514
+ "max": 2550.0
515
+ },
516
+ {
517
+ "current": 2550.0,
518
+ "min": 1500.0,
519
+ "max": 2550.0
520
+ },
521
+ {
522
+ "current": 2550.0,
523
+ "min": 1500.0,
524
+ "max": 2550.0
525
+ },
526
+ {
527
+ "current": 2550.0,
528
+ "min": 1500.0,
529
+ "max": 2550.0
530
+ },
531
+ {
532
+ "current": 2550.0,
533
+ "min": 1500.0,
534
+ "max": 2550.0
535
+ },
536
+ {
537
+ "current": 2550.0,
538
+ "min": 1500.0,
539
+ "max": 2550.0
540
+ },
541
+ {
542
+ "current": 2550.0,
543
+ "min": 1500.0,
544
+ "max": 2550.0
545
+ },
546
+ {
547
+ "current": 2550.0,
548
+ "min": 1500.0,
549
+ "max": 2550.0
550
+ },
551
+ {
552
+ "current": 2550.0,
553
+ "min": 1500.0,
554
+ "max": 2550.0
555
+ },
556
+ {
557
+ "current": 2550.0,
558
+ "min": 1500.0,
559
+ "max": 2550.0
560
+ },
561
+ {
562
+ "current": 2550.0,
563
+ "min": 1500.0,
564
+ "max": 2550.0
565
+ },
566
+ {
567
+ "current": 2550.0,
568
+ "min": 1500.0,
569
+ "max": 2550.0
570
+ },
571
+ {
572
+ "current": 2550.0,
573
+ "min": 1500.0,
574
+ "max": 2550.0
575
+ },
576
+ {
577
+ "current": 2550.0,
578
+ "min": 1500.0,
579
+ "max": 2550.0
580
+ },
581
+ {
582
+ "current": 2550.0,
583
+ "min": 1500.0,
584
+ "max": 2550.0
585
+ },
586
+ {
587
+ "current": 2550.0,
588
+ "min": 1500.0,
589
+ "max": 2550.0
590
+ },
591
+ {
592
+ "current": 2550.0,
593
+ "min": 1500.0,
594
+ "max": 2550.0
595
+ },
596
+ {
597
+ "current": 2550.0,
598
+ "min": 1500.0,
599
+ "max": 2550.0
600
+ },
601
+ {
602
+ "current": 2550.0,
603
+ "min": 1500.0,
604
+ "max": 2550.0
605
+ },
606
+ {
607
+ "current": 2550.0,
608
+ "min": 1500.0,
609
+ "max": 2550.0
610
+ },
611
+ {
612
+ "current": 2550.0,
613
+ "min": 1500.0,
614
+ "max": 2550.0
615
+ },
616
+ {
617
+ "current": 2550.0,
618
+ "min": 1500.0,
619
+ "max": 2550.0
620
+ },
621
+ {
622
+ "current": 2550.0,
623
+ "min": 1500.0,
624
+ "max": 2550.0
625
+ },
626
+ {
627
+ "current": 2550.0,
628
+ "min": 1500.0,
629
+ "max": 2550.0
630
+ },
631
+ {
632
+ "current": 2550.0,
633
+ "min": 1500.0,
634
+ "max": 2550.0
635
+ },
636
+ {
637
+ "current": 2550.0,
638
+ "min": 1500.0,
639
+ "max": 2550.0
640
+ },
641
+ {
642
+ "current": 2550.0,
643
+ "min": 1500.0,
644
+ "max": 2550.0
645
+ },
646
+ {
647
+ "current": 2550.0,
648
+ "min": 1500.0,
649
+ "max": 2550.0
650
+ },
651
+ {
652
+ "current": 2550.0,
653
+ "min": 1500.0,
654
+ "max": 2550.0
655
+ },
656
+ {
657
+ "current": 2550.0,
658
+ "min": 1500.0,
659
+ "max": 2550.0
660
+ },
661
+ {
662
+ "current": 2550.0,
663
+ "min": 1500.0,
664
+ "max": 2550.0
665
+ },
666
+ {
667
+ "current": 2550.0,
668
+ "min": 1500.0,
669
+ "max": 2550.0
670
+ },
671
+ {
672
+ "current": 2550.0,
673
+ "min": 1500.0,
674
+ "max": 2550.0
675
+ },
676
+ {
677
+ "current": 2550.0,
678
+ "min": 1500.0,
679
+ "max": 2550.0
680
+ },
681
+ {
682
+ "current": 2550.0,
683
+ "min": 1500.0,
684
+ "max": 2550.0
685
+ },
686
+ {
687
+ "current": 2550.0,
688
+ "min": 1500.0,
689
+ "max": 2550.0
690
+ },
691
+ {
692
+ "current": 2550.0,
693
+ "min": 1500.0,
694
+ "max": 2550.0
695
+ },
696
+ {
697
+ "current": 3086.157,
698
+ "min": 1500.0,
699
+ "max": 2550.0
700
+ },
701
+ {
702
+ "current": 3249.17,
703
+ "min": 1500.0,
704
+ "max": 2550.0
705
+ },
706
+ {
707
+ "current": 2550.0,
708
+ "min": 1500.0,
709
+ "max": 2550.0
710
+ },
711
+ {
712
+ "current": 2550.0,
713
+ "min": 1500.0,
714
+ "max": 2550.0
715
+ },
716
+ {
717
+ "current": 2550.0,
718
+ "min": 1500.0,
719
+ "max": 2550.0
720
+ },
721
+ {
722
+ "current": 2550.0,
723
+ "min": 1500.0,
724
+ "max": 2550.0
725
+ },
726
+ {
727
+ "current": 2550.0,
728
+ "min": 1500.0,
729
+ "max": 2550.0
730
+ },
731
+ {
732
+ "current": 2550.0,
733
+ "min": 1500.0,
734
+ "max": 2550.0
735
+ },
736
+ {
737
+ "current": 2550.0,
738
+ "min": 1500.0,
739
+ "max": 2550.0
740
+ },
741
+ {
742
+ "current": 2550.0,
743
+ "min": 1500.0,
744
+ "max": 2550.0
745
+ },
746
+ {
747
+ "current": 2550.0,
748
+ "min": 1500.0,
749
+ "max": 2550.0
750
+ },
751
+ {
752
+ "current": 2550.0,
753
+ "min": 1500.0,
754
+ "max": 2550.0
755
+ },
756
+ {
757
+ "current": 2550.0,
758
+ "min": 1500.0,
759
+ "max": 2550.0
760
+ },
761
+ {
762
+ "current": 2550.0,
763
+ "min": 1500.0,
764
+ "max": 2550.0
765
+ },
766
+ {
767
+ "current": 2550.0,
768
+ "min": 1500.0,
769
+ "max": 2550.0
770
+ },
771
+ {
772
+ "current": 2550.0,
773
+ "min": 1500.0,
774
+ "max": 2550.0
775
+ },
776
+ {
777
+ "current": 2550.0,
778
+ "min": 1500.0,
779
+ "max": 2550.0
780
+ },
781
+ {
782
+ "current": 2550.0,
783
+ "min": 1500.0,
784
+ "max": 2550.0
785
+ },
786
+ {
787
+ "current": 2550.0,
788
+ "min": 1500.0,
789
+ "max": 2550.0
790
+ },
791
+ {
792
+ "current": 2550.0,
793
+ "min": 1500.0,
794
+ "max": 2550.0
795
+ },
796
+ {
797
+ "current": 2550.0,
798
+ "min": 1500.0,
799
+ "max": 2550.0
800
+ },
801
+ {
802
+ "current": 2550.0,
803
+ "min": 1500.0,
804
+ "max": 2550.0
805
+ },
806
+ {
807
+ "current": 2550.0,
808
+ "min": 1500.0,
809
+ "max": 2550.0
810
+ },
811
+ {
812
+ "current": 2550.0,
813
+ "min": 1500.0,
814
+ "max": 2550.0
815
+ },
816
+ {
817
+ "current": 2550.0,
818
+ "min": 1500.0,
819
+ "max": 2550.0
820
+ },
821
+ {
822
+ "current": 2550.0,
823
+ "min": 1500.0,
824
+ "max": 2550.0
825
+ },
826
+ {
827
+ "current": 3248.978,
828
+ "min": 1500.0,
829
+ "max": 2550.0
830
+ },
831
+ {
832
+ "current": 2550.0,
833
+ "min": 1500.0,
834
+ "max": 2550.0
835
+ },
836
+ {
837
+ "current": 2550.0,
838
+ "min": 1500.0,
839
+ "max": 2550.0
840
+ },
841
+ {
842
+ "current": 2550.0,
843
+ "min": 1500.0,
844
+ "max": 2550.0
845
+ },
846
+ {
847
+ "current": 2550.0,
848
+ "min": 1500.0,
849
+ "max": 2550.0
850
+ },
851
+ {
852
+ "current": 2550.0,
853
+ "min": 1500.0,
854
+ "max": 2550.0
855
+ },
856
+ {
857
+ "current": 2550.0,
858
+ "min": 1500.0,
859
+ "max": 2550.0
860
+ },
861
+ {
862
+ "current": 2550.0,
863
+ "min": 1500.0,
864
+ "max": 2550.0
865
+ },
866
+ {
867
+ "current": 2550.0,
868
+ "min": 1500.0,
869
+ "max": 2550.0
870
+ },
871
+ {
872
+ "current": 2550.0,
873
+ "min": 1500.0,
874
+ "max": 2550.0
875
+ },
876
+ {
877
+ "current": 2550.0,
878
+ "min": 1500.0,
879
+ "max": 2550.0
880
+ },
881
+ {
882
+ "current": 2550.0,
883
+ "min": 1500.0,
884
+ "max": 2550.0
885
+ },
886
+ {
887
+ "current": 2550.0,
888
+ "min": 1500.0,
889
+ "max": 2550.0
890
+ },
891
+ {
892
+ "current": 2550.0,
893
+ "min": 1500.0,
894
+ "max": 2550.0
895
+ },
896
+ {
897
+ "current": 3249.405,
898
+ "min": 1500.0,
899
+ "max": 2550.0
900
+ },
901
+ {
902
+ "current": 2550.0,
903
+ "min": 1500.0,
904
+ "max": 2550.0
905
+ },
906
+ {
907
+ "current": 2550.0,
908
+ "min": 1500.0,
909
+ "max": 2550.0
910
+ },
911
+ {
912
+ "current": 2550.0,
913
+ "min": 1500.0,
914
+ "max": 2550.0
915
+ },
916
+ {
917
+ "current": 2550.0,
918
+ "min": 1500.0,
919
+ "max": 2550.0
920
+ },
921
+ {
922
+ "current": 2550.0,
923
+ "min": 1500.0,
924
+ "max": 2550.0
925
+ },
926
+ {
927
+ "current": 2550.0,
928
+ "min": 1500.0,
929
+ "max": 2550.0
930
+ },
931
+ {
932
+ "current": 2550.0,
933
+ "min": 1500.0,
934
+ "max": 2550.0
935
+ },
936
+ {
937
+ "current": 2550.0,
938
+ "min": 1500.0,
939
+ "max": 2550.0
940
+ },
941
+ {
942
+ "current": 2550.0,
943
+ "min": 1500.0,
944
+ "max": 2550.0
945
+ },
946
+ {
947
+ "current": 2550.0,
948
+ "min": 1500.0,
949
+ "max": 2550.0
950
+ },
951
+ {
952
+ "current": 2550.0,
953
+ "min": 1500.0,
954
+ "max": 2550.0
955
+ },
956
+ {
957
+ "current": 3249.423,
958
+ "min": 1500.0,
959
+ "max": 2550.0
960
+ },
961
+ {
962
+ "current": 2550.0,
963
+ "min": 1500.0,
964
+ "max": 2550.0
965
+ },
966
+ {
967
+ "current": 2550.0,
968
+ "min": 1500.0,
969
+ "max": 2550.0
970
+ },
971
+ {
972
+ "current": 2550.0,
973
+ "min": 1500.0,
974
+ "max": 2550.0
975
+ },
976
+ {
977
+ "current": 2550.0,
978
+ "min": 1500.0,
979
+ "max": 2550.0
980
+ },
981
+ {
982
+ "current": 2550.0,
983
+ "min": 1500.0,
984
+ "max": 2550.0
985
+ },
986
+ {
987
+ "current": 3248.565,
988
+ "min": 1500.0,
989
+ "max": 2550.0
990
+ },
991
+ {
992
+ "current": 2550.0,
993
+ "min": 1500.0,
994
+ "max": 2550.0
995
+ },
996
+ {
997
+ "current": 2550.0,
998
+ "min": 1500.0,
999
+ "max": 2550.0
1000
+ },
1001
+ {
1002
+ "current": 2550.0,
1003
+ "min": 1500.0,
1004
+ "max": 2550.0
1005
+ },
1006
+ {
1007
+ "current": 2550.0,
1008
+ "min": 1500.0,
1009
+ "max": 2550.0
1010
+ },
1011
+ {
1012
+ "current": 2550.0,
1013
+ "min": 1500.0,
1014
+ "max": 2550.0
1015
+ },
1016
+ {
1017
+ "current": 2550.0,
1018
+ "min": 1500.0,
1019
+ "max": 2550.0
1020
+ },
1021
+ {
1022
+ "current": 2550.0,
1023
+ "min": 1500.0,
1024
+ "max": 2550.0
1025
+ },
1026
+ {
1027
+ "current": 2550.0,
1028
+ "min": 1500.0,
1029
+ "max": 2550.0
1030
+ },
1031
+ {
1032
+ "current": 2550.0,
1033
+ "min": 1500.0,
1034
+ "max": 2550.0
1035
+ },
1036
+ {
1037
+ "current": 2550.0,
1038
+ "min": 1500.0,
1039
+ "max": 2550.0
1040
+ },
1041
+ {
1042
+ "current": 2550.0,
1043
+ "min": 1500.0,
1044
+ "max": 2550.0
1045
+ },
1046
+ {
1047
+ "current": 2550.0,
1048
+ "min": 1500.0,
1049
+ "max": 2550.0
1050
+ },
1051
+ {
1052
+ "current": 2550.0,
1053
+ "min": 1500.0,
1054
+ "max": 2550.0
1055
+ },
1056
+ {
1057
+ "current": 2550.0,
1058
+ "min": 1500.0,
1059
+ "max": 2550.0
1060
+ },
1061
+ {
1062
+ "current": 2550.0,
1063
+ "min": 1500.0,
1064
+ "max": 2550.0
1065
+ },
1066
+ {
1067
+ "current": 2550.0,
1068
+ "min": 1500.0,
1069
+ "max": 2550.0
1070
+ },
1071
+ {
1072
+ "current": 2550.0,
1073
+ "min": 1500.0,
1074
+ "max": 2550.0
1075
+ },
1076
+ {
1077
+ "current": 2550.0,
1078
+ "min": 1500.0,
1079
+ "max": 2550.0
1080
+ },
1081
+ {
1082
+ "current": 2550.0,
1083
+ "min": 1500.0,
1084
+ "max": 2550.0
1085
+ },
1086
+ {
1087
+ "current": 2550.0,
1088
+ "min": 1500.0,
1089
+ "max": 2550.0
1090
+ },
1091
+ {
1092
+ "current": 2550.0,
1093
+ "min": 1500.0,
1094
+ "max": 2550.0
1095
+ },
1096
+ {
1097
+ "current": 2550.0,
1098
+ "min": 1500.0,
1099
+ "max": 2550.0
1100
+ },
1101
+ {
1102
+ "current": 2550.0,
1103
+ "min": 1500.0,
1104
+ "max": 2550.0
1105
+ },
1106
+ {
1107
+ "current": 2550.0,
1108
+ "min": 1500.0,
1109
+ "max": 2550.0
1110
+ },
1111
+ {
1112
+ "current": 2550.0,
1113
+ "min": 1500.0,
1114
+ "max": 2550.0
1115
+ },
1116
+ {
1117
+ "current": 2550.0,
1118
+ "min": 1500.0,
1119
+ "max": 2550.0
1120
+ },
1121
+ {
1122
+ "current": 2550.0,
1123
+ "min": 1500.0,
1124
+ "max": 2550.0
1125
+ },
1126
+ {
1127
+ "current": 2550.0,
1128
+ "min": 1500.0,
1129
+ "max": 2550.0
1130
+ },
1131
+ {
1132
+ "current": 2550.0,
1133
+ "min": 1500.0,
1134
+ "max": 2550.0
1135
+ },
1136
+ {
1137
+ "current": 2550.0,
1138
+ "min": 1500.0,
1139
+ "max": 2550.0
1140
+ },
1141
+ {
1142
+ "current": 2550.0,
1143
+ "min": 1500.0,
1144
+ "max": 2550.0
1145
+ },
1146
+ {
1147
+ "current": 2550.0,
1148
+ "min": 1500.0,
1149
+ "max": 2550.0
1150
+ },
1151
+ {
1152
+ "current": 2550.0,
1153
+ "min": 1500.0,
1154
+ "max": 2550.0
1155
+ },
1156
+ {
1157
+ "current": 2550.0,
1158
+ "min": 1500.0,
1159
+ "max": 2550.0
1160
+ },
1161
+ {
1162
+ "current": 2550.0,
1163
+ "min": 1500.0,
1164
+ "max": 2550.0
1165
+ },
1166
+ {
1167
+ "current": 2550.0,
1168
+ "min": 1500.0,
1169
+ "max": 2550.0
1170
+ },
1171
+ {
1172
+ "current": 2550.0,
1173
+ "min": 1500.0,
1174
+ "max": 2550.0
1175
+ },
1176
+ {
1177
+ "current": 2550.0,
1178
+ "min": 1500.0,
1179
+ "max": 2550.0
1180
+ },
1181
+ {
1182
+ "current": 2550.0,
1183
+ "min": 1500.0,
1184
+ "max": 2550.0
1185
+ },
1186
+ {
1187
+ "current": 2550.0,
1188
+ "min": 1500.0,
1189
+ "max": 2550.0
1190
+ },
1191
+ {
1192
+ "current": 2550.0,
1193
+ "min": 1500.0,
1194
+ "max": 2550.0
1195
+ },
1196
+ {
1197
+ "current": 2550.0,
1198
+ "min": 1500.0,
1199
+ "max": 2550.0
1200
+ },
1201
+ {
1202
+ "current": 2550.0,
1203
+ "min": 1500.0,
1204
+ "max": 2550.0
1205
+ },
1206
+ {
1207
+ "current": 2550.0,
1208
+ "min": 1500.0,
1209
+ "max": 2550.0
1210
+ },
1211
+ {
1212
+ "current": 2550.0,
1213
+ "min": 1500.0,
1214
+ "max": 2550.0
1215
+ },
1216
+ {
1217
+ "current": 2550.0,
1218
+ "min": 1500.0,
1219
+ "max": 2550.0
1220
+ },
1221
+ {
1222
+ "current": 2550.0,
1223
+ "min": 1500.0,
1224
+ "max": 2550.0
1225
+ },
1226
+ {
1227
+ "current": 2550.0,
1228
+ "min": 1500.0,
1229
+ "max": 2550.0
1230
+ },
1231
+ {
1232
+ "current": 2550.0,
1233
+ "min": 1500.0,
1234
+ "max": 2550.0
1235
+ },
1236
+ {
1237
+ "current": 2550.0,
1238
+ "min": 1500.0,
1239
+ "max": 2550.0
1240
+ },
1241
+ {
1242
+ "current": 2550.0,
1243
+ "min": 1500.0,
1244
+ "max": 2550.0
1245
+ },
1246
+ {
1247
+ "current": 2550.0,
1248
+ "min": 1500.0,
1249
+ "max": 2550.0
1250
+ },
1251
+ {
1252
+ "current": 2550.0,
1253
+ "min": 1500.0,
1254
+ "max": 2550.0
1255
+ },
1256
+ {
1257
+ "current": 2550.0,
1258
+ "min": 1500.0,
1259
+ "max": 2550.0
1260
+ },
1261
+ {
1262
+ "current": 2550.0,
1263
+ "min": 1500.0,
1264
+ "max": 2550.0
1265
+ },
1266
+ {
1267
+ "current": 2550.0,
1268
+ "min": 1500.0,
1269
+ "max": 2550.0
1270
+ },
1271
+ {
1272
+ "current": 2550.0,
1273
+ "min": 1500.0,
1274
+ "max": 2550.0
1275
+ },
1276
+ {
1277
+ "current": 2550.0,
1278
+ "min": 1500.0,
1279
+ "max": 2550.0
1280
+ },
1281
+ {
1282
+ "current": 2550.0,
1283
+ "min": 1500.0,
1284
+ "max": 2550.0
1285
+ },
1286
+ {
1287
+ "current": 2550.0,
1288
+ "min": 1500.0,
1289
+ "max": 2550.0
1290
+ },
1291
+ {
1292
+ "current": 2550.0,
1293
+ "min": 1500.0,
1294
+ "max": 2550.0
1295
+ },
1296
+ {
1297
+ "current": 2550.0,
1298
+ "min": 1500.0,
1299
+ "max": 2550.0
1300
+ },
1301
+ {
1302
+ "current": 2550.0,
1303
+ "min": 1500.0,
1304
+ "max": 2550.0
1305
+ },
1306
+ {
1307
+ "current": 2550.0,
1308
+ "min": 1500.0,
1309
+ "max": 2550.0
1310
+ },
1311
+ {
1312
+ "current": 2550.0,
1313
+ "min": 1500.0,
1314
+ "max": 2550.0
1315
+ }
1316
+ ],
1317
+ "disk": {
1318
+ "/": {
1319
+ "total": 242.1086883544922,
1320
+ "used": 50.53842544555664
1321
+ }
1322
+ },
1323
+ "gpu": "NVIDIA A100-SXM4-80GB",
1324
+ "gpu_count": 8,
1325
+ "gpu_devices": [
1326
+ {
1327
+ "name": "NVIDIA A100-SXM4-80GB",
1328
+ "memory_total": 85899345920
1329
+ },
1330
+ {
1331
+ "name": "NVIDIA A100-SXM4-80GB",
1332
+ "memory_total": 85899345920
1333
+ },
1334
+ {
1335
+ "name": "NVIDIA A100-SXM4-80GB",
1336
+ "memory_total": 85899345920
1337
+ },
1338
+ {
1339
+ "name": "NVIDIA A100-SXM4-80GB",
1340
+ "memory_total": 85899345920
1341
+ },
1342
+ {
1343
+ "name": "NVIDIA A100-SXM4-80GB",
1344
+ "memory_total": 85899345920
1345
+ },
1346
+ {
1347
+ "name": "NVIDIA A100-SXM4-80GB",
1348
+ "memory_total": 85899345920
1349
+ },
1350
+ {
1351
+ "name": "NVIDIA A100-SXM4-80GB",
1352
+ "memory_total": 85899345920
1353
+ },
1354
+ {
1355
+ "name": "NVIDIA A100-SXM4-80GB",
1356
+ "memory_total": 85899345920
1357
+ }
1358
+ ],
1359
+ "memory": {
1360
+ "total": 2015.6942749023438
1361
+ }
1362
+ }