kushal-tri commited on
Commit
5a60e84
1 Parent(s): 2fe0f98

Training in progress, epoch 0

Browse files
Files changed (27) hide show
  1. .gitattributes +1 -0
  2. added_tokens.json +30 -0
  3. config.json +29 -0
  4. merges.txt +0 -0
  5. model-00001-of-00004.safetensors +3 -0
  6. model-00002-of-00004.safetensors +3 -0
  7. model-00003-of-00004.safetensors +3 -0
  8. model-00004-of-00004.safetensors +3 -0
  9. model.safetensors.index.json +346 -0
  10. special_tokens_map.json +60 -0
  11. tokenizer.json +3 -0
  12. tokenizer_config.json +248 -0
  13. training_args.bin +3 -0
  14. vocab.json +0 -0
  15. wandb/debug-internal.log +373 -0
  16. wandb/debug.log +30 -0
  17. wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/code/qlearning_reasoning/training/sft.py +249 -0
  18. wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/conda-environment.yaml +515 -0
  19. wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/config.yaml +948 -0
  20. wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/diff.patch +49 -0
  21. wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log +0 -0
  22. wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/requirements.txt +289 -0
  23. wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/wandb-metadata.json +1069 -0
  24. wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/wandb-summary.json +1 -0
  25. wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/logs/debug-internal.log +373 -0
  26. wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/logs/debug.log +30 -0
  27. wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/run-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1.wandb +0 -0
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
added_tokens.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "</tool_call>": 151658,
3
+ "<tool_call>": 151657,
4
+ "<|box_end|>": 151649,
5
+ "<|box_start|>": 151648,
6
+ "<|endoftext|>": 151643,
7
+ "<|file_sep|>": 151664,
8
+ "<|fim_middle|>": 151660,
9
+ "<|fim_pad|>": 151662,
10
+ "<|fim_prefix|>": 151659,
11
+ "<|fim_suffix|>": 151661,
12
+ "<|im_end|>": 151645,
13
+ "<|im_start|>": 151644,
14
+ "<|image_pad|>": 151655,
15
+ "<|object_ref_end|>": 151647,
16
+ "<|object_ref_start|>": 151646,
17
+ "<|quad_end|>": 151651,
18
+ "<|quad_start|>": 151650,
19
+ "<|repo_name|>": 151663,
20
+ "<|video_pad|>": 151656,
21
+ "<|vision_end|>": 151653,
22
+ "<|vision_pad|>": 151654,
23
+ "<|vision_start|>": 151652,
24
+ "[/REVISION]": 151670,
25
+ "[/STEP]": 151666,
26
+ "[/TURN]": 151668,
27
+ "[REVISION]": 151669,
28
+ "[STEP]": 151665,
29
+ "[TURN]": 151667
30
+ }
config.json ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "Qwen/Qwen2.5-Coder-7B-Instruct",
3
+ "architectures": [
4
+ "Qwen2ForCausalLM"
5
+ ],
6
+ "attention_dropout": 0.0,
7
+ "eos_token_id": 151645,
8
+ "hidden_act": "silu",
9
+ "hidden_size": 3584,
10
+ "initializer_range": 0.02,
11
+ "intermediate_size": 18944,
12
+ "max_position_embeddings": 32768,
13
+ "max_window_layers": 28,
14
+ "model_type": "qwen2",
15
+ "num_attention_heads": 28,
16
+ "num_hidden_layers": 28,
17
+ "num_key_value_heads": 4,
18
+ "pad_token_id": 151643,
19
+ "rms_norm_eps": 1e-06,
20
+ "rope_scaling": null,
21
+ "rope_theta": 1000000.0,
22
+ "sliding_window": null,
23
+ "tie_word_embeddings": false,
24
+ "torch_dtype": "bfloat16",
25
+ "transformers_version": "4.46.0",
26
+ "use_cache": false,
27
+ "use_sliding_window": false,
28
+ "vocab_size": 151671
29
+ }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
model-00001-of-00004.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a98b4e3194379ba7bec05ef5af0baa479391fb72c51b261749bdd1d251fb9b81
3
+ size 4874843752
model-00002-of-00004.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ab9e2130ff05c8efc6ce715675f96987ccecebb141684c969aa92ece6777278a
3
+ size 4932751008
model-00003-of-00004.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:181225151db82d7c2a4b0d6bf1b800f2aa0b9e19ebcbc3f666e96ff341cd3a0b
3
+ size 4330865200
model-00004-of-00004.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e9fd93165bb07b6dbb7f816180d4f65ad089d8c4efe19e5171a47c0c751f84b0
3
+ size 1087177856
model.safetensors.index.json ADDED
@@ -0,0 +1,346 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "metadata": {
3
+ "total_size": 15225598976
4
+ },
5
+ "weight_map": {
6
+ "lm_head.weight": "model-00004-of-00004.safetensors",
7
+ "model.embed_tokens.weight": "model-00001-of-00004.safetensors",
8
+ "model.layers.0.input_layernorm.weight": "model-00001-of-00004.safetensors",
9
+ "model.layers.0.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
10
+ "model.layers.0.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
11
+ "model.layers.0.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
12
+ "model.layers.0.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
13
+ "model.layers.0.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
14
+ "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
15
+ "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
16
+ "model.layers.0.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
17
+ "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
18
+ "model.layers.0.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
19
+ "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
20
+ "model.layers.1.input_layernorm.weight": "model-00001-of-00004.safetensors",
21
+ "model.layers.1.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
22
+ "model.layers.1.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
23
+ "model.layers.1.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
24
+ "model.layers.1.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
25
+ "model.layers.1.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
26
+ "model.layers.1.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
27
+ "model.layers.1.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
28
+ "model.layers.1.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
29
+ "model.layers.1.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
30
+ "model.layers.1.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
31
+ "model.layers.1.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
32
+ "model.layers.10.input_layernorm.weight": "model-00002-of-00004.safetensors",
33
+ "model.layers.10.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
34
+ "model.layers.10.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
35
+ "model.layers.10.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
36
+ "model.layers.10.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
37
+ "model.layers.10.self_attn.k_proj.bias": "model-00002-of-00004.safetensors",
38
+ "model.layers.10.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
39
+ "model.layers.10.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
40
+ "model.layers.10.self_attn.q_proj.bias": "model-00002-of-00004.safetensors",
41
+ "model.layers.10.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
42
+ "model.layers.10.self_attn.v_proj.bias": "model-00002-of-00004.safetensors",
43
+ "model.layers.10.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
44
+ "model.layers.11.input_layernorm.weight": "model-00002-of-00004.safetensors",
45
+ "model.layers.11.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
46
+ "model.layers.11.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
47
+ "model.layers.11.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
48
+ "model.layers.11.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
49
+ "model.layers.11.self_attn.k_proj.bias": "model-00002-of-00004.safetensors",
50
+ "model.layers.11.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
51
+ "model.layers.11.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
52
+ "model.layers.11.self_attn.q_proj.bias": "model-00002-of-00004.safetensors",
53
+ "model.layers.11.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
54
+ "model.layers.11.self_attn.v_proj.bias": "model-00002-of-00004.safetensors",
55
+ "model.layers.11.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
56
+ "model.layers.12.input_layernorm.weight": "model-00002-of-00004.safetensors",
57
+ "model.layers.12.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
58
+ "model.layers.12.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
59
+ "model.layers.12.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
60
+ "model.layers.12.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
61
+ "model.layers.12.self_attn.k_proj.bias": "model-00002-of-00004.safetensors",
62
+ "model.layers.12.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
63
+ "model.layers.12.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
64
+ "model.layers.12.self_attn.q_proj.bias": "model-00002-of-00004.safetensors",
65
+ "model.layers.12.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
66
+ "model.layers.12.self_attn.v_proj.bias": "model-00002-of-00004.safetensors",
67
+ "model.layers.12.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
68
+ "model.layers.13.input_layernorm.weight": "model-00002-of-00004.safetensors",
69
+ "model.layers.13.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
70
+ "model.layers.13.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
71
+ "model.layers.13.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
72
+ "model.layers.13.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
73
+ "model.layers.13.self_attn.k_proj.bias": "model-00002-of-00004.safetensors",
74
+ "model.layers.13.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
75
+ "model.layers.13.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
76
+ "model.layers.13.self_attn.q_proj.bias": "model-00002-of-00004.safetensors",
77
+ "model.layers.13.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
78
+ "model.layers.13.self_attn.v_proj.bias": "model-00002-of-00004.safetensors",
79
+ "model.layers.13.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
80
+ "model.layers.14.input_layernorm.weight": "model-00002-of-00004.safetensors",
81
+ "model.layers.14.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
82
+ "model.layers.14.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
83
+ "model.layers.14.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
84
+ "model.layers.14.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
85
+ "model.layers.14.self_attn.k_proj.bias": "model-00002-of-00004.safetensors",
86
+ "model.layers.14.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
87
+ "model.layers.14.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
88
+ "model.layers.14.self_attn.q_proj.bias": "model-00002-of-00004.safetensors",
89
+ "model.layers.14.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
90
+ "model.layers.14.self_attn.v_proj.bias": "model-00002-of-00004.safetensors",
91
+ "model.layers.14.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
92
+ "model.layers.15.input_layernorm.weight": "model-00002-of-00004.safetensors",
93
+ "model.layers.15.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
94
+ "model.layers.15.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
95
+ "model.layers.15.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
96
+ "model.layers.15.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
97
+ "model.layers.15.self_attn.k_proj.bias": "model-00002-of-00004.safetensors",
98
+ "model.layers.15.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
99
+ "model.layers.15.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
100
+ "model.layers.15.self_attn.q_proj.bias": "model-00002-of-00004.safetensors",
101
+ "model.layers.15.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
102
+ "model.layers.15.self_attn.v_proj.bias": "model-00002-of-00004.safetensors",
103
+ "model.layers.15.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
104
+ "model.layers.16.input_layernorm.weight": "model-00002-of-00004.safetensors",
105
+ "model.layers.16.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
106
+ "model.layers.16.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
107
+ "model.layers.16.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
108
+ "model.layers.16.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
109
+ "model.layers.16.self_attn.k_proj.bias": "model-00002-of-00004.safetensors",
110
+ "model.layers.16.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
111
+ "model.layers.16.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
112
+ "model.layers.16.self_attn.q_proj.bias": "model-00002-of-00004.safetensors",
113
+ "model.layers.16.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
114
+ "model.layers.16.self_attn.v_proj.bias": "model-00002-of-00004.safetensors",
115
+ "model.layers.16.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
116
+ "model.layers.17.input_layernorm.weight": "model-00002-of-00004.safetensors",
117
+ "model.layers.17.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
118
+ "model.layers.17.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
119
+ "model.layers.17.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
120
+ "model.layers.17.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
121
+ "model.layers.17.self_attn.k_proj.bias": "model-00002-of-00004.safetensors",
122
+ "model.layers.17.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
123
+ "model.layers.17.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
124
+ "model.layers.17.self_attn.q_proj.bias": "model-00002-of-00004.safetensors",
125
+ "model.layers.17.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
126
+ "model.layers.17.self_attn.v_proj.bias": "model-00002-of-00004.safetensors",
127
+ "model.layers.17.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
128
+ "model.layers.18.input_layernorm.weight": "model-00003-of-00004.safetensors",
129
+ "model.layers.18.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
130
+ "model.layers.18.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
131
+ "model.layers.18.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
132
+ "model.layers.18.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
133
+ "model.layers.18.self_attn.k_proj.bias": "model-00002-of-00004.safetensors",
134
+ "model.layers.18.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
135
+ "model.layers.18.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
136
+ "model.layers.18.self_attn.q_proj.bias": "model-00002-of-00004.safetensors",
137
+ "model.layers.18.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
138
+ "model.layers.18.self_attn.v_proj.bias": "model-00002-of-00004.safetensors",
139
+ "model.layers.18.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
140
+ "model.layers.19.input_layernorm.weight": "model-00003-of-00004.safetensors",
141
+ "model.layers.19.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
142
+ "model.layers.19.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
143
+ "model.layers.19.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
144
+ "model.layers.19.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
145
+ "model.layers.19.self_attn.k_proj.bias": "model-00003-of-00004.safetensors",
146
+ "model.layers.19.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
147
+ "model.layers.19.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
148
+ "model.layers.19.self_attn.q_proj.bias": "model-00003-of-00004.safetensors",
149
+ "model.layers.19.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
150
+ "model.layers.19.self_attn.v_proj.bias": "model-00003-of-00004.safetensors",
151
+ "model.layers.19.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
152
+ "model.layers.2.input_layernorm.weight": "model-00001-of-00004.safetensors",
153
+ "model.layers.2.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
154
+ "model.layers.2.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
155
+ "model.layers.2.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
156
+ "model.layers.2.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
157
+ "model.layers.2.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
158
+ "model.layers.2.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
159
+ "model.layers.2.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
160
+ "model.layers.2.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
161
+ "model.layers.2.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
162
+ "model.layers.2.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
163
+ "model.layers.2.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
164
+ "model.layers.20.input_layernorm.weight": "model-00003-of-00004.safetensors",
165
+ "model.layers.20.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
166
+ "model.layers.20.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
167
+ "model.layers.20.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
168
+ "model.layers.20.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
169
+ "model.layers.20.self_attn.k_proj.bias": "model-00003-of-00004.safetensors",
170
+ "model.layers.20.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
171
+ "model.layers.20.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
172
+ "model.layers.20.self_attn.q_proj.bias": "model-00003-of-00004.safetensors",
173
+ "model.layers.20.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
174
+ "model.layers.20.self_attn.v_proj.bias": "model-00003-of-00004.safetensors",
175
+ "model.layers.20.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
176
+ "model.layers.21.input_layernorm.weight": "model-00003-of-00004.safetensors",
177
+ "model.layers.21.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
178
+ "model.layers.21.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
179
+ "model.layers.21.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
180
+ "model.layers.21.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
181
+ "model.layers.21.self_attn.k_proj.bias": "model-00003-of-00004.safetensors",
182
+ "model.layers.21.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
183
+ "model.layers.21.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
184
+ "model.layers.21.self_attn.q_proj.bias": "model-00003-of-00004.safetensors",
185
+ "model.layers.21.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
186
+ "model.layers.21.self_attn.v_proj.bias": "model-00003-of-00004.safetensors",
187
+ "model.layers.21.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
188
+ "model.layers.22.input_layernorm.weight": "model-00003-of-00004.safetensors",
189
+ "model.layers.22.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
190
+ "model.layers.22.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
191
+ "model.layers.22.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
192
+ "model.layers.22.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
193
+ "model.layers.22.self_attn.k_proj.bias": "model-00003-of-00004.safetensors",
194
+ "model.layers.22.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
195
+ "model.layers.22.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
196
+ "model.layers.22.self_attn.q_proj.bias": "model-00003-of-00004.safetensors",
197
+ "model.layers.22.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
198
+ "model.layers.22.self_attn.v_proj.bias": "model-00003-of-00004.safetensors",
199
+ "model.layers.22.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
200
+ "model.layers.23.input_layernorm.weight": "model-00003-of-00004.safetensors",
201
+ "model.layers.23.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
202
+ "model.layers.23.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
203
+ "model.layers.23.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
204
+ "model.layers.23.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
205
+ "model.layers.23.self_attn.k_proj.bias": "model-00003-of-00004.safetensors",
206
+ "model.layers.23.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
207
+ "model.layers.23.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
208
+ "model.layers.23.self_attn.q_proj.bias": "model-00003-of-00004.safetensors",
209
+ "model.layers.23.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
210
+ "model.layers.23.self_attn.v_proj.bias": "model-00003-of-00004.safetensors",
211
+ "model.layers.23.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
212
+ "model.layers.24.input_layernorm.weight": "model-00003-of-00004.safetensors",
213
+ "model.layers.24.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
214
+ "model.layers.24.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
215
+ "model.layers.24.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
216
+ "model.layers.24.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
217
+ "model.layers.24.self_attn.k_proj.bias": "model-00003-of-00004.safetensors",
218
+ "model.layers.24.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
219
+ "model.layers.24.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
220
+ "model.layers.24.self_attn.q_proj.bias": "model-00003-of-00004.safetensors",
221
+ "model.layers.24.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
222
+ "model.layers.24.self_attn.v_proj.bias": "model-00003-of-00004.safetensors",
223
+ "model.layers.24.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
224
+ "model.layers.25.input_layernorm.weight": "model-00003-of-00004.safetensors",
225
+ "model.layers.25.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
226
+ "model.layers.25.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
227
+ "model.layers.25.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
228
+ "model.layers.25.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
229
+ "model.layers.25.self_attn.k_proj.bias": "model-00003-of-00004.safetensors",
230
+ "model.layers.25.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
231
+ "model.layers.25.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
232
+ "model.layers.25.self_attn.q_proj.bias": "model-00003-of-00004.safetensors",
233
+ "model.layers.25.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
234
+ "model.layers.25.self_attn.v_proj.bias": "model-00003-of-00004.safetensors",
235
+ "model.layers.25.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
236
+ "model.layers.26.input_layernorm.weight": "model-00003-of-00004.safetensors",
237
+ "model.layers.26.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
238
+ "model.layers.26.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
239
+ "model.layers.26.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
240
+ "model.layers.26.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
241
+ "model.layers.26.self_attn.k_proj.bias": "model-00003-of-00004.safetensors",
242
+ "model.layers.26.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
243
+ "model.layers.26.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
244
+ "model.layers.26.self_attn.q_proj.bias": "model-00003-of-00004.safetensors",
245
+ "model.layers.26.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
246
+ "model.layers.26.self_attn.v_proj.bias": "model-00003-of-00004.safetensors",
247
+ "model.layers.26.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
248
+ "model.layers.27.input_layernorm.weight": "model-00003-of-00004.safetensors",
249
+ "model.layers.27.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
250
+ "model.layers.27.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
251
+ "model.layers.27.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
252
+ "model.layers.27.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
253
+ "model.layers.27.self_attn.k_proj.bias": "model-00003-of-00004.safetensors",
254
+ "model.layers.27.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
255
+ "model.layers.27.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
256
+ "model.layers.27.self_attn.q_proj.bias": "model-00003-of-00004.safetensors",
257
+ "model.layers.27.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
258
+ "model.layers.27.self_attn.v_proj.bias": "model-00003-of-00004.safetensors",
259
+ "model.layers.27.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
260
+ "model.layers.3.input_layernorm.weight": "model-00001-of-00004.safetensors",
261
+ "model.layers.3.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
262
+ "model.layers.3.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
263
+ "model.layers.3.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
264
+ "model.layers.3.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
265
+ "model.layers.3.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
266
+ "model.layers.3.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
267
+ "model.layers.3.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
268
+ "model.layers.3.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
269
+ "model.layers.3.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
270
+ "model.layers.3.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
271
+ "model.layers.3.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
272
+ "model.layers.4.input_layernorm.weight": "model-00001-of-00004.safetensors",
273
+ "model.layers.4.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
274
+ "model.layers.4.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
275
+ "model.layers.4.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
276
+ "model.layers.4.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
277
+ "model.layers.4.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
278
+ "model.layers.4.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
279
+ "model.layers.4.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
280
+ "model.layers.4.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
281
+ "model.layers.4.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
282
+ "model.layers.4.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
283
+ "model.layers.4.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
284
+ "model.layers.5.input_layernorm.weight": "model-00001-of-00004.safetensors",
285
+ "model.layers.5.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
286
+ "model.layers.5.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
287
+ "model.layers.5.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
288
+ "model.layers.5.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
289
+ "model.layers.5.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
290
+ "model.layers.5.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
291
+ "model.layers.5.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
292
+ "model.layers.5.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
293
+ "model.layers.5.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
294
+ "model.layers.5.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
295
+ "model.layers.5.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
296
+ "model.layers.6.input_layernorm.weight": "model-00001-of-00004.safetensors",
297
+ "model.layers.6.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
298
+ "model.layers.6.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
299
+ "model.layers.6.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
300
+ "model.layers.6.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
301
+ "model.layers.6.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
302
+ "model.layers.6.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
303
+ "model.layers.6.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
304
+ "model.layers.6.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
305
+ "model.layers.6.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
306
+ "model.layers.6.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
307
+ "model.layers.6.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
308
+ "model.layers.7.input_layernorm.weight": "model-00001-of-00004.safetensors",
309
+ "model.layers.7.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
310
+ "model.layers.7.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
311
+ "model.layers.7.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
312
+ "model.layers.7.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
313
+ "model.layers.7.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
314
+ "model.layers.7.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
315
+ "model.layers.7.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
316
+ "model.layers.7.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
317
+ "model.layers.7.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
318
+ "model.layers.7.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
319
+ "model.layers.7.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
320
+ "model.layers.8.input_layernorm.weight": "model-00002-of-00004.safetensors",
321
+ "model.layers.8.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
322
+ "model.layers.8.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
323
+ "model.layers.8.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
324
+ "model.layers.8.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
325
+ "model.layers.8.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
326
+ "model.layers.8.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
327
+ "model.layers.8.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
328
+ "model.layers.8.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
329
+ "model.layers.8.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
330
+ "model.layers.8.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
331
+ "model.layers.8.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
332
+ "model.layers.9.input_layernorm.weight": "model-00002-of-00004.safetensors",
333
+ "model.layers.9.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
334
+ "model.layers.9.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
335
+ "model.layers.9.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
336
+ "model.layers.9.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
337
+ "model.layers.9.self_attn.k_proj.bias": "model-00002-of-00004.safetensors",
338
+ "model.layers.9.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
339
+ "model.layers.9.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
340
+ "model.layers.9.self_attn.q_proj.bias": "model-00002-of-00004.safetensors",
341
+ "model.layers.9.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
342
+ "model.layers.9.self_attn.v_proj.bias": "model-00002-of-00004.safetensors",
343
+ "model.layers.9.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
344
+ "model.norm.weight": "model-00003-of-00004.safetensors"
345
+ }
346
+ }
special_tokens_map.json ADDED
@@ -0,0 +1,60 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ {
4
+ "content": "[STEP]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false
9
+ },
10
+ {
11
+ "content": "[/STEP]",
12
+ "lstrip": false,
13
+ "normalized": false,
14
+ "rstrip": false,
15
+ "single_word": false
16
+ },
17
+ {
18
+ "content": "[TURN]",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ },
24
+ {
25
+ "content": "[/TURN]",
26
+ "lstrip": false,
27
+ "normalized": false,
28
+ "rstrip": false,
29
+ "single_word": false
30
+ },
31
+ {
32
+ "content": "[REVISION]",
33
+ "lstrip": false,
34
+ "normalized": false,
35
+ "rstrip": false,
36
+ "single_word": false
37
+ },
38
+ {
39
+ "content": "[/REVISION]",
40
+ "lstrip": false,
41
+ "normalized": false,
42
+ "rstrip": false,
43
+ "single_word": false
44
+ }
45
+ ],
46
+ "eos_token": {
47
+ "content": "<|im_end|>",
48
+ "lstrip": false,
49
+ "normalized": false,
50
+ "rstrip": false,
51
+ "single_word": false
52
+ },
53
+ "pad_token": {
54
+ "content": "<|endoftext|>",
55
+ "lstrip": false,
56
+ "normalized": false,
57
+ "rstrip": false,
58
+ "single_word": false
59
+ }
60
+ }
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:119474abe8593f47402f00c90e22eac0436e64f8c84c8270309a27e152901418
3
+ size 11423104
tokenizer_config.json ADDED
@@ -0,0 +1,248 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": false,
3
+ "add_prefix_space": false,
4
+ "added_tokens_decoder": {
5
+ "151643": {
6
+ "content": "<|endoftext|>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "151644": {
14
+ "content": "<|im_start|>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "151645": {
22
+ "content": "<|im_end|>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ },
29
+ "151646": {
30
+ "content": "<|object_ref_start|>",
31
+ "lstrip": false,
32
+ "normalized": false,
33
+ "rstrip": false,
34
+ "single_word": false,
35
+ "special": true
36
+ },
37
+ "151647": {
38
+ "content": "<|object_ref_end|>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false,
43
+ "special": true
44
+ },
45
+ "151648": {
46
+ "content": "<|box_start|>",
47
+ "lstrip": false,
48
+ "normalized": false,
49
+ "rstrip": false,
50
+ "single_word": false,
51
+ "special": true
52
+ },
53
+ "151649": {
54
+ "content": "<|box_end|>",
55
+ "lstrip": false,
56
+ "normalized": false,
57
+ "rstrip": false,
58
+ "single_word": false,
59
+ "special": true
60
+ },
61
+ "151650": {
62
+ "content": "<|quad_start|>",
63
+ "lstrip": false,
64
+ "normalized": false,
65
+ "rstrip": false,
66
+ "single_word": false,
67
+ "special": true
68
+ },
69
+ "151651": {
70
+ "content": "<|quad_end|>",
71
+ "lstrip": false,
72
+ "normalized": false,
73
+ "rstrip": false,
74
+ "single_word": false,
75
+ "special": true
76
+ },
77
+ "151652": {
78
+ "content": "<|vision_start|>",
79
+ "lstrip": false,
80
+ "normalized": false,
81
+ "rstrip": false,
82
+ "single_word": false,
83
+ "special": true
84
+ },
85
+ "151653": {
86
+ "content": "<|vision_end|>",
87
+ "lstrip": false,
88
+ "normalized": false,
89
+ "rstrip": false,
90
+ "single_word": false,
91
+ "special": true
92
+ },
93
+ "151654": {
94
+ "content": "<|vision_pad|>",
95
+ "lstrip": false,
96
+ "normalized": false,
97
+ "rstrip": false,
98
+ "single_word": false,
99
+ "special": true
100
+ },
101
+ "151655": {
102
+ "content": "<|image_pad|>",
103
+ "lstrip": false,
104
+ "normalized": false,
105
+ "rstrip": false,
106
+ "single_word": false,
107
+ "special": true
108
+ },
109
+ "151656": {
110
+ "content": "<|video_pad|>",
111
+ "lstrip": false,
112
+ "normalized": false,
113
+ "rstrip": false,
114
+ "single_word": false,
115
+ "special": true
116
+ },
117
+ "151657": {
118
+ "content": "<tool_call>",
119
+ "lstrip": false,
120
+ "normalized": false,
121
+ "rstrip": false,
122
+ "single_word": false,
123
+ "special": false
124
+ },
125
+ "151658": {
126
+ "content": "</tool_call>",
127
+ "lstrip": false,
128
+ "normalized": false,
129
+ "rstrip": false,
130
+ "single_word": false,
131
+ "special": false
132
+ },
133
+ "151659": {
134
+ "content": "<|fim_prefix|>",
135
+ "lstrip": false,
136
+ "normalized": false,
137
+ "rstrip": false,
138
+ "single_word": false,
139
+ "special": false
140
+ },
141
+ "151660": {
142
+ "content": "<|fim_middle|>",
143
+ "lstrip": false,
144
+ "normalized": false,
145
+ "rstrip": false,
146
+ "single_word": false,
147
+ "special": false
148
+ },
149
+ "151661": {
150
+ "content": "<|fim_suffix|>",
151
+ "lstrip": false,
152
+ "normalized": false,
153
+ "rstrip": false,
154
+ "single_word": false,
155
+ "special": false
156
+ },
157
+ "151662": {
158
+ "content": "<|fim_pad|>",
159
+ "lstrip": false,
160
+ "normalized": false,
161
+ "rstrip": false,
162
+ "single_word": false,
163
+ "special": false
164
+ },
165
+ "151663": {
166
+ "content": "<|repo_name|>",
167
+ "lstrip": false,
168
+ "normalized": false,
169
+ "rstrip": false,
170
+ "single_word": false,
171
+ "special": false
172
+ },
173
+ "151664": {
174
+ "content": "<|file_sep|>",
175
+ "lstrip": false,
176
+ "normalized": false,
177
+ "rstrip": false,
178
+ "single_word": false,
179
+ "special": false
180
+ },
181
+ "151665": {
182
+ "content": "[STEP]",
183
+ "lstrip": false,
184
+ "normalized": false,
185
+ "rstrip": false,
186
+ "single_word": false,
187
+ "special": true
188
+ },
189
+ "151666": {
190
+ "content": "[/STEP]",
191
+ "lstrip": false,
192
+ "normalized": false,
193
+ "rstrip": false,
194
+ "single_word": false,
195
+ "special": true
196
+ },
197
+ "151667": {
198
+ "content": "[TURN]",
199
+ "lstrip": false,
200
+ "normalized": false,
201
+ "rstrip": false,
202
+ "single_word": false,
203
+ "special": true
204
+ },
205
+ "151668": {
206
+ "content": "[/TURN]",
207
+ "lstrip": false,
208
+ "normalized": false,
209
+ "rstrip": false,
210
+ "single_word": false,
211
+ "special": true
212
+ },
213
+ "151669": {
214
+ "content": "[REVISION]",
215
+ "lstrip": false,
216
+ "normalized": false,
217
+ "rstrip": false,
218
+ "single_word": false,
219
+ "special": true
220
+ },
221
+ "151670": {
222
+ "content": "[/REVISION]",
223
+ "lstrip": false,
224
+ "normalized": false,
225
+ "rstrip": false,
226
+ "single_word": false,
227
+ "special": true
228
+ }
229
+ },
230
+ "additional_special_tokens": [
231
+ "[STEP]",
232
+ "[/STEP]",
233
+ "[TURN]",
234
+ "[/TURN]",
235
+ "[REVISION]",
236
+ "[/REVISION]"
237
+ ],
238
+ "bos_token": null,
239
+ "chat_template": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0]['role'] == 'system' %}\n {{- messages[0]['content'] }}\n {%- else %}\n {{- 'You are Qwen, created by Alibaba Cloud. You are a helpful assistant.' }}\n {%- endif %}\n {{- \"\\n\\n# Tools\\n\\nYou may call one or more functions to assist with the user query.\\n\\nYou are provided with function signatures within <tools></tools> XML tags:\\n<tools>\" }}\n {%- for tool in tools %}\n {{- \"\\n\" }}\n {{- tool | tojson }}\n {%- endfor %}\n {{- \"\\n</tools>\\n\\nFor each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:\\n<tool_call>\\n{\\\"name\\\": <function-name>, \\\"arguments\\\": <args-json-object>}\\n</tool_call><|im_end|>\\n\" }}\n{%- else %}\n {%- if messages[0]['role'] == 'system' %}\n {{- '<|im_start|>system\\n' + messages[0]['content'] + '<|im_end|>\\n' }}\n {%- else %}\n {{- '<|im_start|>system\\nYou are Qwen, created by Alibaba Cloud. You are a helpful assistant.<|im_end|>\\n' }}\n {%- endif %}\n{%- endif %}\n{%- for message in messages %}\n {%- if (message.role == \"user\") or (message.role == \"system\" and not loop.first) or (message.role == \"assistant\" and not message.tool_calls) %}\n {{- '<|im_start|>' + message.role + '\\n' + message.content + '<|im_end|>' + '\\n' }}\n {%- elif message.role == \"assistant\" %}\n {{- '<|im_start|>' + message.role }}\n {%- if message.content %}\n {{- '\\n' + message.content }}\n {%- endif %}\n {%- for tool_call in message.tool_calls %}\n {%- if tool_call.function is defined %}\n {%- set tool_call = tool_call.function %}\n {%- endif %}\n {{- '\\n<tool_call>\\n{\"name\": \"' }}\n {{- tool_call.name }}\n {{- '\", \"arguments\": ' }}\n {{- tool_call.arguments | tojson }}\n {{- '}\\n</tool_call>' }}\n {%- endfor %}\n {{- '<|im_end|>\\n' }}\n {%- elif message.role == \"tool\" %}\n {%- if (loop.index0 == 0) or (messages[loop.index0 - 1].role != \"tool\") %}\n {{- '<|im_start|>user' }}\n {%- endif %}\n {{- '\\n<tool_response>\\n' }}\n {{- message.content }}\n {{- '\\n</tool_response>' }}\n {%- if loop.last or (messages[loop.index0 + 1].role != \"tool\") %}\n {{- '<|im_end|>\\n' }}\n {%- endif %}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|im_start|>assistant\\n' }}\n{%- endif %}\n",
240
+ "clean_up_tokenization_spaces": false,
241
+ "eos_token": "<|im_end|>",
242
+ "errors": "replace",
243
+ "model_max_length": 131072,
244
+ "pad_token": "<|endoftext|>",
245
+ "split_special_tokens": false,
246
+ "tokenizer_class": "Qwen2Tokenizer",
247
+ "unk_token": null
248
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5d8b58b2fd3affe635bf267c95f88ed337f48f927e28276811b6aa45b9a3dd53
3
+ size 6328
vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
wandb/debug-internal.log ADDED
@@ -0,0 +1,373 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-11-13 19:02:46,431 INFO StreamThr :1939 [internal.py:wandb_internal():86] W&B internal server running at pid: 1939, started at: 2024-11-13 19:02:46.431351
2
+ 2024-11-13 19:02:46,433 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status
3
+ 2024-11-13 19:02:46,434 INFO WriterThread:1939 [datastore.py:open_for_write():87] open: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/run-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1.wandb
4
+ 2024-11-13 19:02:46,435 DEBUG SenderThread:1939 [sender.py:send():382] send: header
5
+ 2024-11-13 19:02:46,442 DEBUG SenderThread:1939 [sender.py:send():382] send: run
6
+ 2024-11-13 19:02:46,772 INFO SenderThread:1939 [dir_watcher.py:__init__():211] watching files in: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files
7
+ 2024-11-13 19:02:46,772 INFO SenderThread:1939 [sender.py:_start_run_threads():1136] run started: kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1 with start time 1731524566.431221
8
+ 2024-11-13 19:02:46,782 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: check_version
9
+ 2024-11-13 19:02:46,782 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: check_version
10
+ 2024-11-13 19:02:46,859 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: run_start
11
+ 2024-11-13 19:02:46,947 DEBUG HandlerThread:1939 [system_info.py:__init__():27] System info init
12
+ 2024-11-13 19:02:46,947 DEBUG HandlerThread:1939 [system_info.py:__init__():42] System info init done
13
+ 2024-11-13 19:02:46,947 INFO HandlerThread:1939 [system_monitor.py:start():194] Starting system monitor
14
+ 2024-11-13 19:02:46,947 INFO SystemMonitor:1939 [system_monitor.py:_start():158] Starting system asset monitoring threads
15
+ 2024-11-13 19:02:46,948 INFO HandlerThread:1939 [system_monitor.py:probe():214] Collecting system info
16
+ 2024-11-13 19:02:46,948 INFO SystemMonitor:1939 [interfaces.py:start():190] Started cpu monitoring
17
+ 2024-11-13 19:02:46,949 INFO SystemMonitor:1939 [interfaces.py:start():190] Started disk monitoring
18
+ 2024-11-13 19:02:46,949 INFO SystemMonitor:1939 [interfaces.py:start():190] Started gpu monitoring
19
+ 2024-11-13 19:02:46,951 INFO SystemMonitor:1939 [interfaces.py:start():190] Started memory monitoring
20
+ 2024-11-13 19:02:46,952 INFO SystemMonitor:1939 [interfaces.py:start():190] Started network monitoring
21
+ 2024-11-13 19:02:47,010 DEBUG HandlerThread:1939 [system_info.py:probe():151] Probing system
22
+ 2024-11-13 19:02:47,013 DEBUG HandlerThread:1939 [system_info.py:_probe_git():136] Probing git
23
+ 2024-11-13 19:02:47,021 DEBUG HandlerThread:1939 [system_info.py:_probe_git():144] Probing git done
24
+ 2024-11-13 19:02:47,021 DEBUG HandlerThread:1939 [system_info.py:probe():199] Probing system done
25
+ 2024-11-13 19:02:47,022 DEBUG HandlerThread:1939 [system_monitor.py:probe():223] {'os': 'Linux-5.10.226-214.879.amzn2.x86_64-x86_64-with-glibc2.31', 'python': '3.10.12', 'heartbeatAt': '2024-11-13T19:02:47.010671', 'startedAt': '2024-11-13T19:02:46.426230', 'docker': None, 'cuda': None, 'args': ('--batch_size', '64', '--cache_dir', '/opt/ml/data/input/.cache', '--dataset', 'code-contests', '--gradient_accumulation_steps', '8', '--logging_steps', '5', '--lr', '1e-6', '--max_seq_length', '2048', '--model_name', 'Qwen/Qwen2.5-Coder-7B-Instruct', '--num_train_epochs', '2', '--output_dir', '/opt/ml/model/', '--push_to_hub', 'True', '--run_name', 'sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048', '--save_steps', '100', '--schedule', 'cosine', '--wandb_project', 'sft-codecontests-1112', '--weight_decay', '0.0'), 'state': 'running', 'program': '/opt/ml/code/qlearning_reasoning/training/sft.py', 'codePathLocal': 'qlearning_reasoning/training/sft.py', 'codePath': 'qlearning_reasoning/training/sft.py', 'git': {'remote': 'git@github.com:TRI-ML/reasoning-value-verifiers.git', 'commit': 'cbfd05147f15c55be83f4996eebdb305c17f119d'}, 'email': None, 'root': '/opt/ml/code', 'host': 'algo-1', 'username': 'root', 'executable': '/opt/conda/bin/python3.10', 'cpu_count': 96, 'cpu_count_logical': 192, 'cpu_freq': {'current': 2829.910838541667, 'min': 0.0, 'max': 0.0}, 'cpu_freq_per_core': [{'current': 2850.788, 'min': 0.0, 'max': 0.0}, {'current': 3406.159, 'min': 0.0, 'max': 0.0}, {'current': 2605.944, 'min': 0.0, 'max': 0.0}, {'current': 2608.162, 'min': 0.0, 'max': 0.0}, {'current': 2818.169, 'min': 0.0, 'max': 0.0}, {'current': 2591.551, 'min': 0.0, 'max': 0.0}, {'current': 2621.486, 'min': 0.0, 'max': 0.0}, {'current': 2607.504, 'min': 0.0, 'max': 0.0}, {'current': 2999.734, 'min': 0.0, 'max': 0.0}, {'current': 2991.449, 'min': 0.0, 'max': 0.0}, {'current': 3123.177, 'min': 0.0, 'max': 0.0}, {'current': 3053.641, 'min': 0.0, 'max': 0.0}, {'current': 3003.014, 'min': 0.0, 'max': 0.0}, {'current': 3000.178, 'min': 0.0, 'max': 0.0}, {'current': 3432.256, 'min': 0.0, 'max': 0.0}, {'current': 2999.802, 'min': 0.0, 'max': 0.0}, {'current': 3005.725, 'min': 0.0, 'max': 0.0}, {'current': 3565.885, 'min': 0.0, 'max': 0.0}, {'current': 2992.339, 'min': 0.0, 'max': 0.0}, {'current': 3104.468, 'min': 0.0, 'max': 0.0}, {'current': 3182.411, 'min': 0.0, 'max': 0.0}, {'current': 3013.217, 'min': 0.0, 'max': 0.0}, {'current': 3131.904, 'min': 0.0, 'max': 0.0}, {'current': 3007.314, 'min': 0.0, 'max': 0.0}, {'current': 2625.175, 'min': 0.0, 'max': 0.0}, {'current': 2653.407, 'min': 0.0, 'max': 0.0}, {'current': 2610.871, 'min': 0.0, 'max': 0.0}, {'current': 2793.855, 'min': 0.0, 'max': 0.0}, {'current': 2619.232, 'min': 0.0, 'max': 0.0}, {'current': 2627.886, 'min': 0.0, 'max': 0.0}, {'current': 2613.748, 'min': 0.0, 'max': 0.0}, {'current': 2318.039, 'min': 0.0, 'max': 0.0}, {'current': 3546.661, 'min': 0.0, 'max': 0.0}, {'current': 3022.56, 'min': 0.0, 'max': 0.0}, {'current': 3035.28, 'min': 0.0, 'max': 0.0}, {'current': 3541.443, 'min': 0.0, 'max': 0.0}, {'current': 3163.55, 'min': 0.0, 'max': 0.0}, {'current': 3023.79, 'min': 0.0, 'max': 0.0}, {'current': 3033.204, 'min': 0.0, 'max': 0.0}, {'current': 3031.695, 'min': 0.0, 'max': 0.0}, {'current': 3034.587, 'min': 0.0, 'max': 0.0}, {'current': 3031.71, 'min': 0.0, 'max': 0.0}, {'current': 3029.116, 'min': 0.0, 'max': 0.0}, {'current': 3293.701, 'min': 0.0, 'max': 0.0}, {'current': 3033.78, 'min': 0.0, 'max': 0.0}, {'current': 3032.034, 'min': 0.0, 'max': 0.0}, {'current': 3057.365, 'min': 0.0, 'max': 0.0}, {'current': 3028.183, 'min': 0.0, 'max': 0.0}, {'current': 3596.439, 'min': 0.0, 'max': 0.0}, {'current': 3327.853, 'min': 0.0, 'max': 0.0}, {'current': 3226.463, 'min': 0.0, 'max': 0.0}, {'current': 3259.361, 'min': 0.0, 'max': 0.0}, {'current': 3284.076, 'min': 0.0, 'max': 0.0}, {'current': 3257.698, 'min': 0.0, 'max': 0.0}, {'current': 3280.289, 'min': 0.0, 'max': 0.0}, {'current': 3258.834, 'min': 0.0, 'max': 0.0}, {'current': 2517.346, 'min': 0.0, 'max': 0.0}, {'current': 2553.091, 'min': 0.0, 'max': 0.0}, {'current': 2552.693, 'min': 0.0, 'max': 0.0}, {'current': 2689.091, 'min': 0.0, 'max': 0.0}, {'current': 2503.301, 'min': 0.0, 'max': 0.0}, {'current': 2524.4, 'min': 0.0, 'max': 0.0}, {'current': 2521.768, 'min': 0.0, 'max': 0.0}, {'current': 2530.364, 'min': 0.0, 'max': 0.0}, {'current': 3258.225, 'min': 0.0, 'max': 0.0}, {'current': 3597.406, 'min': 0.0, 'max': 0.0}, {'current': 3279.658, 'min': 0.0, 'max': 0.0}, {'current': 3292.257, 'min': 0.0, 'max': 0.0}, {'current': 3265.351, 'min': 0.0, 'max': 0.0}, {'current': 3286.715, 'min': 0.0, 'max': 0.0}, {'current': 3268.777, 'min': 0.0, 'max': 0.0}, {'current': 3242.068, 'min': 0.0, 'max': 0.0}, {'current': 3288.908, 'min': 0.0, 'max': 0.0}, {'current': 3278.645, 'min': 0.0, 'max': 0.0}, {'current': 3215.578, 'min': 0.0, 'max': 0.0}, {'current': 3327.101, 'min': 0.0, 'max': 0.0}, {'current': 3357.075, 'min': 0.0, 'max': 0.0}, {'current': 3144.838, 'min': 0.0, 'max': 0.0}, {'current': 3005.517, 'min': 0.0, 'max': 0.0}, {'current': 3597.381, 'min': 0.0, 'max': 0.0}, {'current': 3596.175, 'min': 0.0, 'max': 0.0}, {'current': 3317.258, 'min': 0.0, 'max': 0.0}, {'current': 3310.249, 'min': 0.0, 'max': 0.0}, {'current': 3140.687, 'min': 0.0, 'max': 0.0}, {'current': 3398.954, 'min': 0.0, 'max': 0.0}, {'current': 3318.433, 'min': 0.0, 'max': 0.0}, {'current': 3342.35, 'min': 0.0, 'max': 0.0}, {'current': 3328.575, 'min': 0.0, 'max': 0.0}, {'current': 3499.892, 'min': 0.0, 'max': 0.0}, {'current': 3344.864, 'min': 0.0, 'max': 0.0}, {'current': 3355.802, 'min': 0.0, 'max': 0.0}, {'current': 3305.032, 'min': 0.0, 'max': 0.0}, {'current': 3318.07, 'min': 0.0, 'max': 0.0}, {'current': 3336.546, 'min': 0.0, 'max': 0.0}, {'current': 3335.458, 'min': 0.0, 'max': 0.0}, {'current': 3596.561, 'min': 0.0, 'max': 0.0}, {'current': 2884.473, 'min': 0.0, 'max': 0.0}, {'current': 2934.098, 'min': 0.0, 'max': 0.0}, {'current': 2621.598, 'min': 0.0, 'max': 0.0}, {'current': 2645.184, 'min': 0.0, 'max': 0.0}, {'current': 2480.726, 'min': 0.0, 'max': 0.0}, {'current': 2515.912, 'min': 0.0, 'max': 0.0}, {'current': 2702.57, 'min': 0.0, 'max': 0.0}, {'current': 2671.473, 'min': 0.0, 'max': 0.0}, {'current': 3092.321, 'min': 0.0, 'max': 0.0}, {'current': 3092.204, 'min': 0.0, 'max': 0.0}, {'current': 3217.287, 'min': 0.0, 'max': 0.0}, {'current': 3161.646, 'min': 0.0, 'max': 0.0}, {'current': 3089.273, 'min': 0.0, 'max': 0.0}, {'current': 3088.447, 'min': 0.0, 'max': 0.0}, {'current': 3559.214, 'min': 0.0, 'max': 0.0}, {'current': 3218.611, 'min': 0.0, 'max': 0.0}, {'current': 3021.622, 'min': 0.0, 'max': 0.0}, {'current': 3332.084, 'min': 0.0, 'max': 0.0}, {'current': 3019.854, 'min': 0.0, 'max': 0.0}, {'current': 3213.831, 'min': 0.0, 'max': 0.0}, {'current': 3092.774, 'min': 0.0, 'max': 0.0}, {'current': 3015.868, 'min': 0.0, 'max': 0.0}, {'current': 3082.752, 'min': 0.0, 'max': 0.0}, {'current': 3029.559, 'min': 0.0, 'max': 0.0}, {'current': 2873.142, 'min': 0.0, 'max': 0.0}, {'current': 2215.288, 'min': 0.0, 'max': 0.0}, {'current': 3193.443, 'min': 0.0, 'max': 0.0}, {'current': 2928.717, 'min': 0.0, 'max': 0.0}, {'current': 2990.366, 'min': 0.0, 'max': 0.0}, {'current': 2923.847, 'min': 0.0, 'max': 0.0}, {'current': 2869.364, 'min': 0.0, 'max': 0.0}, {'current': 2859.189, 'min': 0.0, 'max': 0.0}, {'current': 3049.001, 'min': 0.0, 'max': 0.0}, {'current': 2805.206, 'min': 0.0, 'max': 0.0}, {'current': 3094.081, 'min': 0.0, 'max': 0.0}, {'current': 3187.222, 'min': 0.0, 'max': 0.0}, {'current': 2853.945, 'min': 0.0, 'max': 0.0}, {'current': 2944.73, 'min': 0.0, 'max': 0.0}, {'current': 2809.161, 'min': 0.0, 'max': 0.0}, {'current': 2806.787, 'min': 0.0, 'max': 0.0}, {'current': 3024.225, 'min': 0.0, 'max': 0.0}, {'current': 3032.173, 'min': 0.0, 'max': 0.0}, {'current': 3033.195, 'min': 0.0, 'max': 0.0}, {'current': 3561.287, 'min': 0.0, 'max': 0.0}, {'current': 2987.872, 'min': 0.0, 'max': 0.0}, {'current': 3051.252, 'min': 0.0, 'max': 0.0}, {'current': 3379.017, 'min': 0.0, 'max': 0.0}, {'current': 3044.319, 'min': 0.0, 'max': 0.0}, {'current': 3595.62, 'min': 0.0, 'max': 0.0}, {'current': 3261.737, 'min': 0.0, 'max': 0.0}, {'current': 3257.013, 'min': 0.0, 'max': 0.0}, {'current': 3288.317, 'min': 0.0, 'max': 0.0}, {'current': 3268.675, 'min': 0.0, 'max': 0.0}, {'current': 3269.579, 'min': 0.0, 'max': 0.0}, {'current': 3266.213, 'min': 0.0, 'max': 0.0}, {'current': 3265.148, 'min': 0.0, 'max': 0.0}, {'current': 2554.152, 'min': 0.0, 'max': 0.0}, {'current': 2556.63, 'min': 0.0, 'max': 0.0}, {'current': 2540.571, 'min': 0.0, 'max': 0.0}, {'current': 2512.005, 'min': 0.0, 'max': 0.0}, {'current': 2560.992, 'min': 0.0, 'max': 0.0}, {'current': 2526.641, 'min': 0.0, 'max': 0.0}, {'current': 2528.267, 'min': 0.0, 'max': 0.0}, {'current': 2583.061, 'min': 0.0, 'max': 0.0}, {'current': 3257.652, 'min': 0.0, 'max': 0.0}, {'current': 3598.217, 'min': 0.0, 'max': 0.0}, {'current': 3282.578, 'min': 0.0, 'max': 0.0}, {'current': 3261.664, 'min': 0.0, 'max': 0.0}, {'current': 3268.246, 'min': 0.0, 'max': 0.0}, {'current': 3331.937, 'min': 0.0, 'max': 0.0}, {'current': 3265.603, 'min': 0.0, 'max': 0.0}, {'current': 3258.501, 'min': 0.0, 'max': 0.0}, {'current': 3337.045, 'min': 0.0, 'max': 0.0}, {'current': 3259.949, 'min': 0.0, 'max': 0.0}, {'current': 3226.058, 'min': 0.0, 'max': 0.0}, {'current': 3250.308, 'min': 0.0, 'max': 0.0}, {'current': 3208.026, 'min': 0.0, 'max': 0.0}, {'current': 2974.043, 'min': 0.0, 'max': 0.0}, {'current': 3216.706, 'min': 0.0, 'max': 0.0}, {'current': 3598.55, 'min': 0.0, 'max': 0.0}, {'current': 3598.261, 'min': 0.0, 'max': 0.0}, {'current': 3222.637, 'min': 0.0, 'max': 0.0}, {'current': 3227.662, 'min': 0.0, 'max': 0.0}, {'current': 3143.781, 'min': 0.0, 'max': 0.0}, {'current': 3215.567, 'min': 0.0, 'max': 0.0}, {'current': 3250.612, 'min': 0.0, 'max': 0.0}, {'current': 3195.784, 'min': 0.0, 'max': 0.0}, {'current': 3219.289, 'min': 0.0, 'max': 0.0}, {'current': 3289.163, 'min': 0.0, 'max': 0.0}, {'current': 3247.875, 'min': 0.0, 'max': 0.0}, {'current': 3249.112, 'min': 0.0, 'max': 0.0}, {'current': 3209.3, 'min': 0.0, 'max': 0.0}, {'current': 3187.212, 'min': 0.0, 'max': 0.0}, {'current': 3288.003, 'min': 0.0, 'max': 0.0}, {'current': 3169.417, 'min': 0.0, 'max': 0.0}, {'current': 3596.071, 'min': 0.0, 'max': 0.0}], 'disk': {'/': {'total': 119.94140625, 'used': 56.4015998840332}}, 'gpu': 'NVIDIA H100 80GB HBM3', 'gpu_count': 8, 'gpu_devices': [{'name': 'NVIDIA H100 80GB HBM3', 'memory_total': 85520809984}, {'name': 'NVIDIA H100 80GB HBM3', 'memory_total': 85520809984}, {'name': 'NVIDIA H100 80GB HBM3', 'memory_total': 85520809984}, {'name': 'NVIDIA H100 80GB HBM3', 'memory_total': 85520809984}, {'name': 'NVIDIA H100 80GB HBM3', 'memory_total': 85520809984}, {'name': 'NVIDIA H100 80GB HBM3', 'memory_total': 85520809984}, {'name': 'NVIDIA H100 80GB HBM3', 'memory_total': 85520809984}, {'name': 'NVIDIA H100 80GB HBM3', 'memory_total': 85520809984}], 'memory': {'total': 1999.9661331176758}}
26
+ 2024-11-13 19:02:47,022 INFO HandlerThread:1939 [system_monitor.py:probe():224] Finished collecting system info
27
+ 2024-11-13 19:02:47,022 INFO HandlerThread:1939 [system_monitor.py:probe():227] Publishing system info
28
+ 2024-11-13 19:02:47,022 DEBUG HandlerThread:1939 [system_info.py:_save_conda():208] Saving list of conda packages installed into the current environment
29
+ 2024-11-13 19:02:47,775 INFO Thread-12 :1939 [dir_watcher.py:_on_file_created():271] file/dir created: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/conda-environment.yaml
30
+ 2024-11-13 19:03:01,345 DEBUG HandlerThread:1939 [system_info.py:_save_conda():220] Saving conda packages done
31
+ 2024-11-13 19:03:01,345 DEBUG HandlerThread:1939 [system_info.py:_save_code():45] Saving code
32
+ 2024-11-13 19:03:01,349 DEBUG HandlerThread:1939 [system_info.py:_save_code():66] Saving code done
33
+ 2024-11-13 19:03:01,349 DEBUG HandlerThread:1939 [system_info.py:_save_patches():83] Saving git patches
34
+ 2024-11-13 19:03:01,776 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/conda-environment.yaml
35
+ 2024-11-13 19:03:01,776 INFO Thread-12 :1939 [dir_watcher.py:_on_file_created():271] file/dir created: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/code/qlearning_reasoning/training/sft.py
36
+ 2024-11-13 19:03:01,776 INFO Thread-12 :1939 [dir_watcher.py:_on_file_created():271] file/dir created: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/code/qlearning_reasoning
37
+ 2024-11-13 19:03:01,776 INFO Thread-12 :1939 [dir_watcher.py:_on_file_created():271] file/dir created: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/code/qlearning_reasoning/training
38
+ 2024-11-13 19:03:01,776 INFO Thread-12 :1939 [dir_watcher.py:_on_file_created():271] file/dir created: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/code
39
+ 2024-11-13 19:03:01,989 DEBUG HandlerThread:1939 [system_info.py:_save_patches():125] Saving git patches done
40
+ 2024-11-13 19:03:01,991 INFO HandlerThread:1939 [system_monitor.py:probe():229] Finished publishing system info
41
+ 2024-11-13 19:03:01,993 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
42
+ 2024-11-13 19:03:01,993 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: keepalive
43
+ 2024-11-13 19:03:01,994 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
44
+ 2024-11-13 19:03:01,994 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: keepalive
45
+ 2024-11-13 19:03:01,994 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
46
+ 2024-11-13 19:03:01,994 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: keepalive
47
+ 2024-11-13 19:03:01,994 DEBUG SenderThread:1939 [sender.py:send():382] send: files
48
+ 2024-11-13 19:03:01,995 INFO SenderThread:1939 [sender.py:_save_file():1403] saving file wandb-metadata.json with policy now
49
+ 2024-11-13 19:03:01,995 INFO SenderThread:1939 [sender.py:_save_file():1403] saving file code/qlearning_reasoning/training/sft.py with policy now
50
+ 2024-11-13 19:03:01,995 INFO SenderThread:1939 [sender.py:_save_file():1403] saving file diff.patch with policy now
51
+ 2024-11-13 19:03:01,999 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: python_packages
52
+ 2024-11-13 19:03:01,999 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: python_packages
53
+ 2024-11-13 19:03:02,001 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
54
+ 2024-11-13 19:03:02,003 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
55
+ 2024-11-13 19:03:02,003 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
56
+ 2024-11-13 19:03:02,111 DEBUG SenderThread:1939 [sender.py:send():382] send: telemetry
57
+ 2024-11-13 19:03:02,111 DEBUG SenderThread:1939 [sender.py:send():382] send: config
58
+ 2024-11-13 19:03:02,112 DEBUG SenderThread:1939 [sender.py:send():382] send: metric
59
+ 2024-11-13 19:03:02,112 DEBUG SenderThread:1939 [sender.py:send():382] send: telemetry
60
+ 2024-11-13 19:03:02,112 DEBUG SenderThread:1939 [sender.py:send():382] send: metric
61
+ 2024-11-13 19:03:02,113 WARNING SenderThread:1939 [sender.py:send_metric():1354] Seen metric with glob (shouldn't happen)
62
+ 2024-11-13 19:03:02,113 DEBUG SenderThread:1939 [sender.py:send():382] send: telemetry
63
+ 2024-11-13 19:03:02,113 DEBUG SenderThread:1939 [sender.py:send():382] send: telemetry
64
+ 2024-11-13 19:03:02,113 DEBUG SenderThread:1939 [sender.py:send():382] send: config
65
+ 2024-11-13 19:03:02,551 INFO wandb-upload_1:1939 [upload_job.py:push():131] Uploaded file /tmp/tmpm_h7h9v0wandb/pjhigh2e-code/qlearning_reasoning/training/sft.py
66
+ 2024-11-13 19:03:02,554 INFO wandb-upload_0:1939 [upload_job.py:push():131] Uploaded file /tmp/tmpm_h7h9v0wandb/f98paeu8-wandb-metadata.json
67
+ 2024-11-13 19:03:02,556 INFO wandb-upload_2:1939 [upload_job.py:push():131] Uploaded file /tmp/tmpm_h7h9v0wandb/cc8zuviy-diff.patch
68
+ 2024-11-13 19:03:02,776 INFO Thread-12 :1939 [dir_watcher.py:_on_file_created():271] file/dir created: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/diff.patch
69
+ 2024-11-13 19:03:02,776 INFO Thread-12 :1939 [dir_watcher.py:_on_file_created():271] file/dir created: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
70
+ 2024-11-13 19:03:02,777 INFO Thread-12 :1939 [dir_watcher.py:_on_file_created():271] file/dir created: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/requirements.txt
71
+ 2024-11-13 19:03:02,777 INFO Thread-12 :1939 [dir_watcher.py:_on_file_created():271] file/dir created: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/wandb-metadata.json
72
+ 2024-11-13 19:03:04,777 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
73
+ 2024-11-13 19:03:07,114 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
74
+ 2024-11-13 19:03:12,115 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
75
+ 2024-11-13 19:03:16,999 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
76
+ 2024-11-13 19:03:17,000 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
77
+ 2024-11-13 19:03:17,002 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
78
+ 2024-11-13 19:03:18,090 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
79
+ 2024-11-13 19:03:18,782 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/config.yaml
80
+ 2024-11-13 19:03:23,250 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
81
+ 2024-11-13 19:03:26,095 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: partial_history
82
+ 2024-11-13 19:03:26,097 DEBUG SenderThread:1939 [sender.py:send():382] send: metric
83
+ 2024-11-13 19:03:26,097 DEBUG SenderThread:1939 [sender.py:send():382] send: metric
84
+ 2024-11-13 19:03:26,097 DEBUG SenderThread:1939 [sender.py:send():382] send: metric
85
+ 2024-11-13 19:03:26,097 DEBUG SenderThread:1939 [sender.py:send():382] send: metric
86
+ 2024-11-13 19:03:26,097 DEBUG SenderThread:1939 [sender.py:send():382] send: history
87
+ 2024-11-13 19:03:26,098 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: summary_record
88
+ 2024-11-13 19:03:26,099 INFO SenderThread:1939 [sender.py:_save_file():1403] saving file wandb-summary.json with policy end
89
+ 2024-11-13 19:03:26,785 INFO Thread-12 :1939 [dir_watcher.py:_on_file_created():271] file/dir created: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/wandb-summary.json
90
+ 2024-11-13 19:03:28,785 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
91
+ 2024-11-13 19:03:29,100 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
92
+ 2024-11-13 19:03:31,999 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
93
+ 2024-11-13 19:03:32,000 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
94
+ 2024-11-13 19:03:32,002 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
95
+ 2024-11-13 19:03:34,147 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
96
+ 2024-11-13 19:03:39,147 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
97
+ 2024-11-13 19:03:44,148 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
98
+ 2024-11-13 19:03:46,955 DEBUG SystemMonitor:1939 [system_monitor.py:_start():172] Starting system metrics aggregation loop
99
+ 2024-11-13 19:03:46,959 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
100
+ 2024-11-13 19:03:47,000 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
101
+ 2024-11-13 19:03:47,000 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
102
+ 2024-11-13 19:03:47,043 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
103
+ 2024-11-13 19:03:49,636 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
104
+ 2024-11-13 19:03:50,813 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
105
+ 2024-11-13 19:03:50,814 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/config.yaml
106
+ 2024-11-13 19:03:54,814 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
107
+ 2024-11-13 19:03:59,814 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
108
+ 2024-11-13 19:04:02,000 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
109
+ 2024-11-13 19:04:02,000 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
110
+ 2024-11-13 19:04:02,002 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
111
+ 2024-11-13 19:04:05,061 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
112
+ 2024-11-13 19:04:10,062 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
113
+ 2024-11-13 19:04:14,822 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
114
+ 2024-11-13 19:04:15,419 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
115
+ 2024-11-13 19:04:16,962 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
116
+ 2024-11-13 19:04:17,000 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
117
+ 2024-11-13 19:04:17,000 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
118
+ 2024-11-13 19:04:17,002 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
119
+ 2024-11-13 19:04:21,087 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
120
+ 2024-11-13 19:04:26,088 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
121
+ 2024-11-13 19:04:31,089 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
122
+ 2024-11-13 19:04:32,000 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
123
+ 2024-11-13 19:04:32,000 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
124
+ 2024-11-13 19:04:32,043 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
125
+ 2024-11-13 19:04:36,587 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
126
+ 2024-11-13 19:04:36,829 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
127
+ 2024-11-13 19:04:41,588 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
128
+ 2024-11-13 19:04:46,588 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
129
+ 2024-11-13 19:04:46,965 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
130
+ 2024-11-13 19:04:47,000 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
131
+ 2024-11-13 19:04:47,000 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
132
+ 2024-11-13 19:04:47,002 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
133
+ 2024-11-13 19:04:52,111 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
134
+ 2024-11-13 19:04:57,111 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
135
+ 2024-11-13 19:04:57,383 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: partial_history
136
+ 2024-11-13 19:04:57,384 DEBUG SenderThread:1939 [sender.py:send():382] send: history
137
+ 2024-11-13 19:04:57,385 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: summary_record
138
+ 2024-11-13 19:04:57,385 INFO SenderThread:1939 [sender.py:_save_file():1403] saving file wandb-summary.json with policy end
139
+ 2024-11-13 19:04:57,837 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/wandb-summary.json
140
+ 2024-11-13 19:04:58,837 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
141
+ 2024-11-13 19:04:59,002 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: partial_history
142
+ 2024-11-13 19:04:59,004 DEBUG SenderThread:1939 [sender.py:send():382] send: metric
143
+ 2024-11-13 19:04:59,005 DEBUG SenderThread:1939 [sender.py:send():382] send: metric
144
+ 2024-11-13 19:04:59,005 DEBUG SenderThread:1939 [sender.py:send():382] send: metric
145
+ 2024-11-13 19:04:59,005 DEBUG SenderThread:1939 [sender.py:send():382] send: metric
146
+ 2024-11-13 19:04:59,006 DEBUG SenderThread:1939 [sender.py:send():382] send: history
147
+ 2024-11-13 19:04:59,006 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: summary_record
148
+ 2024-11-13 19:04:59,006 INFO SenderThread:1939 [sender.py:_save_file():1403] saving file wandb-summary.json with policy end
149
+ 2024-11-13 19:04:59,838 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/wandb-summary.json
150
+ 2024-11-13 19:05:00,838 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
151
+ 2024-11-13 19:05:02,000 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
152
+ 2024-11-13 19:05:02,001 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
153
+ 2024-11-13 19:05:02,002 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
154
+ 2024-11-13 19:05:03,111 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
155
+ 2024-11-13 19:05:08,112 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
156
+ 2024-11-13 19:05:13,113 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
157
+ 2024-11-13 19:05:16,967 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
158
+ 2024-11-13 19:05:17,000 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
159
+ 2024-11-13 19:05:17,001 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
160
+ 2024-11-13 19:05:17,043 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
161
+ 2024-11-13 19:05:18,123 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
162
+ 2024-11-13 19:05:22,846 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
163
+ 2024-11-13 19:05:23,982 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
164
+ 2024-11-13 19:05:25,847 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/config.yaml
165
+ 2024-11-13 19:05:30,486 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
166
+ 2024-11-13 19:05:32,000 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
167
+ 2024-11-13 19:05:32,001 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
168
+ 2024-11-13 19:05:32,002 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
169
+ 2024-11-13 19:05:36,117 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
170
+ 2024-11-13 19:05:41,118 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
171
+ 2024-11-13 19:05:46,513 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
172
+ 2024-11-13 19:05:46,854 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
173
+ 2024-11-13 19:05:46,970 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
174
+ 2024-11-13 19:05:47,000 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
175
+ 2024-11-13 19:05:47,001 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
176
+ 2024-11-13 19:05:47,003 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
177
+ 2024-11-13 19:05:52,097 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
178
+ 2024-11-13 19:05:57,098 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
179
+ 2024-11-13 19:06:02,001 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
180
+ 2024-11-13 19:06:02,001 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
181
+ 2024-11-13 19:06:02,043 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
182
+ 2024-11-13 19:06:02,152 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
183
+ 2024-11-13 19:06:07,153 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
184
+ 2024-11-13 19:06:08,862 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
185
+ 2024-11-13 19:06:12,639 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
186
+ 2024-11-13 19:06:16,973 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
187
+ 2024-11-13 19:06:17,001 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
188
+ 2024-11-13 19:06:17,002 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
189
+ 2024-11-13 19:06:17,004 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
190
+ 2024-11-13 19:06:18,127 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
191
+ 2024-11-13 19:06:23,127 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
192
+ 2024-11-13 19:06:28,128 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
193
+ 2024-11-13 19:06:30,870 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
194
+ 2024-11-13 19:06:32,001 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
195
+ 2024-11-13 19:06:32,001 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
196
+ 2024-11-13 19:06:32,003 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
197
+ 2024-11-13 19:06:33,143 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
198
+ 2024-11-13 19:06:38,144 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
199
+ 2024-11-13 19:06:43,144 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
200
+ 2024-11-13 19:06:46,975 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
201
+ 2024-11-13 19:06:47,001 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
202
+ 2024-11-13 19:06:47,001 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
203
+ 2024-11-13 19:06:47,043 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
204
+ 2024-11-13 19:06:49,137 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
205
+ 2024-11-13 19:06:51,931 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: partial_history
206
+ 2024-11-13 19:06:51,932 DEBUG SenderThread:1939 [sender.py:send():382] send: history
207
+ 2024-11-13 19:06:51,933 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: summary_record
208
+ 2024-11-13 19:06:51,934 INFO SenderThread:1939 [sender.py:_save_file():1403] saving file wandb-summary.json with policy end
209
+ 2024-11-13 19:06:52,878 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
210
+ 2024-11-13 19:06:52,878 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/wandb-summary.json
211
+ 2024-11-13 19:06:53,541 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: partial_history
212
+ 2024-11-13 19:06:53,543 DEBUG SenderThread:1939 [sender.py:send():382] send: history
213
+ 2024-11-13 19:06:53,543 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: summary_record
214
+ 2024-11-13 19:06:53,544 INFO SenderThread:1939 [sender.py:_save_file():1403] saving file wandb-summary.json with policy end
215
+ 2024-11-13 19:06:53,878 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/wandb-summary.json
216
+ 2024-11-13 19:06:54,545 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
217
+ 2024-11-13 19:06:54,878 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
218
+ 2024-11-13 19:06:59,546 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
219
+ 2024-11-13 19:07:02,001 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
220
+ 2024-11-13 19:07:02,002 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
221
+ 2024-11-13 19:07:02,004 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
222
+ 2024-11-13 19:07:05,164 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
223
+ 2024-11-13 19:07:10,165 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
224
+ 2024-11-13 19:07:15,165 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
225
+ 2024-11-13 19:07:16,886 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
226
+ 2024-11-13 19:07:16,978 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
227
+ 2024-11-13 19:07:17,001 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
228
+ 2024-11-13 19:07:17,002 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
229
+ 2024-11-13 19:07:17,003 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
230
+ 2024-11-13 19:07:21,108 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
231
+ 2024-11-13 19:07:26,108 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
232
+ 2024-11-13 19:07:31,109 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
233
+ 2024-11-13 19:07:32,001 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
234
+ 2024-11-13 19:07:32,002 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
235
+ 2024-11-13 19:07:32,043 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
236
+ 2024-11-13 19:07:36,934 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
237
+ 2024-11-13 19:07:38,894 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
238
+ 2024-11-13 19:07:41,935 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
239
+ 2024-11-13 19:07:46,935 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
240
+ 2024-11-13 19:07:46,981 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
241
+ 2024-11-13 19:07:47,002 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
242
+ 2024-11-13 19:07:47,002 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
243
+ 2024-11-13 19:07:47,004 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
244
+ 2024-11-13 19:07:52,135 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
245
+ 2024-11-13 19:07:57,136 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
246
+ 2024-11-13 19:08:02,136 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
247
+ 2024-11-13 19:08:02,856 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
248
+ 2024-11-13 19:08:02,923 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
249
+ 2024-11-13 19:08:02,924 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
250
+ 2024-11-13 19:08:08,076 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
251
+ 2024-11-13 19:08:13,077 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
252
+ 2024-11-13 19:08:16,985 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
253
+ 2024-11-13 19:08:17,407 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
254
+ 2024-11-13 19:08:17,407 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
255
+ 2024-11-13 19:08:17,447 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
256
+ 2024-11-13 19:08:18,545 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
257
+ 2024-11-13 19:08:23,546 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
258
+ 2024-11-13 19:08:28,546 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
259
+ 2024-11-13 19:08:32,914 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
260
+ 2024-11-13 19:08:33,843 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
261
+ 2024-11-13 19:08:34,729 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
262
+ 2024-11-13 19:08:34,729 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
263
+ 2024-11-13 19:08:34,771 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
264
+ 2024-11-13 19:08:39,802 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
265
+ 2024-11-13 19:08:44,803 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
266
+ 2024-11-13 19:08:46,987 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
267
+ 2024-11-13 19:08:48,007 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
268
+ 2024-11-13 19:08:48,007 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
269
+ 2024-11-13 19:08:48,008 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
270
+ 2024-11-13 19:08:48,921 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
271
+ 2024-11-13 19:08:49,945 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
272
+ 2024-11-13 19:08:50,922 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
273
+ 2024-11-13 19:08:52,923 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
274
+ 2024-11-13 19:08:54,924 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
275
+ 2024-11-13 19:08:55,154 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
276
+ 2024-11-13 19:08:56,925 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
277
+ 2024-11-13 19:08:58,926 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
278
+ 2024-11-13 19:09:00,371 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
279
+ 2024-11-13 19:09:00,948 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
280
+ 2024-11-13 19:09:03,950 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
281
+ 2024-11-13 19:09:05,541 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
282
+ 2024-11-13 19:09:05,782 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
283
+ 2024-11-13 19:09:06,102 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
284
+ 2024-11-13 19:09:06,102 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
285
+ 2024-11-13 19:09:11,263 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
286
+ 2024-11-13 19:09:16,264 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
287
+ 2024-11-13 19:09:16,990 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
288
+ 2024-11-13 19:09:20,961 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
289
+ 2024-11-13 19:09:22,073 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
290
+ 2024-11-13 19:09:22,272 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
291
+ 2024-11-13 19:09:22,272 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
292
+ 2024-11-13 19:09:27,401 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
293
+ 2024-11-13 19:09:32,401 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
294
+ 2024-11-13 19:09:36,126 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
295
+ 2024-11-13 19:09:36,666 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
296
+ 2024-11-13 19:09:36,666 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
297
+ 2024-11-13 19:09:38,052 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
298
+ 2024-11-13 19:09:39,964 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
299
+ 2024-11-13 19:09:43,769 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
300
+ 2024-11-13 19:09:46,992 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
301
+ 2024-11-13 19:09:47,846 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
302
+ 2024-11-13 19:09:47,847 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
303
+ 2024-11-13 19:09:47,847 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
304
+ 2024-11-13 19:09:49,230 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
305
+ 2024-11-13 19:09:54,231 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
306
+ 2024-11-13 19:09:59,231 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
307
+ 2024-11-13 19:10:02,847 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
308
+ 2024-11-13 19:10:02,848 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
309
+ 2024-11-13 19:10:02,848 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
310
+ 2024-11-13 19:10:05,230 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
311
+ 2024-11-13 19:10:10,231 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
312
+ 2024-11-13 19:10:11,977 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
313
+ 2024-11-13 19:10:15,248 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
314
+ 2024-11-13 19:10:16,996 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
315
+ 2024-11-13 19:10:17,847 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
316
+ 2024-11-13 19:10:17,848 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
317
+ 2024-11-13 19:10:17,848 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
318
+ 2024-11-13 19:10:21,230 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
319
+ 2024-11-13 19:10:26,231 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
320
+ 2024-11-13 19:10:31,231 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
321
+ 2024-11-13 19:10:32,847 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
322
+ 2024-11-13 19:10:32,847 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
323
+ 2024-11-13 19:10:32,847 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
324
+ 2024-11-13 19:10:33,985 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
325
+ 2024-11-13 19:10:37,230 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
326
+ 2024-11-13 19:10:42,231 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
327
+ 2024-11-13 19:10:46,998 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
328
+ 2024-11-13 19:10:47,847 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
329
+ 2024-11-13 19:10:47,847 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
330
+ 2024-11-13 19:10:47,887 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
331
+ 2024-11-13 19:10:47,921 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
332
+ 2024-11-13 19:10:53,230 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
333
+ 2024-11-13 19:10:53,575 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: partial_history
334
+ 2024-11-13 19:10:53,576 DEBUG SenderThread:1939 [sender.py:send():382] send: history
335
+ 2024-11-13 19:10:53,576 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: summary_record
336
+ 2024-11-13 19:10:53,577 INFO SenderThread:1939 [sender.py:_save_file():1403] saving file wandb-summary.json with policy end
337
+ 2024-11-13 19:10:53,992 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/wandb-summary.json
338
+ 2024-11-13 19:10:54,993 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
339
+ 2024-11-13 19:10:55,200 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: partial_history
340
+ 2024-11-13 19:10:55,201 DEBUG SenderThread:1939 [sender.py:send():382] send: history
341
+ 2024-11-13 19:10:55,201 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: summary_record
342
+ 2024-11-13 19:10:55,203 INFO SenderThread:1939 [sender.py:_save_file():1403] saving file wandb-summary.json with policy end
343
+ 2024-11-13 19:10:55,993 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
344
+ 2024-11-13 19:10:55,994 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/wandb-summary.json
345
+ 2024-11-13 19:10:56,994 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
346
+ 2024-11-13 19:10:57,994 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
347
+ 2024-11-13 19:10:59,204 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
348
+ 2024-11-13 19:11:02,848 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
349
+ 2024-11-13 19:11:02,848 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
350
+ 2024-11-13 19:11:02,849 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
351
+ 2024-11-13 19:11:04,230 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
352
+ 2024-11-13 19:11:09,231 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
353
+ 2024-11-13 19:11:14,231 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
354
+ 2024-11-13 19:11:17,000 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
355
+ 2024-11-13 19:11:17,847 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
356
+ 2024-11-13 19:11:17,848 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
357
+ 2024-11-13 19:11:17,848 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
358
+ 2024-11-13 19:11:19,311 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
359
+ 2024-11-13 19:11:20,004 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
360
+ 2024-11-13 19:11:24,312 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
361
+ 2024-11-13 19:11:29,312 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
362
+ 2024-11-13 19:11:32,848 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
363
+ 2024-11-13 19:11:32,848 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
364
+ 2024-11-13 19:11:32,849 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
365
+ 2024-11-13 19:11:35,231 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
366
+ 2024-11-13 19:11:40,232 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
367
+ 2024-11-13 19:11:44,014 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
368
+ 2024-11-13 19:11:46,173 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
369
+ 2024-11-13 19:11:47,003 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
370
+ 2024-11-13 19:11:47,847 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
371
+ 2024-11-13 19:11:47,847 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
372
+ 2024-11-13 19:11:47,848 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
373
+ 2024-11-13 19:11:51,230 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
wandb/debug.log ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-11-13 19:02:46,428 INFO MainThread:173 [wandb_setup.py:_flush():76] Current SDK version is 0.16.3
2
+ 2024-11-13 19:02:46,428 INFO MainThread:173 [wandb_setup.py:_flush():76] Configure stats pid to 173
3
+ 2024-11-13 19:02:46,428 INFO MainThread:173 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
4
+ 2024-11-13 19:02:46,428 INFO MainThread:173 [wandb_setup.py:_flush():76] Loading settings from /opt/ml/code/wandb/settings
5
+ 2024-11-13 19:02:46,428 INFO MainThread:173 [wandb_setup.py:_flush():76] Loading settings from environment variables: {'root_dir': '/opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048', 'api_key': '***REDACTED***', 'project': 'sft-codecontests-1112', 'run_name': 'sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048'}
6
+ 2024-11-13 19:02:46,428 INFO MainThread:173 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
7
+ 2024-11-13 19:02:46,428 INFO MainThread:173 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': 'qlearning_reasoning/training/sft.py', 'program_abspath': '/opt/ml/code/qlearning_reasoning/training/sft.py', 'program': '/opt/ml/code/qlearning_reasoning/training/sft.py'}
8
+ 2024-11-13 19:02:46,428 INFO MainThread:173 [wandb_init.py:_log_setup():526] Logging user logs to /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/logs/debug.log
9
+ 2024-11-13 19:02:46,428 INFO MainThread:173 [wandb_init.py:_log_setup():527] Logging internal logs to /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/logs/debug-internal.log
10
+ 2024-11-13 19:02:46,428 INFO MainThread:173 [wandb_init.py:init():566] calling init triggers
11
+ 2024-11-13 19:02:46,428 INFO MainThread:173 [wandb_init.py:init():573] wandb.init called with sweep_config: {}
12
+ config: {'sagemaker_training_job_name': 'kushalarora-rvv-main-2024-11-13-17-55-42-664', 'batch_size': 64, 'cache_dir': '/opt/ml/data/input/.cache', 'dataset': 'code-contests', 'gradient_accumulation_steps': 8, 'logging_steps': 5, 'lr': '1e-6', 'max_seq_length': 2048, 'model_name': 'Qwen/Qwen2.5-Coder-7B-Instruct', 'num_train_epochs': 2, 'output_dir': '/opt/ml/output', 'push_to_hub': 'true', 'run_name': 'sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048', 'sagemaker_container_log_level': 20, 'sagemaker_instance_type': 'ml.p5.48xlarge', 'sagemaker_job_name': 'kushalarora-rvv-main-2024-11-13-17-55-42-664', 'sagemaker_program': 'sft.py', 'sagemaker_region': 'us-east-1', 'sagemaker_submit_directory': 's3://tri-ml-datasets/kushalarora-rvv-main-2024-11-13-17-55-42-664/source/sourcedir.tar.gz', 'sagemaker_torch_distributed_enabled': 'true', 'save_steps': 100, 'schedule': 'cosine', 'wandb_project': 'sft-codecontests-1112', 'weight_decay': 0.0, 'additional_framework_parameters': {'sagemaker_instance_type': 'ml.p5.48xlarge', 'sagemaker_torch_distributed_enabled': True}, 'channel_input_dirs': {}, 'current_host': 'algo-1', 'current_instance_group': 'homogeneousCluster', 'current_instance_group_hosts': ['algo-1'], 'current_instance_type': 'ml.p5.48xlarge', 'distribution_hosts': ['algo-1'], 'distribution_instance_groups': ['homogeneousCluster'], 'framework_module': 'sagemaker_pytorch_container.training:main', 'hosts': ['algo-1'], 'hyperparameters': {'batch_size': 64, 'cache_dir': '/opt/ml/data/input/.cache', 'dataset': 'code-contests', 'gradient_accumulation_steps': 8, 'logging_steps': 5, 'lr': '1e-6', 'max_seq_length': 2048, 'model_name': 'Qwen/Qwen2.5-Coder-7B-Instruct', 'num_train_epochs': 2, 'output_dir': '/opt/ml/model/', 'push_to_hub': True, 'run_name': 'sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048', 'save_steps': 100, 'schedule': 'cosine', 'wandb_project': 'sft-codecontests-1112', 'weight_decay': 0.0}, 'input_config_dir': '/opt/ml/input/config', 'input_data_config': {}, 'input_dir': '/opt/ml/input', 'instance_groups': ['homogeneousCluster'], 'instance_groups_dict': {'homogeneousCluster': {'hosts': ['algo-1'], 'instance_group_name': 'homogeneousCluster', 'instance_type': 'ml.p5.48xlarge'}}, 'is_hetero': False, 'is_master': True, 'is_modelparallel_enabled': None, 'is_smddpmprun_installed': False, 'is_smddprun_installed': True, 'job_name': 'kushalarora-rvv-main-2024-11-13-17-55-42-664', 'log_level': 20, 'master_hostname': 'algo-1', 'model_dir': '/opt/ml/model', 'module_dir': 's3://tri-ml-datasets/kushalarora-rvv-main-2024-11-13-17-55-42-664/source/sourcedir.tar.gz', 'module_name': '/opt/ml/code/qlearning_reasoning/training/sft', 'network_interface_name': 'eth0', 'num_cpus': 192, 'num_gpus': 8, 'num_neurons': 0, 'output_data_dir': '/opt/ml/output/data', 'output_intermediate_dir': '/opt/ml/output/intermediate', 'resource_config': {'current_group_name': 'homogeneousCluster', 'current_host': 'algo-1', 'current_instance_type': 'ml.p5.48xlarge', 'hosts': ['algo-1'], 'instance_groups': [{'hosts': ['algo-1'], 'instance_group_name': 'homogeneousCluster', 'instance_type': 'ml.p5.48xlarge'}], 'network_interface_name': 'eth0'}, 'user_entry_point': '/opt/ml/code/qlearning_reasoning/training/sft.py'}
13
+ 2024-11-13 19:02:46,428 INFO MainThread:173 [wandb_init.py:init():616] starting backend
14
+ 2024-11-13 19:02:46,428 INFO MainThread:173 [wandb_init.py:init():620] setting up manager
15
+ 2024-11-13 19:02:46,429 INFO MainThread:173 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
16
+ 2024-11-13 19:02:46,431 INFO MainThread:173 [wandb_init.py:init():628] backend started and connected
17
+ 2024-11-13 19:02:46,434 INFO MainThread:173 [wandb_init.py:init():720] updated telemetry
18
+ 2024-11-13 19:02:46,440 INFO MainThread:173 [wandb_init.py:init():753] communicating run to backend with 90.0 second timeout
19
+ 2024-11-13 19:02:46,781 INFO MainThread:173 [wandb_run.py:_on_init():2262] communicating current version
20
+ 2024-11-13 19:02:46,812 INFO MainThread:173 [wandb_run.py:_on_init():2271] got version response upgrade_message: "wandb version 0.18.6 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
21
+
22
+ 2024-11-13 19:02:46,813 INFO MainThread:173 [wandb_init.py:init():804] starting run threads in backend
23
+ 2024-11-13 19:03:02,000 INFO MainThread:173 [wandb_run.py:_console_start():2241] atexit reg
24
+ 2024-11-13 19:03:02,000 INFO MainThread:173 [wandb_run.py:_redirect():2096] redirect: wrap_raw
25
+ 2024-11-13 19:03:02,000 INFO MainThread:173 [wandb_run.py:_redirect():2161] Wrapping output streams.
26
+ 2024-11-13 19:03:02,001 INFO MainThread:173 [wandb_run.py:_redirect():2186] Redirects installed.
27
+ 2024-11-13 19:03:02,001 INFO MainThread:173 [wandb_init.py:init():847] run started, returning control to user process
28
+ 2024-11-13 19:03:02,004 INFO MainThread:173 [wandb_run.py:_config_callback():1343] config_cb None None {'vocab_size': 151671, 'max_position_embeddings': 32768, 'hidden_size': 3584, 'intermediate_size': 18944, 'num_hidden_layers': 28, 'num_attention_heads': 28, 'use_sliding_window': False, 'sliding_window': None, 'max_window_layers': 28, 'num_key_value_heads': 4, 'hidden_act': 'silu', 'initializer_range': 0.02, 'rms_norm_eps': 1e-06, 'use_cache': False, 'rope_theta': 1000000.0, 'rope_scaling': None, 'attention_dropout': 0.0, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'bfloat16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['Qwen2ForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': 151643, 'eos_token_id': 151645, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'Qwen/Qwen2.5-Coder-7B-Instruct', '_attn_implementation_autoset': True, 'transformers_version': '4.46.0', 'model_type': 'qwen2', 'output_dir': '/opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'eval_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 8, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 8, 'eval_accumulation_steps': None, 'eval_delay': 0, 'torch_empty_cache_steps': None, 'learning_rate': 1e-06, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 2, 'max_steps': -1, 'lr_scheduler_type': 'cosine', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/runs/Nov13_19-02-15_algo-1', 'logging_strategy': 'steps', 'logging_first_step': True, 'logging_steps': 5, 'logging_nan_inf_filter': True, 'save_strategy': 'epoch', 'save_steps': 100, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'restore_callback_states_from_checkpoint': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': 5, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': 'sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': ['full_shard', 'auto_wrap'], 'fsdp_min_num_params': 0, 'fsdp_config': {'activation_checkpointing': False, 'auto_wrap_policy': 'TRANSFORMER_BASED_WRAP', 'backward_prefetch': 'BACKWARD_PRE', 'cpu_ram_efficient_loading': True, 'forward_prefetch': False, 'offload_params': False, 'sharding_strategy': 'FULL_SHARD', 'state_dict_type': 'FULL_STATE_DICT', 'sync_module_states': True, 'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': 'sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_', 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': {'use_reentrant': False}, 'include_inputs_for_metrics': False, 'include_for_metrics': [], 'eval_do_concat_batches': True, 'fp16_backend': 'auto', 'evaluation_strategy': None, 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None, 'batch_eval_metrics': False, 'eval_on_start': False, 'use_liger_kernel': False, 'eval_use_gather_object': False, 'dataset_text_field': 'text', 'packing': False, 'max_seq_length': 2048, 'dataset_num_proc': None, 'dataset_batch_size': 64, 'model_init_kwargs': None, 'dataset_kwargs': {}, 'eval_packing': None, 'num_of_sequences': 1024, 'chars_per_token': '<CHARS_PER_TOKEN>', 'use_liger': False}
29
+ 2024-11-13 19:03:02,008 INFO MainThread:173 [wandb_config.py:__setitem__():151] config set model/num_parameters = 951599936 - <bound method Run._config_callback of <wandb.sdk.wandb_run.Run object at 0x7fb464957310>>
30
+ 2024-11-13 19:03:02,009 INFO MainThread:173 [wandb_run.py:_config_callback():1343] config_cb model/num_parameters 951599936 None
wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/code/qlearning_reasoning/training/sft.py ADDED
@@ -0,0 +1,249 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ from datasets import load_dataset, concatenate_datasets
2
+ from transformers import AutoTokenizer, AutoModelForCausalLM, TrainingArguments
3
+ import torch
4
+ from transformers import get_constant_schedule_with_warmup, get_cosine_schedule_with_warmup, get_linear_schedule_with_warmup, get_inverse_sqrt_schedule
5
+ from trl import SFTTrainer, SFTConfig, DataCollatorForCompletionOnlyLM
6
+ import os
7
+ from textwrap import dedent
8
+ from absl import app, flags
9
+ FLAGS = flags.FLAGS
10
+ import warnings
11
+ warnings.filterwarnings("ignore")
12
+ from trl import setup_chat_format
13
+
14
+ ALL_DATASET_NAMES = [
15
+ 'prm800k',
16
+ 'prm800k-improvement',
17
+ 'math',
18
+ 'math-v2',
19
+ 'math-prm800k',
20
+ 'math-revision',
21
+ 'og-math',
22
+ 'leetcode',
23
+ 'taco',
24
+ 'math-revision-onesamp',
25
+ 'math-revision-onesamp-overlap',
26
+ 'math-revision-onesamp-correct-overlap',
27
+ 'math-revision-onesamp-correct-overlap',
28
+ 'code-contests'
29
+ ]
30
+
31
+ flags.DEFINE_enum('dataset', 'prm800k', ALL_DATASET_NAMES, 'Name of the dataset')
32
+ flags.DEFINE_string('cache_dir', '/scratch/bclj/asingh15/.cache', 'Directory to store cache files')
33
+ flags.DEFINE_string('output_dir', '/scratch/bclj/asingh15/sft_checkpoints/', 'Directory to store output files')
34
+ flags.DEFINE_string('run_name', 'sft', 'Name of the run')
35
+ flags.DEFINE_string('model_name', 'deepseek-ai/DeepSeek-Coder-V2-Lite-Instruct', 'Name of the model')
36
+ flags.DEFINE_string('token', None, 'Hugging Face API token')
37
+ flags.DEFINE_string('wandb_project', 'sft-prm800k-pregen-0501', 'Wandb project name')
38
+ flags.DEFINE_string('optim', 'adamw_torch', 'Optimizer')
39
+ flags.DEFINE_string('schedule', 'constant', 'Learning rate schedule')
40
+ flags.DEFINE_float('lr', 5e-5, 'Learning rate')
41
+ flags.DEFINE_float('weight_decay', 0.0, 'Weight decay')
42
+ flags.DEFINE_integer('warmup_steps', 0, 'Number of warmup steps')
43
+ flags.DEFINE_integer('num_train_epochs', 1, 'Number of training epochs')
44
+ flags.DEFINE_integer('batch_size', 256, 'Batch size')
45
+ flags.DEFINE_integer('gradient_accumulation_steps', 64, 'Gradient accumulation steps')
46
+ flags.DEFINE_integer('max_seq_length', 1024, 'Maximum sequence length')
47
+ flags.DEFINE_integer('logging_steps', 100, 'Logging steps')
48
+ flags.DEFINE_integer('save_steps', 1000, 'Save steps')
49
+ flags.DEFINE_integer('seed', 42, 'Random seed')
50
+ flags.DEFINE_string('fsdp', "full_shard auto_wrap", 'If to use FSDP')
51
+ flags.DEFINE_string('fsdp_config', "qlearning_reasoning/accelerate_configs/fsdp_config.json", 'FSDP config to use.')
52
+ flags.DEFINE_boolean("push_to_hub", False, "Should the model be uploaded to Huggingface hub?")
53
+ flags.DEFINE_boolean('add_step_tokens', True, 'Add step tokens')
54
+ flags.DEFINE_string("hub_model_id", None, "Hub model id. If not defined it will be truncated run name.")
55
+
56
+ START_OF_STEP_TOKEN = '[STEP]'
57
+ END_OF_STEP_TOKEN = '[/STEP]'
58
+ START_OF_TURN_TOKEN = '[TURN]'
59
+ END_OF_TURN_TOKEN = '[/TURN]'
60
+ START_OF_REVISION_TOKEN = '[REVISION]'
61
+ END_OF_REVISION_TOKEN = '[/REVISION]'
62
+
63
+ def main(_):
64
+
65
+ if FLAGS.hub_model_id is None:
66
+ FLAGS.hub_model_id = FLAGS.run_name[:96]
67
+
68
+ if not os.path.exists(FLAGS.output_dir):
69
+ os.makedirs(FLAGS.output_dir, exist_ok=True)
70
+ os.makedirs(f"{FLAGS.output_dir}/{FLAGS.run_name}", exist_ok=True)
71
+
72
+ os.environ["WANDB_PROJECT"] = FLAGS.wandb_project
73
+ os.environ["WANDB_DIR"] = f"{FLAGS.output_dir}/{FLAGS.run_name}"
74
+ os.environ["WANDB_NAME"] = FLAGS.run_name
75
+ os.environ["HF_DATASETS_CACHE"] = FLAGS.cache_dir
76
+
77
+ shared_kwargs = dict(
78
+ cache_dir=FLAGS.cache_dir,
79
+ trust_remote_code=True,
80
+ )
81
+ tokenizer = AutoTokenizer.from_pretrained(
82
+ FLAGS.model_name,
83
+ **shared_kwargs,
84
+ )
85
+ if tokenizer.pad_token is None:
86
+ tokenizer.pad_token = tokenizer.eos_token
87
+
88
+ if FLAGS.add_step_tokens:
89
+ tokenizer.add_special_tokens(
90
+ {'additional_special_tokens': [START_OF_STEP_TOKEN, END_OF_STEP_TOKEN, START_OF_TURN_TOKEN, END_OF_TURN_TOKEN, START_OF_REVISION_TOKEN, END_OF_REVISION_TOKEN]}
91
+ )
92
+
93
+ print(f"Tokenizer vocab size: {len(tokenizer)}")
94
+ print('EOS token:', tokenizer.eos_token)
95
+ print('PAD token:', tokenizer.pad_token)
96
+ print('Special tokens:', tokenizer.special_tokens_map)
97
+
98
+ instruction_template = "# Question\n\n"
99
+ if FLAGS.dataset == 'prm800k':
100
+ dataset_name = 'Asap7772/sft_prm800k_processed'
101
+ response_template = "# Solution\n\n"
102
+ elif FLAGS.dataset == 'math':
103
+ dataset_name = 'Asap7772/sft_math_processed'
104
+ response_template = "# Solution\n\n"
105
+ elif FLAGS.dataset == 'math-prm800k':
106
+ dataset_name = ['Asap7772/sft_prm800k_processed', 'Asap7772/sft_math_processed']
107
+ response_template = "# Solution\n\n"
108
+ elif FLAGS.dataset == 'og-math':
109
+ dataset_name = 'Asap7772/hendrycks_math_sft_formatted'
110
+ response_template = "# Solution\n\n"
111
+ elif FLAGS.dataset == 'leetcode':
112
+ dataset_name = 'Asap7772/leetcode-rosetta-processed'
113
+ instruction_template = "# Question"
114
+ response_template = "# Solution"
115
+ elif FLAGS.dataset == 'prm800k-improvement':
116
+ dataset_name = 'Asap7772/prm800k_improvement_groundtruth_values'
117
+ instruction_template = 'Solution Steps:'
118
+ response_template = "Improved Step:"
119
+ elif FLAGS.dataset == 'taco':
120
+ dataset_name = 'Asap7772/TACO-Processed'
121
+ instruction_template = "# Question"
122
+ response_template = "# Solution"
123
+ elif FLAGS.dataset == 'math-revision':
124
+ dataset_name = 'Asap7772/hendrycks-math-mc-llama-sftnoic-revisions-filtered'
125
+ instruction_template = "# Question"
126
+ response_template = f"{END_OF_TURN_TOKEN}\n\n{START_OF_TURN_TOKEN}"
127
+ elif FLAGS.dataset == 'math-revision-ic':
128
+ dataset_name = 'Asap7772/hendrycks-math-mc-llama-revisions-filtered'
129
+ instruction_template = "# Question"
130
+ response_template = f"{END_OF_TURN_TOKEN}\n\n{START_OF_TURN_TOKEN}"
131
+ elif FLAGS.dataset == 'math-revision-onesamp':
132
+ dataset_name = 'Asap7772/hendrycks-math-onesamp-revision'
133
+ instruction_template = "# Question"
134
+ response_template = "# Solution"
135
+ elif FLAGS.dataset == 'math-revision-onesamp-overlap':
136
+ dataset_name = 'Asap7772/hendrycks-math-onesamp-revision-overlap'
137
+ instruction_template = "# Question"
138
+ response_template = "# Solution"
139
+ elif FLAGS.dataset == 'math-v2':
140
+ dataset_name = 'Asap7772/sft_math_405b'
141
+ instruction_template = "# Question"
142
+ response_template = "# Solution"
143
+ elif FLAGS.dataset == 'math-revision-onesamp-correct-overlap':
144
+ dataset_name = 'Asap7772/hendrycks-math-onesamp-revision-correct-overlap'
145
+ instruction_template = "# Question"
146
+ response_template = "# Solution"
147
+ elif FLAGS.dataset == 'code-contests':
148
+ dataset_name = 'Asap7772/code_contests'
149
+ instruction_template = "Q:"
150
+ response_template = "A:"
151
+ else:
152
+ raise ValueError(f"Invalid dataset: {FLAGS.dataset}")
153
+
154
+ if isinstance(dataset_name, list):
155
+ all_datasets = [load_dataset(name, cache_dir=FLAGS.cache_dir) for name in dataset_name]
156
+ train_dataset = concatenate_datasets([dataset['train'] for dataset in all_datasets])
157
+ eval_dataset = concatenate_datasets([dataset['test'] for dataset in all_datasets])
158
+ print(f"example: {train_dataset[0]}")
159
+ elif isinstance(dataset_name, str):
160
+ dataset_dict = load_dataset(dataset_name, cache_dir=FLAGS.cache_dir)
161
+ train_dataset, eval_dataset = dataset_dict['train'], dataset_dict['test']
162
+ else:
163
+ raise ValueError(f"Invalid dataset name: {dataset_name}")
164
+ train_dataset, eval_dataset = train_dataset.shuffle(seed=FLAGS.seed), eval_dataset.shuffle(seed=FLAGS.seed)
165
+
166
+ num_train_epochs = FLAGS.num_train_epochs
167
+
168
+ def formatting_prompts_func(example):
169
+ output_texts = []
170
+ for i in range(len(example['text'])):
171
+ if FLAGS.dataset in ALL_DATASET_NAMES:
172
+ text = f"{example['text'][i]}{tokenizer.eos_token}"
173
+ else:
174
+ raise ValueError(f"Invalid dataset: {FLAGS.dataset}")
175
+
176
+ text = dedent(text).strip()
177
+ output_texts.append(text)
178
+ return output_texts
179
+ format_example = formatting_prompts_func(train_dataset[0:1])
180
+ print(f"Example formatted prompt: \n\n{format_example[0]}\n\n")
181
+ assert instruction_template in format_example[0], f"Instruction template not found in text."
182
+ assert response_template in format_example[0], f"Response template not found in text."
183
+ assert tokenizer.eos_token in format_example[0], f"EOS token not found in text."
184
+
185
+ instruction_template_ids = tokenizer.encode(instruction_template, add_special_tokens=False)[:2]
186
+ response_template_ids = tokenizer.encode(response_template, add_special_tokens=False)[:2]
187
+ collator = DataCollatorForCompletionOnlyLM(instruction_template=instruction_template_ids, response_template=response_template_ids, tokenizer=tokenizer, mlm=False)
188
+
189
+ model = AutoModelForCausalLM.from_pretrained(
190
+ FLAGS.model_name,
191
+ low_cpu_mem_usage=True,
192
+ torch_dtype=torch.bfloat16,
193
+ **shared_kwargs,
194
+ )
195
+ model.resize_token_embeddings(len(tokenizer))
196
+ if getattr(model, "config", None) is not None:
197
+ model.config.pad_token_id = tokenizer.pad_token_id
198
+ model.config.bos_token_id = tokenizer.bos_token_id
199
+ model.config.eos_token_id = tokenizer.eos_token_id
200
+ # Update the generation config to use the new eos & bos token
201
+ if getattr(model, "generation_config", None) is not None:
202
+ model.generation_config.bos_token_id = tokenizer.bos_token_id
203
+ model.generation_config.eos_token_id = tokenizer.eos_token_id
204
+ model.generation_config.pad_token_id = tokenizer.pad_token_id
205
+
206
+ model.config.use_cache = False
207
+ model.gradient_checkpointing_enable()
208
+
209
+ transformer_args = SFTConfig(
210
+ run_name=FLAGS.run_name,
211
+ output_dir=f"{FLAGS.output_dir}/{FLAGS.run_name}",
212
+ learning_rate=FLAGS.lr,
213
+ eval_strategy="steps",
214
+ logging_strategy="steps",
215
+ logging_first_step=True,
216
+ save_strategy="epoch",
217
+ dataset_batch_size=FLAGS.batch_size,
218
+ max_seq_length=FLAGS.max_seq_length,
219
+ gradient_accumulation_steps=FLAGS.gradient_accumulation_steps,
220
+ gradient_checkpointing_kwargs={'use_reentrant':False},
221
+ logging_steps=FLAGS.logging_steps,
222
+ save_steps=FLAGS.save_steps,
223
+ lr_scheduler_type=FLAGS.schedule,
224
+ optim=FLAGS.optim,
225
+ weight_decay=FLAGS.weight_decay,
226
+ warmup_steps=FLAGS.warmup_steps,
227
+ fsdp=FLAGS.fsdp,
228
+ fsdp_config=FLAGS.fsdp_config if FLAGS.fsdp != '' else None,
229
+ push_to_hub=FLAGS.push_to_hub,
230
+ num_train_epochs=num_train_epochs,
231
+ hub_model_id=FLAGS.hub_model_id
232
+ )
233
+
234
+ flags_dict = FLAGS.flag_values_dict()
235
+ trainer = SFTTrainer(
236
+ model=model,
237
+ args=transformer_args,
238
+ tokenizer=tokenizer,
239
+ train_dataset=train_dataset,
240
+ eval_dataset=eval_dataset,
241
+ formatting_func=formatting_prompts_func,
242
+ data_collator=collator,
243
+ dataset_batch_size=FLAGS.batch_size,
244
+ )
245
+
246
+ trainer.train()
247
+
248
+ if __name__ == '__main__':
249
+ app.run(main)
wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/conda-environment.yaml ADDED
@@ -0,0 +1,515 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ name: base
2
+ channels:
3
+ - fastai
4
+ - nvidia
5
+ - https://aws-ml-conda.s3.us-west-2.amazonaws.com
6
+ - conda-forge
7
+ dependencies:
8
+ - _libgcc_mutex=0.1=conda_forge
9
+ - _openmp_mutex=4.5=2_kmp_llvm
10
+ - alsa-lib=1.2.12=h4ab18f5_0
11
+ - annotated-types=0.7.0=pyhd8ed1ab_0
12
+ - asttokens=2.4.1=pyhd8ed1ab_0
13
+ - attr=2.5.1=h166bdaf_1
14
+ - aws-ofi-nccl=1.9.1=aws_efa1.26.1_0
15
+ - blas=1.0=mkl
16
+ - boltons=23.0.0=pyhd8ed1ab_0
17
+ - brotli-bin=1.1.0=hd590300_0
18
+ - bzip2=1.0.8=h7f98852_4
19
+ - c-ares=1.32.3=h4bc722e_0
20
+ - ca-certificates=2024.7.4=hbcca054_0
21
+ - cached-property=1.5.2=hd8ed1ab_1
22
+ - cached_property=1.5.2=pyha770c72_1
23
+ - cairo=1.18.0=h3faef2a_0
24
+ - catalogue=2.0.10=py310hff52083_0
25
+ - certifi=2024.7.4=pyhd8ed1ab_0
26
+ - cffi=1.15.1=py310h255011f_3
27
+ - charset-normalizer=3.2.0=pyhd8ed1ab_0
28
+ - click=8.1.7=unix_pyh707e725_0
29
+ - cloudpathlib=0.18.1=pyhd8ed1ab_0
30
+ - colorama=0.4.6=pyhd8ed1ab_0
31
+ - comm=0.2.2=pyhd8ed1ab_0
32
+ - conda=23.3.1=py310hff52083_0
33
+ - conda-libmamba-solver=23.3.0=pyhd8ed1ab_0
34
+ - conda-package-handling=2.2.0=pyh38be061_0
35
+ - conda-package-streaming=0.9.0=pyhd8ed1ab_0
36
+ - confection=0.1.4=py310h17c5347_0
37
+ - contourpy=1.2.1=py310hd41b1e2_0
38
+ - cuda-cccl=12.4.127=0
39
+ - cuda-compiler=12.1.1=0
40
+ - cuda-cudart=12.1.105=0
41
+ - cuda-cudart-dev=12.1.105=0
42
+ - cuda-cudart-static=12.1.105=0
43
+ - cuda-cuobjdump=12.4.127=0
44
+ - cuda-cupti=12.1.105=0
45
+ - cuda-cupti-static=12.1.105=0
46
+ - cuda-cuxxfilt=12.4.127=0
47
+ - cuda-driver-dev=12.4.127=0
48
+ - cuda-libraries=12.1.0=0
49
+ - cuda-libraries-dev=12.1.0=0
50
+ - cuda-libraries-static=12.1.1=0
51
+ - cuda-nvcc=12.1.105=0
52
+ - cuda-nvprune=12.4.127=0
53
+ - cuda-nvrtc=12.1.105=0
54
+ - cuda-nvrtc-dev=12.1.105=0
55
+ - cuda-nvrtc-static=12.1.105=0
56
+ - cuda-nvtx=12.1.105=0
57
+ - cuda-opencl=12.4.127=0
58
+ - cuda-opencl-dev=12.4.127=0
59
+ - cuda-profiler-api=12.4.127=0
60
+ - cuda-runtime=12.1.0=0
61
+ - cuda-version=12.1=h1d6eff3_3
62
+ - curl=8.8.0=he654da7_1
63
+ - cycler=0.12.1=pyhd8ed1ab_0
64
+ - cymem=2.0.8=py310hc6cd4ac_1
65
+ - cython=3.0.11=py310hea249c9_0
66
+ - cython-blis=0.7.10=py310h1f7b6fc_2
67
+ - dbus=1.13.6=h5008d03_3
68
+ - debugpy=1.8.5=py310hea249c9_0
69
+ - decorator=5.1.1=pyhd8ed1ab_0
70
+ - docutils=0.16=py310hff52083_4
71
+ - exceptiongroup=1.2.2=pyhd8ed1ab_0
72
+ - executing=2.0.1=pyhd8ed1ab_0
73
+ - expat=2.6.2=h59595ed_0
74
+ - fastai=2.7.16=py_0
75
+ - fastcore=1.5.55=py_0
76
+ - fastdownload=0.0.7=py_0
77
+ - fastprogress=1.0.3=py_0
78
+ - ffmpeg=4.2=h3fd9d12_1
79
+ - filelock=3.15.4=pyhd8ed1ab_0
80
+ - fmt=9.1.0=h924138e_0
81
+ - font-ttf-dejavu-sans-mono=2.37=hab24e00_0
82
+ - font-ttf-inconsolata=3.000=h77eed37_0
83
+ - font-ttf-source-code-pro=2.038=h77eed37_0
84
+ - font-ttf-ubuntu=0.83=h77eed37_2
85
+ - fontconfig=2.14.2=h14ed4e7_0
86
+ - fonts-conda-ecosystem=1=0
87
+ - fonts-conda-forge=1=0
88
+ - fonttools=4.53.1=py310h5b4e0ec_0
89
+ - freetype=2.12.1=h267a509_2
90
+ - gettext=0.22.5=h59595ed_2
91
+ - gettext-tools=0.22.5=h59595ed_2
92
+ - glib=2.80.2=hf974151_0
93
+ - glib-tools=2.80.2=hb6ce0ca_0
94
+ - gmp=6.3.0=hac33072_2
95
+ - gmpy2=2.1.5=py310hc7909c9_1
96
+ - gnutls=3.6.15=he1e5248_0
97
+ - graphite2=1.3.13=h59595ed_1003
98
+ - gst-plugins-base=1.22.9=hfa15dee_1
99
+ - gstreamer=1.22.9=h98fc4e7_1
100
+ - h5py=3.11.0=nompi_py310hf054cd7_102
101
+ - harfbuzz=8.5.0=hfac3d4d_0
102
+ - hdf5=1.14.3=nompi_hdf9ad27_105
103
+ - hwloc=2.9.2=h2bc3f7f_0
104
+ - icu=73.2=h59595ed_0
105
+ - idna=3.7=pyhd8ed1ab_0
106
+ - importlib_metadata=8.2.0=hd8ed1ab_0
107
+ - jedi=0.19.1=pyhd8ed1ab_0
108
+ - jinja2=3.1.4=pyhd8ed1ab_0
109
+ - jmespath=1.0.1=pyhd8ed1ab_0
110
+ - joblib=1.4.2=pyhd8ed1ab_0
111
+ - jsonpatch=1.32=pyhd8ed1ab_0
112
+ - jsonpointer=2.0=py_0
113
+ - jupyter_client=8.6.2=pyhd8ed1ab_0
114
+ - jupyter_core=5.7.2=py310hff52083_0
115
+ - keyutils=1.6.1=h166bdaf_0
116
+ - kiwisolver=1.4.5=py310hd41b1e2_1
117
+ - krb5=1.21.3=h659f571_0
118
+ - lame=3.100=h166bdaf_1003
119
+ - langcodes=3.4.0=pyhd8ed1ab_0
120
+ - language-data=1.2.0=pyhd8ed1ab_0
121
+ - lcms2=2.16=hb7c19ff_0
122
+ - ld_impl_linux-64=2.40=h41732ed_0
123
+ - lerc=4.0.0=h27087fc_0
124
+ - libaec=1.1.3=h59595ed_0
125
+ - libarchive=3.6.2=h039dbb9_1
126
+ - libasprintf=0.22.5=h661eb56_2
127
+ - libasprintf-devel=0.22.5=h661eb56_2
128
+ - libblas=3.9.0=21_linux64_mkl
129
+ - libbrotlicommon=1.1.0=hd590300_0
130
+ - libbrotlidec=1.1.0=hd590300_0
131
+ - libbrotlienc=1.1.0=hd590300_0
132
+ - libcap=2.69=h0f662aa_0
133
+ - libcblas=3.9.0=21_linux64_mkl
134
+ - libclang=15.0.7=default_h127d8a8_5
135
+ - libclang13=15.0.7=default_h5d6823c_5
136
+ - libcublas=12.1.0.26=0
137
+ - libcublas-dev=12.1.0.26=0
138
+ - libcublas-static=12.1.3.1=hd3aeb46_0
139
+ - libcufft=11.0.2.4=0
140
+ - libcufft-dev=11.0.2.4=0
141
+ - libcufft-static=11.0.2.54=hd3aeb46_0
142
+ - libcufile=1.9.1.3=0
143
+ - libcufile-dev=1.9.1.3=0
144
+ - libcufile-static=1.9.1.3=0
145
+ - libcups=2.3.3=h4637d8d_4
146
+ - libcurand=10.3.5.147=0
147
+ - libcurand-dev=10.3.5.147=0
148
+ - libcurand-static=10.3.5.147=0
149
+ - libcurl=8.8.0=hca28451_1
150
+ - libcusolver=11.4.4.55=0
151
+ - libcusolver-dev=11.4.4.55=0
152
+ - libcusolver-static=11.4.5.107=hd3aeb46_0
153
+ - libcusparse=12.0.2.55=0
154
+ - libcusparse-dev=12.0.2.55=0
155
+ - libcusparse-static=12.1.0.106=hd3aeb46_0
156
+ - libdeflate=1.20=hd590300_0
157
+ - libedit=3.1.20191231=he28a2e2_2
158
+ - libev=4.33=h516909a_1
159
+ - libevent=2.1.12=hf998b51_1
160
+ - libexpat=2.6.2=h59595ed_0
161
+ - libffi=3.4.2=h7f98852_5
162
+ - libflac=1.4.3=h59595ed_0
163
+ - libgcc=7.2.0=h69d50b8_2
164
+ - libgcc-ng=13.1.0=he5830b7_0
165
+ - libgcrypt=1.11.0=h4ab18f5_1
166
+ - libgettextpo=0.22.5=h59595ed_2
167
+ - libgettextpo-devel=0.22.5=h59595ed_2
168
+ - libgfortran-ng=13.2.0=h69a702a_0
169
+ - libgfortran5=13.2.0=ha4646dd_0
170
+ - libglib=2.80.2=hf974151_0
171
+ - libgomp=13.1.0=he5830b7_0
172
+ - libgpg-error=1.50=h4f305b6_0
173
+ - libhwloc=2.9.3=default_h554bfaf_1009
174
+ - libiconv=1.17=h166bdaf_0
175
+ - libidn2=2.3.7=hd590300_0
176
+ - libjpeg-turbo=3.0.0=hd590300_1
177
+ - liblapack=3.9.0=21_linux64_mkl
178
+ - libllvm15=15.0.7=h5cf9203_3
179
+ - libmamba=1.4.2=hcea66bb_0
180
+ - libmambapy=1.4.2=py310h1428755_0
181
+ - libnghttp2=1.58.0=h47da74e_1
182
+ - libnl=3.10.0=h4bc722e_0
183
+ - libnpp=12.0.2.50=0
184
+ - libnpp-dev=12.0.2.50=0
185
+ - libnpp-static=12.1.0.40=hd3aeb46_0
186
+ - libnsl=2.0.0=h7f98852_0
187
+ - libnvjitlink=12.1.105=0
188
+ - libnvjitlink-dev=12.1.105=0
189
+ - libnvjitlink-static=12.1.105=hd3aeb46_0
190
+ - libnvjpeg=12.1.1.14=0
191
+ - libnvjpeg-dev=12.1.1.14=0
192
+ - libnvjpeg-static=12.2.0.2=ha770c72_0
193
+ - libogg=1.3.5=h4ab18f5_0
194
+ - libopenblas=0.3.27=pthreads_hac2b453_1
195
+ - libopus=1.3.1=h7f98852_1
196
+ - libpng=1.6.43=h2797004_0
197
+ - libpq=16.4=h482b261_0
198
+ - libsndfile=1.2.2=hc60ed4a_1
199
+ - libsodium=1.0.18=h36c2ea0_1
200
+ - libsolv=0.7.29=ha6fb4c9_0
201
+ - libsqlite=3.46.0=hde9e2c9_0
202
+ - libssh2=1.11.0=h0841786_0
203
+ - libstdcxx-ng=13.1.0=hfd8a6a1_0
204
+ - libsystemd0=255=h3516f8a_1
205
+ - libtasn1=4.19.0=h166bdaf_0
206
+ - libtiff=4.6.0=h1dd3fc0_3
207
+ - libunistring=0.9.10=h7f98852_0
208
+ - libuuid=2.38.1=h0b41bf4_0
209
+ - libuv=1.48.0=hd590300_0
210
+ - libvorbis=1.3.7=h9c3ff4c_0
211
+ - libwebp-base=1.4.0=hd590300_0
212
+ - libxcb=1.15=h0b41bf4_0
213
+ - libxkbcommon=1.6.0=h5d7e998_0
214
+ - libxml2=2.11.6=h232c23b_0
215
+ - libzlib=1.2.13=h4ab18f5_6
216
+ - llvm-openmp=18.1.7=ha31de31_0
217
+ - lz4-c=1.9.4=hcb278e6_0
218
+ - lzo=2.10=h516909a_1000
219
+ - mamba=1.4.2=py310h51d5547_0
220
+ - marisa-trie=1.1.0=py310hc6cd4ac_1
221
+ - markdown-it-py=3.0.0=pyhd8ed1ab_0
222
+ - markupsafe=2.1.5=py310h2372a71_0
223
+ - matplotlib=3.9.1=py310hff52083_1
224
+ - matplotlib-base=3.9.1=py310hf02ac8c_2
225
+ - matplotlib-inline=0.1.7=pyhd8ed1ab_0
226
+ - mdurl=0.1.2=pyhd8ed1ab_0
227
+ - mkl=2024.0.0=ha957f24_49657
228
+ - mkl-include=2024.2.0=ha957f24_665
229
+ - mpc=1.3.1=hfe3b2da_0
230
+ - mpfr=4.2.1=h38ae2d0_2
231
+ - mpg123=1.32.6=h59595ed_0
232
+ - mpi=1.0=openmpi
233
+ - mpi4py=3.1.6=py310hb2ba3f8_1
234
+ - mpmath=1.3.0=pyhd8ed1ab_0
235
+ - munkres=1.1.4=pyh9f0ad1d_0
236
+ - murmurhash=1.0.10=py310hc6cd4ac_1
237
+ - mysql-common=8.0.33=hf1915f5_6
238
+ - mysql-libs=8.0.33=hca2cd23_6
239
+ - ncurses=6.5=h59595ed_0
240
+ - nest-asyncio=1.6.0=pyhd8ed1ab_0
241
+ - nettle=3.7.3=hbbd107a_1
242
+ - networkx=3.3=pyhd8ed1ab_1
243
+ - nspr=4.35=h27087fc_0
244
+ - nss=3.100=hca3bf56_0
245
+ - openh264=2.1.1=h780b84a_0
246
+ - openjpeg=2.5.2=h488ebb8_0
247
+ - openmpi=5.0.3=h817cd4e_101
248
+ - openssl=3.3.1=h4bc722e_2
249
+ - packaging=23.1=pyhd8ed1ab_0
250
+ - pandas=2.2.2=py310hf9f9076_1
251
+ - parso=0.8.4=pyhd8ed1ab_0
252
+ - pcre2=10.43=hcad00b1_0
253
+ - pexpect=4.9.0=pyhd8ed1ab_0
254
+ - pickleshare=0.7.5=py_1003
255
+ - pillow=10.3.0=py310hf73ecf8_0
256
+ - pixman=0.43.2=h59595ed_0
257
+ - platformdirs=4.2.2=pyhd8ed1ab_0
258
+ - pluggy=1.3.0=pyhd8ed1ab_0
259
+ - ply=3.11=pyhd8ed1ab_2
260
+ - preshed=3.0.9=py310hc6cd4ac_1
261
+ - prompt-toolkit=3.0.47=pyha770c72_0
262
+ - psutil=6.0.0=py310hc51659f_0
263
+ - pthread-stubs=0.4=h36c2ea0_1001
264
+ - ptyprocess=0.7.0=pyhd3deb0d_0
265
+ - pulseaudio-client=16.1=hb77b528_5
266
+ - pure_eval=0.2.3=pyhd8ed1ab_0
267
+ - pyasn1=0.6.0=pyhd8ed1ab_0
268
+ - pybind11=2.13.1=py310h25c7140_0
269
+ - pybind11-abi=4=hd8ed1ab_3
270
+ - pybind11-global=2.13.1=py310h25c7140_0
271
+ - pycosat=0.6.4=py310h5764c6d_1
272
+ - pycparser=2.21=pyhd8ed1ab_0
273
+ - pydantic=2.8.2=pyhd8ed1ab_0
274
+ - pydantic-core=2.20.1=py310h42e942d_0
275
+ - pygments=2.18.0=pyhd8ed1ab_0
276
+ - pyparsing=3.1.2=pyhd8ed1ab_0
277
+ - pyqt=5.15.9=py310h04931ad_5
278
+ - pyqt5-sip=12.12.2=py310hc6cd4ac_5
279
+ - pysocks=1.7.1=pyha2e5f31_6
280
+ - python=3.10.12=hd12c33a_0_cpython
281
+ - python-dateutil=2.9.0=pyhd8ed1ab_0
282
+ - python-tzdata=2024.1=pyhd8ed1ab_0
283
+ - python_abi=3.10=3_cp310
284
+ - pytorch-cuda=12.1=ha16c6d3_5
285
+ - pytorch-mutex=1.0=cuda
286
+ - pytz=2024.1=pyhd8ed1ab_0
287
+ - pyyaml=6.0.2=py310h5b4e0ec_0
288
+ - pyzmq=26.1.0=py310h7d2b5bf_0
289
+ - qhull=2020.2=h434a139_5
290
+ - qt-main=5.15.8=h82b777d_17
291
+ - readline=8.2=h8228510_1
292
+ - reproc=14.2.4=h0b41bf4_0
293
+ - reproc-cpp=14.2.4=hcb278e6_0
294
+ - requests=2.32.3=pyhd8ed1ab_0
295
+ - rhash=1.4.4=hd590300_0
296
+ - rich=13.7.1=pyhd8ed1ab_0
297
+ - rsa=4.7.2=pyh44b312d_0
298
+ - ruamel.yaml=0.17.32=py310h2372a71_0
299
+ - ruamel.yaml.clib=0.2.7=py310h1fa729e_1
300
+ - s3transfer=0.10.2=pyhd8ed1ab_0
301
+ - scipy=1.14.0=py310h93e2701_1
302
+ - shellingham=1.5.4=pyhd8ed1ab_0
303
+ - sip=6.7.12=py310hc6cd4ac_0
304
+ - six=1.16.0=pyh6c4a22f_0
305
+ - smart-open=7.0.4=hd8ed1ab_0
306
+ - smart_open=7.0.4=pyhd8ed1ab_0
307
+ - spacy=3.7.5=py310h9fd543b_1
308
+ - spacy-legacy=3.0.12=pyhd8ed1ab_0
309
+ - spacy-loggers=1.0.5=pyhd8ed1ab_0
310
+ - srsly=2.4.8=py310hc6cd4ac_1
311
+ - stack_data=0.6.2=pyhd8ed1ab_0
312
+ - sympy=1.13.0=pypyh2585a3b_103
313
+ - tbb=2021.8.0=hdb19cb5_0
314
+ - thinc=8.2.5=py310h9fd543b_0
315
+ - threadpoolctl=3.5.0=pyhc1e730c_0
316
+ - tk=8.6.13=noxft_h4845f30_101
317
+ - toml=0.10.2=pyhd8ed1ab_0
318
+ - tomli=2.0.1=pyhd8ed1ab_0
319
+ - toolz=0.12.0=pyhd8ed1ab_0
320
+ - torchdata=0.7.0=py310
321
+ - torchtext=0.16.0=py310
322
+ - tornado=6.4.1=py310hc51659f_0
323
+ - tqdm=4.66.5=pyhd8ed1ab_0
324
+ - traitlets=5.14.3=pyhd8ed1ab_0
325
+ - typer=0.12.3=pyhd8ed1ab_0
326
+ - typer-slim=0.12.3=pyhd8ed1ab_0
327
+ - typer-slim-standard=0.12.3=hd8ed1ab_0
328
+ - typing=3.10.0.0=pyhd8ed1ab_1
329
+ - typing-extensions=4.12.2=hd8ed1ab_0
330
+ - typing_extensions=4.12.2=pyha770c72_0
331
+ - tzdata=2023c=h71feb2d_0
332
+ - unicodedata2=15.1.0=py310h2372a71_0
333
+ - urllib3=1.26.19=pyhd8ed1ab_0
334
+ - wasabi=1.1.2=py310hff52083_1
335
+ - wcwidth=0.2.13=pyhd8ed1ab_0
336
+ - weasel=0.4.1=pyhd8ed1ab_1
337
+ - wheel=0.41.2=pyhd8ed1ab_0
338
+ - wrapt=1.16.0=py310h2372a71_0
339
+ - xcb-util=0.4.0=hd590300_1
340
+ - xcb-util-image=0.4.0=h8ee46fc_1
341
+ - xcb-util-keysyms=0.4.0=h8ee46fc_1
342
+ - xcb-util-renderutil=0.3.9=hd590300_1
343
+ - xcb-util-wm=0.4.1=h8ee46fc_1
344
+ - xkeyboard-config=2.42=h4ab18f5_0
345
+ - xorg-kbproto=1.0.7=h7f98852_1002
346
+ - xorg-libice=1.1.1=hd590300_0
347
+ - xorg-libsm=1.2.4=h7391055_0
348
+ - xorg-libx11=1.8.9=h8ee46fc_0
349
+ - xorg-libxau=1.0.11=hd590300_0
350
+ - xorg-libxdmcp=1.1.3=h7f98852_0
351
+ - xorg-libxext=1.3.4=h0b41bf4_2
352
+ - xorg-libxrender=0.9.11=hd590300_0
353
+ - xorg-renderproto=0.11.1=h7f98852_1002
354
+ - xorg-xextproto=7.3.0=h0b41bf4_1003
355
+ - xorg-xf86vidmodeproto=2.3.1=h7f98852_1002
356
+ - xorg-xproto=7.0.31=h7f98852_1007
357
+ - xz=5.2.6=h166bdaf_0
358
+ - yaml=0.2.5=h7f98852_2
359
+ - yaml-cpp=0.7.0=h27087fc_2
360
+ - zeromq=4.3.5=h75354e8_4
361
+ - zipp=3.19.2=pyhd8ed1ab_0
362
+ - zlib=1.2.13=h4ab18f5_6
363
+ - zstandard=0.19.0=py310h5764c6d_0
364
+ - zstd=1.5.6=ha6fb4c9_0
365
+ - pip:
366
+ - absl-py==2.1.0
367
+ - accelerate==1.1.1
368
+ - aiohappyeyeballs==2.4.3
369
+ - aiohttp==3.10.10
370
+ - aiosignal==1.3.1
371
+ - antlr4-python3-runtime==4.9.3
372
+ - apex==0.1
373
+ - appdirs==1.4.4
374
+ - argparse==1.4.0
375
+ - async-timeout==4.0.3
376
+ - attrdict==2.0.1
377
+ - attrs==23.2.0
378
+ - awscli==1.33.40
379
+ - backcall==0.2.0
380
+ - bcrypt==4.2.0
381
+ - beautifulsoup4==4.12.3
382
+ - bleach==6.2.0
383
+ - bokeh==3.5.1
384
+ - boto3==1.34.158
385
+ - botocore==1.34.158
386
+ - brotli==1.0.9
387
+ - cachetools==5.5.0
388
+ - chardet==5.2.0
389
+ - cloudpickle==2.2.1
390
+ - cmake==3.30.2
391
+ - cryptography==41.0.7
392
+ - datasets==3.1.0
393
+ - defusedxml==0.7.1
394
+ - dill==0.3.8
395
+ - docker==7.1.0
396
+ - docker-pycreds==0.4.0
397
+ - docopt==0.6.2
398
+ - docstring-parser==0.16
399
+ - einops==0.8.0
400
+ - fastjsonschema==2.20.0
401
+ - fire==0.6.0
402
+ - flash-attn==2.5.5
403
+ - frozenlist==1.5.0
404
+ - fsspec==2024.6.1
405
+ - gevent==24.2.1
406
+ - gitdb==4.0.11
407
+ - gitpython==3.1.43
408
+ - google-pasta==0.2.0
409
+ - greenlet==3.0.3
410
+ - huggingface-hub==0.26.2
411
+ - hydra-core==1.3.2
412
+ - imageio==2.34.2
413
+ - importlib-metadata==6.11.0
414
+ - importlib-resources==6.1.1
415
+ - inotify-simple==1.2.1
416
+ - ipykernel==6.29.1
417
+ - ipython==8.12.3
418
+ - jsonschema==4.23.0
419
+ - jsonschema-specifications==2023.12.1
420
+ - jupyterlab-pygments==0.3.0
421
+ - llvmlite==0.43.0
422
+ - mistune==3.0.2
423
+ - multidict==6.1.0
424
+ - multiprocess==0.70.16
425
+ - nbclient==0.10.0
426
+ - nbconvert==7.16.4
427
+ - nbformat==5.10.4
428
+ - ninja==1.11.1.1
429
+ - numba==0.60.0
430
+ - numpy==1.24.3
431
+ - nvidia-cublas-cu12==12.1.3.1
432
+ - nvidia-cuda-cupti-cu12==12.1.105
433
+ - nvidia-cuda-nvrtc-cu12==12.1.105
434
+ - nvidia-cuda-runtime-cu12==12.1.105
435
+ - nvidia-cudnn-cu12==8.9.2.26
436
+ - nvidia-cufft-cu12==11.0.2.54
437
+ - nvidia-curand-cu12==10.3.2.106
438
+ - nvidia-cusolver-cu12==11.4.5.107
439
+ - nvidia-cusparse-cu12==12.1.0.106
440
+ - nvidia-ml-py==12.535.161
441
+ - nvidia-nccl-cu12==2.19.3
442
+ - nvidia-nvjitlink-cu12==12.6.77
443
+ - nvidia-nvtx-cu12==12.1.105
444
+ - nvitop==1.3.2
445
+ - omegaconf==2.3.0
446
+ - opencv-python==4.10.0.84
447
+ - pandocfilters==1.5.1
448
+ - paramiko==3.4.0
449
+ - pathos==0.3.2
450
+ - pip==24.1.2
451
+ - pip-chill==1.0.3
452
+ - pipreqs==0.5.0
453
+ - plotly==5.23.0
454
+ - pox==0.3.4
455
+ - ppft==1.7.6.8
456
+ - propcache==0.2.0
457
+ - protobuf==3.20.3
458
+ - pyarrow==17.0.0
459
+ - pyfunctional==1.5.0
460
+ - pynacl==1.5.0
461
+ - pyopenssl==23.2.0
462
+ - referencing==0.35.1
463
+ - regex==2024.11.6
464
+ - retrying==1.3.4
465
+ - rpds-py==0.20.0
466
+ - s3fs==0.4.2
467
+ - safetensors==0.4.5
468
+ - sagemaker==2.224.2
469
+ - sagemaker-experiments==0.1.45
470
+ - sagemaker-pytorch-training==2.8.1
471
+ - sagemaker-training==4.8.1
472
+ - schema==0.7.7
473
+ - scikit-learn==1.3.2
474
+ - seaborn==0.13.2
475
+ - sentencepiece==0.2.0
476
+ - sentry-sdk==2.18.0
477
+ - setproctitle==1.3.3
478
+ - setuptools==72.1.0
479
+ - shap==0.46.0
480
+ - shtab==1.7.1
481
+ - slicer==0.0.8
482
+ - smclarify==0.5
483
+ - smdebug-rulesconfig==1.0.1
484
+ - smdistributed-dataparallel==2.1.0
485
+ - smmap==5.0.1
486
+ - smprof==0.3.334
487
+ - soupsieve==2.6
488
+ - tabulate==0.9.0
489
+ - tblib==3.0.0
490
+ - tenacity==9.0.0
491
+ - tensor-parallel==2.0.0
492
+ - termcolor==2.5.0
493
+ - tinycss2==1.4.0
494
+ - tokenizers==0.20.3
495
+ - torch==2.2.0
496
+ - torchaudio==2.2.0
497
+ - torchnet==0.0.4
498
+ - torchvision==0.17.0
499
+ - transformer-engine==0.12.0+170797
500
+ - transformers==4.46.0
501
+ - triton==2.2.0
502
+ - trl==0.12.0
503
+ - tyro==0.8.5
504
+ - visdom==0.2.4
505
+ - wandb==0.16.3
506
+ - webencodings==0.5.1
507
+ - websocket-client==1.8.0
508
+ - werkzeug==3.0.3
509
+ - xxhash==3.5.0
510
+ - xyzservices==2024.6.0
511
+ - yarg==0.1.9
512
+ - yarl==1.17.1
513
+ - zope-event==5.0
514
+ - zope-interface==7.0.1
515
+ prefix: /opt/conda
wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/config.yaml ADDED
@@ -0,0 +1,948 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wandb_version: 1
2
+
3
+ sagemaker_training_job_name:
4
+ desc: null
5
+ value: kushalarora-rvv-main-2024-11-13-17-55-42-664
6
+ batch_size:
7
+ desc: null
8
+ value: 64
9
+ cache_dir:
10
+ desc: null
11
+ value: /opt/ml/data/input/.cache
12
+ dataset:
13
+ desc: null
14
+ value: code-contests
15
+ gradient_accumulation_steps:
16
+ desc: null
17
+ value: 8
18
+ logging_steps:
19
+ desc: null
20
+ value: 5
21
+ lr:
22
+ desc: null
23
+ value: 1e-6
24
+ max_seq_length:
25
+ desc: null
26
+ value: 2048
27
+ model_name:
28
+ desc: null
29
+ value: Qwen/Qwen2.5-Coder-7B-Instruct
30
+ num_train_epochs:
31
+ desc: null
32
+ value: 2
33
+ output_dir:
34
+ desc: null
35
+ value: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048
36
+ push_to_hub:
37
+ desc: null
38
+ value: true
39
+ run_name:
40
+ desc: null
41
+ value: sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048
42
+ sagemaker_container_log_level:
43
+ desc: null
44
+ value: 20
45
+ sagemaker_instance_type:
46
+ desc: null
47
+ value: ml.p5.48xlarge
48
+ sagemaker_job_name:
49
+ desc: null
50
+ value: kushalarora-rvv-main-2024-11-13-17-55-42-664
51
+ sagemaker_program:
52
+ desc: null
53
+ value: sft.py
54
+ sagemaker_region:
55
+ desc: null
56
+ value: us-east-1
57
+ sagemaker_submit_directory:
58
+ desc: null
59
+ value: s3://tri-ml-datasets/kushalarora-rvv-main-2024-11-13-17-55-42-664/source/sourcedir.tar.gz
60
+ sagemaker_torch_distributed_enabled:
61
+ desc: null
62
+ value: 'true'
63
+ save_steps:
64
+ desc: null
65
+ value: 100
66
+ schedule:
67
+ desc: null
68
+ value: cosine
69
+ wandb_project:
70
+ desc: null
71
+ value: sft-codecontests-1112
72
+ weight_decay:
73
+ desc: null
74
+ value: 0.0
75
+ additional_framework_parameters:
76
+ desc: null
77
+ value:
78
+ sagemaker_instance_type: ml.p5.48xlarge
79
+ sagemaker_torch_distributed_enabled: true
80
+ channel_input_dirs:
81
+ desc: null
82
+ value: {}
83
+ current_host:
84
+ desc: null
85
+ value: algo-1
86
+ current_instance_group:
87
+ desc: null
88
+ value: homogeneousCluster
89
+ current_instance_group_hosts:
90
+ desc: null
91
+ value:
92
+ - algo-1
93
+ current_instance_type:
94
+ desc: null
95
+ value: ml.p5.48xlarge
96
+ distribution_hosts:
97
+ desc: null
98
+ value:
99
+ - algo-1
100
+ distribution_instance_groups:
101
+ desc: null
102
+ value:
103
+ - homogeneousCluster
104
+ framework_module:
105
+ desc: null
106
+ value: sagemaker_pytorch_container.training:main
107
+ hosts:
108
+ desc: null
109
+ value:
110
+ - algo-1
111
+ hyperparameters:
112
+ desc: null
113
+ value:
114
+ batch_size: 64
115
+ cache_dir: /opt/ml/data/input/.cache
116
+ dataset: code-contests
117
+ gradient_accumulation_steps: 8
118
+ logging_steps: 5
119
+ lr: 1e-6
120
+ max_seq_length: 2048
121
+ model_name: Qwen/Qwen2.5-Coder-7B-Instruct
122
+ num_train_epochs: 2
123
+ output_dir: /opt/ml/model/
124
+ push_to_hub: true
125
+ run_name: sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048
126
+ save_steps: 100
127
+ schedule: cosine
128
+ wandb_project: sft-codecontests-1112
129
+ weight_decay: 0.0
130
+ input_config_dir:
131
+ desc: null
132
+ value: /opt/ml/input/config
133
+ input_data_config:
134
+ desc: null
135
+ value: {}
136
+ input_dir:
137
+ desc: null
138
+ value: /opt/ml/input
139
+ instance_groups:
140
+ desc: null
141
+ value:
142
+ - homogeneousCluster
143
+ instance_groups_dict:
144
+ desc: null
145
+ value:
146
+ homogeneousCluster:
147
+ hosts:
148
+ - algo-1
149
+ instance_group_name: homogeneousCluster
150
+ instance_type: ml.p5.48xlarge
151
+ is_hetero:
152
+ desc: null
153
+ value: false
154
+ is_master:
155
+ desc: null
156
+ value: true
157
+ is_modelparallel_enabled:
158
+ desc: null
159
+ value: null
160
+ is_smddpmprun_installed:
161
+ desc: null
162
+ value: false
163
+ is_smddprun_installed:
164
+ desc: null
165
+ value: true
166
+ job_name:
167
+ desc: null
168
+ value: kushalarora-rvv-main-2024-11-13-17-55-42-664
169
+ log_level:
170
+ desc: null
171
+ value: passive
172
+ master_hostname:
173
+ desc: null
174
+ value: algo-1
175
+ model_dir:
176
+ desc: null
177
+ value: /opt/ml/model
178
+ module_dir:
179
+ desc: null
180
+ value: s3://tri-ml-datasets/kushalarora-rvv-main-2024-11-13-17-55-42-664/source/sourcedir.tar.gz
181
+ module_name:
182
+ desc: null
183
+ value: /opt/ml/code/qlearning_reasoning/training/sft
184
+ network_interface_name:
185
+ desc: null
186
+ value: eth0
187
+ num_cpus:
188
+ desc: null
189
+ value: 192
190
+ num_gpus:
191
+ desc: null
192
+ value: 8
193
+ num_neurons:
194
+ desc: null
195
+ value: 0
196
+ output_data_dir:
197
+ desc: null
198
+ value: /opt/ml/output/data
199
+ output_intermediate_dir:
200
+ desc: null
201
+ value: /opt/ml/output/intermediate
202
+ resource_config:
203
+ desc: null
204
+ value:
205
+ current_group_name: homogeneousCluster
206
+ current_host: algo-1
207
+ current_instance_type: ml.p5.48xlarge
208
+ hosts:
209
+ - algo-1
210
+ instance_groups:
211
+ - hosts:
212
+ - algo-1
213
+ instance_group_name: homogeneousCluster
214
+ instance_type: ml.p5.48xlarge
215
+ network_interface_name: eth0
216
+ user_entry_point:
217
+ desc: null
218
+ value: /opt/ml/code/qlearning_reasoning/training/sft.py
219
+ _wandb:
220
+ desc: null
221
+ value:
222
+ code_path: code/qlearning_reasoning/training/sft.py
223
+ python_version: 3.10.12
224
+ cli_version: 0.16.3
225
+ framework: huggingface
226
+ huggingface_version: 4.46.0
227
+ is_jupyter_run: false
228
+ is_kaggle_kernel: false
229
+ start_time: 1731524566.431221
230
+ t:
231
+ 1:
232
+ - 1
233
+ - 5
234
+ - 11
235
+ - 41
236
+ - 49
237
+ - 51
238
+ - 53
239
+ - 55
240
+ - 71
241
+ - 84
242
+ 2:
243
+ - 1
244
+ - 5
245
+ - 11
246
+ - 41
247
+ - 49
248
+ - 51
249
+ - 53
250
+ - 55
251
+ - 71
252
+ - 84
253
+ 3:
254
+ - 7
255
+ - 9
256
+ - 13
257
+ - 19
258
+ - 23
259
+ 4: 3.10.12
260
+ 5: 0.16.3
261
+ 6: 4.46.0
262
+ 8:
263
+ - 5
264
+ 9:
265
+ 1: transformers_trainer
266
+ 13: linux-x86_64
267
+ m:
268
+ - 1: train/global_step
269
+ 6:
270
+ - 3
271
+ - 1: train/loss
272
+ 5: 1
273
+ 6:
274
+ - 1
275
+ - 1: train/grad_norm
276
+ 5: 1
277
+ 6:
278
+ - 1
279
+ - 1: train/learning_rate
280
+ 5: 1
281
+ 6:
282
+ - 1
283
+ - 1: train/epoch
284
+ 5: 1
285
+ 6:
286
+ - 1
287
+ - 1: eval/loss
288
+ 5: 1
289
+ 6:
290
+ - 1
291
+ - 1: eval/runtime
292
+ 5: 1
293
+ 6:
294
+ - 1
295
+ - 1: eval/samples_per_second
296
+ 5: 1
297
+ 6:
298
+ - 1
299
+ - 1: eval/steps_per_second
300
+ 5: 1
301
+ 6:
302
+ - 1
303
+ vocab_size:
304
+ desc: null
305
+ value: 151671
306
+ max_position_embeddings:
307
+ desc: null
308
+ value: 32768
309
+ hidden_size:
310
+ desc: null
311
+ value: 3584
312
+ intermediate_size:
313
+ desc: null
314
+ value: 18944
315
+ num_hidden_layers:
316
+ desc: null
317
+ value: 28
318
+ num_attention_heads:
319
+ desc: null
320
+ value: 28
321
+ use_sliding_window:
322
+ desc: null
323
+ value: false
324
+ sliding_window:
325
+ desc: null
326
+ value: null
327
+ max_window_layers:
328
+ desc: null
329
+ value: 28
330
+ num_key_value_heads:
331
+ desc: null
332
+ value: 4
333
+ hidden_act:
334
+ desc: null
335
+ value: silu
336
+ initializer_range:
337
+ desc: null
338
+ value: 0.02
339
+ rms_norm_eps:
340
+ desc: null
341
+ value: 1.0e-06
342
+ use_cache:
343
+ desc: null
344
+ value: false
345
+ rope_theta:
346
+ desc: null
347
+ value: 1000000.0
348
+ rope_scaling:
349
+ desc: null
350
+ value: null
351
+ attention_dropout:
352
+ desc: null
353
+ value: 0.0
354
+ return_dict:
355
+ desc: null
356
+ value: true
357
+ output_hidden_states:
358
+ desc: null
359
+ value: false
360
+ output_attentions:
361
+ desc: null
362
+ value: false
363
+ torchscript:
364
+ desc: null
365
+ value: false
366
+ torch_dtype:
367
+ desc: null
368
+ value: bfloat16
369
+ use_bfloat16:
370
+ desc: null
371
+ value: false
372
+ tf_legacy_loss:
373
+ desc: null
374
+ value: false
375
+ pruned_heads:
376
+ desc: null
377
+ value: {}
378
+ tie_word_embeddings:
379
+ desc: null
380
+ value: false
381
+ chunk_size_feed_forward:
382
+ desc: null
383
+ value: 0
384
+ is_encoder_decoder:
385
+ desc: null
386
+ value: false
387
+ is_decoder:
388
+ desc: null
389
+ value: false
390
+ cross_attention_hidden_size:
391
+ desc: null
392
+ value: null
393
+ add_cross_attention:
394
+ desc: null
395
+ value: false
396
+ tie_encoder_decoder:
397
+ desc: null
398
+ value: false
399
+ max_length:
400
+ desc: null
401
+ value: 20
402
+ min_length:
403
+ desc: null
404
+ value: 0
405
+ do_sample:
406
+ desc: null
407
+ value: false
408
+ early_stopping:
409
+ desc: null
410
+ value: false
411
+ num_beams:
412
+ desc: null
413
+ value: 1
414
+ num_beam_groups:
415
+ desc: null
416
+ value: 1
417
+ diversity_penalty:
418
+ desc: null
419
+ value: 0.0
420
+ temperature:
421
+ desc: null
422
+ value: 1.0
423
+ top_k:
424
+ desc: null
425
+ value: 50
426
+ top_p:
427
+ desc: null
428
+ value: 1.0
429
+ typical_p:
430
+ desc: null
431
+ value: 1.0
432
+ repetition_penalty:
433
+ desc: null
434
+ value: 1.0
435
+ length_penalty:
436
+ desc: null
437
+ value: 1.0
438
+ no_repeat_ngram_size:
439
+ desc: null
440
+ value: 0
441
+ encoder_no_repeat_ngram_size:
442
+ desc: null
443
+ value: 0
444
+ bad_words_ids:
445
+ desc: null
446
+ value: null
447
+ num_return_sequences:
448
+ desc: null
449
+ value: 1
450
+ output_scores:
451
+ desc: null
452
+ value: false
453
+ return_dict_in_generate:
454
+ desc: null
455
+ value: false
456
+ forced_bos_token_id:
457
+ desc: null
458
+ value: null
459
+ forced_eos_token_id:
460
+ desc: null
461
+ value: null
462
+ remove_invalid_values:
463
+ desc: null
464
+ value: false
465
+ exponential_decay_length_penalty:
466
+ desc: null
467
+ value: null
468
+ suppress_tokens:
469
+ desc: null
470
+ value: null
471
+ begin_suppress_tokens:
472
+ desc: null
473
+ value: null
474
+ architectures:
475
+ desc: null
476
+ value:
477
+ - Qwen2ForCausalLM
478
+ finetuning_task:
479
+ desc: null
480
+ value: null
481
+ id2label:
482
+ desc: null
483
+ value:
484
+ '0': LABEL_0
485
+ '1': LABEL_1
486
+ label2id:
487
+ desc: null
488
+ value:
489
+ LABEL_0: 0
490
+ LABEL_1: 1
491
+ tokenizer_class:
492
+ desc: null
493
+ value: null
494
+ prefix:
495
+ desc: null
496
+ value: null
497
+ bos_token_id:
498
+ desc: null
499
+ value: null
500
+ pad_token_id:
501
+ desc: null
502
+ value: 151643
503
+ eos_token_id:
504
+ desc: null
505
+ value: 151645
506
+ sep_token_id:
507
+ desc: null
508
+ value: null
509
+ decoder_start_token_id:
510
+ desc: null
511
+ value: null
512
+ task_specific_params:
513
+ desc: null
514
+ value: null
515
+ problem_type:
516
+ desc: null
517
+ value: null
518
+ _name_or_path:
519
+ desc: null
520
+ value: Qwen/Qwen2.5-Coder-7B-Instruct
521
+ _attn_implementation_autoset:
522
+ desc: null
523
+ value: true
524
+ transformers_version:
525
+ desc: null
526
+ value: 4.46.0
527
+ model_type:
528
+ desc: null
529
+ value: qwen2
530
+ overwrite_output_dir:
531
+ desc: null
532
+ value: false
533
+ do_train:
534
+ desc: null
535
+ value: false
536
+ do_eval:
537
+ desc: null
538
+ value: true
539
+ do_predict:
540
+ desc: null
541
+ value: false
542
+ eval_strategy:
543
+ desc: null
544
+ value: steps
545
+ prediction_loss_only:
546
+ desc: null
547
+ value: false
548
+ per_device_train_batch_size:
549
+ desc: null
550
+ value: 8
551
+ per_device_eval_batch_size:
552
+ desc: null
553
+ value: 8
554
+ per_gpu_train_batch_size:
555
+ desc: null
556
+ value: null
557
+ per_gpu_eval_batch_size:
558
+ desc: null
559
+ value: null
560
+ eval_accumulation_steps:
561
+ desc: null
562
+ value: null
563
+ eval_delay:
564
+ desc: null
565
+ value: 0
566
+ torch_empty_cache_steps:
567
+ desc: null
568
+ value: null
569
+ learning_rate:
570
+ desc: null
571
+ value: 1.0e-06
572
+ adam_beta1:
573
+ desc: null
574
+ value: 0.9
575
+ adam_beta2:
576
+ desc: null
577
+ value: 0.999
578
+ adam_epsilon:
579
+ desc: null
580
+ value: 1.0e-08
581
+ max_grad_norm:
582
+ desc: null
583
+ value: 1.0
584
+ max_steps:
585
+ desc: null
586
+ value: -1
587
+ lr_scheduler_type:
588
+ desc: null
589
+ value: cosine
590
+ lr_scheduler_kwargs:
591
+ desc: null
592
+ value: {}
593
+ warmup_ratio:
594
+ desc: null
595
+ value: 0.0
596
+ warmup_steps:
597
+ desc: null
598
+ value: 0
599
+ log_level_replica:
600
+ desc: null
601
+ value: warning
602
+ log_on_each_node:
603
+ desc: null
604
+ value: true
605
+ logging_dir:
606
+ desc: null
607
+ value: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/runs/Nov13_19-02-15_algo-1
608
+ logging_strategy:
609
+ desc: null
610
+ value: steps
611
+ logging_first_step:
612
+ desc: null
613
+ value: true
614
+ logging_nan_inf_filter:
615
+ desc: null
616
+ value: true
617
+ save_strategy:
618
+ desc: null
619
+ value: epoch
620
+ save_total_limit:
621
+ desc: null
622
+ value: null
623
+ save_safetensors:
624
+ desc: null
625
+ value: true
626
+ save_on_each_node:
627
+ desc: null
628
+ value: false
629
+ save_only_model:
630
+ desc: null
631
+ value: false
632
+ restore_callback_states_from_checkpoint:
633
+ desc: null
634
+ value: false
635
+ no_cuda:
636
+ desc: null
637
+ value: false
638
+ use_cpu:
639
+ desc: null
640
+ value: false
641
+ use_mps_device:
642
+ desc: null
643
+ value: false
644
+ seed:
645
+ desc: null
646
+ value: 42
647
+ data_seed:
648
+ desc: null
649
+ value: null
650
+ jit_mode_eval:
651
+ desc: null
652
+ value: false
653
+ use_ipex:
654
+ desc: null
655
+ value: false
656
+ bf16:
657
+ desc: null
658
+ value: false
659
+ fp16:
660
+ desc: null
661
+ value: false
662
+ fp16_opt_level:
663
+ desc: null
664
+ value: O1
665
+ half_precision_backend:
666
+ desc: null
667
+ value: auto
668
+ bf16_full_eval:
669
+ desc: null
670
+ value: false
671
+ fp16_full_eval:
672
+ desc: null
673
+ value: false
674
+ tf32:
675
+ desc: null
676
+ value: null
677
+ local_rank:
678
+ desc: null
679
+ value: 0
680
+ ddp_backend:
681
+ desc: null
682
+ value: null
683
+ tpu_num_cores:
684
+ desc: null
685
+ value: null
686
+ tpu_metrics_debug:
687
+ desc: null
688
+ value: false
689
+ debug:
690
+ desc: null
691
+ value: []
692
+ dataloader_drop_last:
693
+ desc: null
694
+ value: false
695
+ eval_steps:
696
+ desc: null
697
+ value: 5
698
+ dataloader_num_workers:
699
+ desc: null
700
+ value: 0
701
+ dataloader_prefetch_factor:
702
+ desc: null
703
+ value: null
704
+ past_index:
705
+ desc: null
706
+ value: -1
707
+ disable_tqdm:
708
+ desc: null
709
+ value: false
710
+ remove_unused_columns:
711
+ desc: null
712
+ value: true
713
+ label_names:
714
+ desc: null
715
+ value: null
716
+ load_best_model_at_end:
717
+ desc: null
718
+ value: false
719
+ metric_for_best_model:
720
+ desc: null
721
+ value: null
722
+ greater_is_better:
723
+ desc: null
724
+ value: null
725
+ ignore_data_skip:
726
+ desc: null
727
+ value: false
728
+ fsdp:
729
+ desc: null
730
+ value:
731
+ - full_shard
732
+ - auto_wrap
733
+ fsdp_min_num_params:
734
+ desc: null
735
+ value: 0
736
+ fsdp_config:
737
+ desc: null
738
+ value:
739
+ activation_checkpointing: false
740
+ auto_wrap_policy: TRANSFORMER_BASED_WRAP
741
+ backward_prefetch: BACKWARD_PRE
742
+ cpu_ram_efficient_loading: true
743
+ forward_prefetch: false
744
+ offload_params: false
745
+ sharding_strategy: FULL_SHARD
746
+ state_dict_type: FULL_STATE_DICT
747
+ sync_module_states: true
748
+ min_num_params: 0
749
+ xla: false
750
+ xla_fsdp_v2: false
751
+ xla_fsdp_grad_ckpt: false
752
+ fsdp_transformer_layer_cls_to_wrap:
753
+ desc: null
754
+ value: null
755
+ accelerator_config:
756
+ desc: null
757
+ value:
758
+ split_batches: false
759
+ dispatch_batches: null
760
+ even_batches: true
761
+ use_seedable_sampler: true
762
+ non_blocking: false
763
+ gradient_accumulation_kwargs: null
764
+ deepspeed:
765
+ desc: null
766
+ value: null
767
+ label_smoothing_factor:
768
+ desc: null
769
+ value: 0.0
770
+ optim:
771
+ desc: null
772
+ value: adamw_torch
773
+ optim_args:
774
+ desc: null
775
+ value: null
776
+ adafactor:
777
+ desc: null
778
+ value: false
779
+ group_by_length:
780
+ desc: null
781
+ value: false
782
+ length_column_name:
783
+ desc: null
784
+ value: length
785
+ report_to:
786
+ desc: null
787
+ value:
788
+ - wandb
789
+ ddp_find_unused_parameters:
790
+ desc: null
791
+ value: null
792
+ ddp_bucket_cap_mb:
793
+ desc: null
794
+ value: null
795
+ ddp_broadcast_buffers:
796
+ desc: null
797
+ value: null
798
+ dataloader_pin_memory:
799
+ desc: null
800
+ value: true
801
+ dataloader_persistent_workers:
802
+ desc: null
803
+ value: false
804
+ skip_memory_metrics:
805
+ desc: null
806
+ value: true
807
+ use_legacy_prediction_loop:
808
+ desc: null
809
+ value: false
810
+ resume_from_checkpoint:
811
+ desc: null
812
+ value: null
813
+ hub_model_id:
814
+ desc: null
815
+ value: sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_
816
+ hub_strategy:
817
+ desc: null
818
+ value: every_save
819
+ hub_token:
820
+ desc: null
821
+ value: <HUB_TOKEN>
822
+ hub_private_repo:
823
+ desc: null
824
+ value: false
825
+ hub_always_push:
826
+ desc: null
827
+ value: false
828
+ gradient_checkpointing:
829
+ desc: null
830
+ value: false
831
+ gradient_checkpointing_kwargs:
832
+ desc: null
833
+ value:
834
+ use_reentrant: false
835
+ include_inputs_for_metrics:
836
+ desc: null
837
+ value: false
838
+ include_for_metrics:
839
+ desc: null
840
+ value: []
841
+ eval_do_concat_batches:
842
+ desc: null
843
+ value: true
844
+ fp16_backend:
845
+ desc: null
846
+ value: auto
847
+ evaluation_strategy:
848
+ desc: null
849
+ value: null
850
+ push_to_hub_model_id:
851
+ desc: null
852
+ value: null
853
+ push_to_hub_organization:
854
+ desc: null
855
+ value: null
856
+ push_to_hub_token:
857
+ desc: null
858
+ value: <PUSH_TO_HUB_TOKEN>
859
+ mp_parameters:
860
+ desc: null
861
+ value: ''
862
+ auto_find_batch_size:
863
+ desc: null
864
+ value: false
865
+ full_determinism:
866
+ desc: null
867
+ value: false
868
+ torchdynamo:
869
+ desc: null
870
+ value: null
871
+ ray_scope:
872
+ desc: null
873
+ value: last
874
+ ddp_timeout:
875
+ desc: null
876
+ value: 1800
877
+ torch_compile:
878
+ desc: null
879
+ value: false
880
+ torch_compile_backend:
881
+ desc: null
882
+ value: null
883
+ torch_compile_mode:
884
+ desc: null
885
+ value: null
886
+ dispatch_batches:
887
+ desc: null
888
+ value: null
889
+ split_batches:
890
+ desc: null
891
+ value: null
892
+ include_tokens_per_second:
893
+ desc: null
894
+ value: false
895
+ include_num_input_tokens_seen:
896
+ desc: null
897
+ value: false
898
+ neftune_noise_alpha:
899
+ desc: null
900
+ value: null
901
+ optim_target_modules:
902
+ desc: null
903
+ value: null
904
+ batch_eval_metrics:
905
+ desc: null
906
+ value: false
907
+ eval_on_start:
908
+ desc: null
909
+ value: false
910
+ use_liger_kernel:
911
+ desc: null
912
+ value: false
913
+ eval_use_gather_object:
914
+ desc: null
915
+ value: false
916
+ dataset_text_field:
917
+ desc: null
918
+ value: text
919
+ packing:
920
+ desc: null
921
+ value: false
922
+ dataset_num_proc:
923
+ desc: null
924
+ value: null
925
+ dataset_batch_size:
926
+ desc: null
927
+ value: 64
928
+ model_init_kwargs:
929
+ desc: null
930
+ value: null
931
+ dataset_kwargs:
932
+ desc: null
933
+ value: {}
934
+ eval_packing:
935
+ desc: null
936
+ value: null
937
+ num_of_sequences:
938
+ desc: null
939
+ value: 1024
940
+ chars_per_token:
941
+ desc: null
942
+ value: <CHARS_PER_TOKEN>
943
+ use_liger:
944
+ desc: null
945
+ value: false
946
+ model/num_parameters:
947
+ desc: null
948
+ value: 951599936
wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/diff.patch ADDED
@@ -0,0 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ diff --git a/qlearning_reasoning/training/sft.py b/qlearning_reasoning/training/sft.py
2
+ index 583fba2..1c2b359 100644
3
+ --- a/qlearning_reasoning/training/sft.py
4
+ +++ b/qlearning_reasoning/training/sft.py
5
+ @@ -51,6 +51,7 @@ flags.DEFINE_string('fsdp', "full_shard auto_wrap", 'If to use FSDP')
6
+ flags.DEFINE_string('fsdp_config', "qlearning_reasoning/accelerate_configs/fsdp_config.json", 'FSDP config to use.')
7
+ flags.DEFINE_boolean("push_to_hub", False, "Should the model be uploaded to Huggingface hub?")
8
+ flags.DEFINE_boolean('add_step_tokens', True, 'Add step tokens')
9
+ +flags.DEFINE_string("hub_model_id", None, "Hub model id. If not defined it will be truncated run name.")
10
+
11
+ START_OF_STEP_TOKEN = '[STEP]'
12
+ END_OF_STEP_TOKEN = '[/STEP]'
13
+ @@ -60,6 +61,10 @@ START_OF_REVISION_TOKEN = '[REVISION]'
14
+ END_OF_REVISION_TOKEN = '[/REVISION]'
15
+
16
+ def main(_):
17
+ +
18
+ + if FLAGS.hub_model_id is None:
19
+ + FLAGS.hub_model_id = FLAGS.run_name[:96]
20
+ +
21
+ if not os.path.exists(FLAGS.output_dir):
22
+ os.makedirs(FLAGS.output_dir, exist_ok=True)
23
+ os.makedirs(f"{FLAGS.output_dir}/{FLAGS.run_name}", exist_ok=True)
24
+ @@ -219,11 +224,11 @@ def main(_):
25
+ optim=FLAGS.optim,
26
+ weight_decay=FLAGS.weight_decay,
27
+ warmup_steps=FLAGS.warmup_steps,
28
+ - max_steps=num_steps,
29
+ fsdp=FLAGS.fsdp,
30
+ fsdp_config=FLAGS.fsdp_config if FLAGS.fsdp != '' else None,
31
+ push_to_hub=FLAGS.push_to_hub,
32
+ num_train_epochs=num_train_epochs,
33
+ + hub_model_id=FLAGS.hub_model_id
34
+ )
35
+
36
+ flags_dict = FLAGS.flag_values_dict()
37
+ diff --git a/sagemaker/launch_sagemaker.py b/sagemaker/launch_sagemaker.py
38
+ index 3316ede..98ed205 100644
39
+ --- a/sagemaker/launch_sagemaker.py
40
+ +++ b/sagemaker/launch_sagemaker.py
41
+ @@ -224,7 +224,7 @@ def main(argv):
42
+ # TODO: verify if this breaks the code
43
+ environment = {
44
+ "PYTHONPATH": "/opt/ml/code/qlearning_reasoning/",
45
+ - "HF_HOME": "/opt/ml/input/data/training/hf-cache",
46
+ + "HF_HOME": "/opt/ml/data/input/.cache",
47
+ "HF_TOKEN": get_hf_token(),
48
+ "WANDB_API_KEY": get_wandb_api_key(),
49
+ "SM_USE_RESERVED_CAPACITY": "1",
wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log ADDED
The diff for this file is too large to render. See raw diff
 
wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/requirements.txt ADDED
@@ -0,0 +1,289 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ absl-py==2.1.0
2
+ accelerate==1.1.1
3
+ aiohappyeyeballs==2.4.3
4
+ aiohttp==3.10.10
5
+ aiosignal==1.3.1
6
+ annotated-types==0.7.0
7
+ antlr4-python3-runtime==4.9.3
8
+ apex==0.1
9
+ appdirs==1.4.4
10
+ argparse==1.4.0
11
+ asttokens==2.4.1
12
+ async-timeout==4.0.3
13
+ attrdict==2.0.1
14
+ attrs==23.2.0
15
+ autocommand==2.2.2
16
+ awscli==1.33.40
17
+ backcall==0.2.0
18
+ backports.tarfile==1.2.0
19
+ bcrypt==4.2.0
20
+ beautifulsoup4==4.12.3
21
+ bleach==6.2.0
22
+ blis==0.7.10
23
+ bokeh==3.5.1
24
+ boltons==23.0.0
25
+ boto3==1.34.158
26
+ botocore==1.34.158
27
+ brotli==1.0.9
28
+ cached-property==1.5.2
29
+ cachetools==5.5.0
30
+ catalogue==2.0.10
31
+ certifi==2024.7.4
32
+ cffi==1.15.1
33
+ chardet==5.2.0
34
+ charset-normalizer==3.2.0
35
+ click==8.1.7
36
+ cloudpathlib==0.18.1
37
+ cloudpickle==2.2.1
38
+ cmake==3.30.2
39
+ colorama==0.4.6
40
+ comm==0.2.2
41
+ conda-libmamba-solver==23.3.0
42
+ conda-package-handling==2.2.0
43
+ conda-package-streaming==0.9.0
44
+ conda==23.3.1
45
+ confection==0.1.4
46
+ contourpy==1.2.1
47
+ cryptography==41.0.7
48
+ cycler==0.12.1
49
+ cymem==2.0.8
50
+ cython==3.0.11
51
+ datasets==3.1.0
52
+ debugpy==1.8.5
53
+ decorator==5.1.1
54
+ defusedxml==0.7.1
55
+ dill==0.3.8
56
+ docker-pycreds==0.4.0
57
+ docker==7.1.0
58
+ docopt==0.6.2
59
+ docstring-parser==0.16
60
+ docutils==0.16
61
+ einops==0.8.0
62
+ exceptiongroup==1.2.2
63
+ executing==2.0.1
64
+ fastai==2.7.16
65
+ fastcore==1.5.55
66
+ fastdownload==0.0.7
67
+ fastjsonschema==2.20.0
68
+ fastprogress==1.0.3
69
+ filelock==3.15.4
70
+ fire==0.6.0
71
+ flash-attn==2.5.5
72
+ fonttools==4.53.1
73
+ frozenlist==1.5.0
74
+ fsspec==2024.6.1
75
+ gevent==24.2.1
76
+ gitdb==4.0.11
77
+ gitpython==3.1.43
78
+ gmpy2==2.1.5
79
+ google-pasta==0.2.0
80
+ greenlet==3.0.3
81
+ h5py==3.11.0
82
+ huggingface-hub==0.26.2
83
+ hydra-core==1.3.2
84
+ idna==3.7
85
+ imageio==2.34.2
86
+ importlib-metadata==6.11.0
87
+ importlib-resources==6.1.1
88
+ inflect==7.3.1
89
+ inotify-simple==1.2.1
90
+ ipykernel==6.29.1
91
+ ipython==8.12.3
92
+ jaraco.context==5.3.0
93
+ jaraco.functools==4.0.1
94
+ jaraco.text==3.12.1
95
+ jedi==0.19.1
96
+ jinja2==3.1.4
97
+ jmespath==1.0.1
98
+ joblib==1.4.2
99
+ jsonpatch==1.32
100
+ jsonpointer==2.0
101
+ jsonschema-specifications==2023.12.1
102
+ jsonschema==4.23.0
103
+ jupyter-client==8.6.2
104
+ jupyter-core==5.7.2
105
+ jupyterlab-pygments==0.3.0
106
+ kiwisolver==1.4.5
107
+ langcodes==3.4.0
108
+ language-data==1.2.0
109
+ libmambapy==1.4.2
110
+ llvmlite==0.43.0
111
+ mamba==1.4.2
112
+ marisa-trie==1.1.0
113
+ markdown-it-py==3.0.0
114
+ markupsafe==2.1.5
115
+ matplotlib-inline==0.1.7
116
+ matplotlib==3.9.1
117
+ mdurl==0.1.2
118
+ mistune==3.0.2
119
+ more-itertools==10.3.0
120
+ mpi4py==3.1.6
121
+ mpmath==1.3.0
122
+ multidict==6.1.0
123
+ multiprocess==0.70.16
124
+ munkres==1.1.4
125
+ murmurhash==1.0.10
126
+ nbclient==0.10.0
127
+ nbconvert==7.16.4
128
+ nbformat==5.10.4
129
+ nest-asyncio==1.6.0
130
+ networkx==3.3
131
+ ninja==1.11.1.1
132
+ numba==0.60.0
133
+ numpy==1.24.3
134
+ nvidia-cublas-cu12==12.1.3.1
135
+ nvidia-cuda-cupti-cu12==12.1.105
136
+ nvidia-cuda-nvrtc-cu12==12.1.105
137
+ nvidia-cuda-runtime-cu12==12.1.105
138
+ nvidia-cudnn-cu12==8.9.2.26
139
+ nvidia-cufft-cu12==11.0.2.54
140
+ nvidia-curand-cu12==10.3.2.106
141
+ nvidia-cusolver-cu12==11.4.5.107
142
+ nvidia-cusparse-cu12==12.1.0.106
143
+ nvidia-ml-py==12.535.161
144
+ nvidia-nccl-cu12==2.19.3
145
+ nvidia-nvjitlink-cu12==12.6.77
146
+ nvidia-nvtx-cu12==12.1.105
147
+ nvitop==1.3.2
148
+ omegaconf==2.3.0
149
+ opencv-python==4.10.0.84
150
+ ordered-set==4.1.0
151
+ packaging==23.1
152
+ pandas==2.2.2
153
+ pandocfilters==1.5.1
154
+ paramiko==3.4.0
155
+ parso==0.8.4
156
+ pathos==0.3.2
157
+ pexpect==4.9.0
158
+ pickleshare==0.7.5
159
+ pillow==10.3.0
160
+ pip-chill==1.0.3
161
+ pip==24.1.2
162
+ pipreqs==0.5.0
163
+ platformdirs==4.2.2
164
+ plotly==5.23.0
165
+ pluggy==1.3.0
166
+ ply==3.11
167
+ pox==0.3.4
168
+ ppft==1.7.6.8
169
+ preshed==3.0.9
170
+ prompt-toolkit==3.0.47
171
+ propcache==0.2.0
172
+ protobuf==3.20.3
173
+ psutil==6.0.0
174
+ ptyprocess==0.7.0
175
+ pure-eval==0.2.3
176
+ pyarrow==17.0.0
177
+ pyasn1==0.6.0
178
+ pybind11-global==2.13.1
179
+ pybind11==2.13.1
180
+ pycosat==0.6.4
181
+ pycparser==2.21
182
+ pydantic-core==2.20.1
183
+ pydantic==2.8.2
184
+ pyfunctional==1.5.0
185
+ pygments==2.18.0
186
+ pynacl==1.5.0
187
+ pyopenssl==23.2.0
188
+ pyparsing==3.1.2
189
+ pyqt5-sip==12.12.2
190
+ pyqt5==5.15.9
191
+ pysocks==1.7.1
192
+ python-dateutil==2.9.0
193
+ pytz==2024.1
194
+ pyyaml==6.0.2
195
+ pyzmq==26.1.0
196
+ referencing==0.35.1
197
+ regex==2024.11.6
198
+ requests==2.32.3
199
+ retrying==1.3.4
200
+ rich==13.7.1
201
+ rpds-py==0.20.0
202
+ rsa==4.7.2
203
+ ruamel.yaml.clib==0.2.7
204
+ ruamel.yaml==0.17.32
205
+ s3fs==0.4.2
206
+ s3transfer==0.10.2
207
+ safetensors==0.4.5
208
+ sagemaker-experiments==0.1.45
209
+ sagemaker-pytorch-training==2.8.1
210
+ sagemaker-training==4.8.1
211
+ sagemaker==2.224.2
212
+ schema==0.7.7
213
+ scikit-learn==1.3.2
214
+ scipy==1.14.0
215
+ seaborn==0.13.2
216
+ sentencepiece==0.2.0
217
+ sentry-sdk==2.18.0
218
+ setproctitle==1.3.3
219
+ setuptools==72.1.0
220
+ shap==0.46.0
221
+ shellingham==1.5.4
222
+ shtab==1.7.1
223
+ sip==6.7.12
224
+ six==1.16.0
225
+ slicer==0.0.8
226
+ smart-open==7.0.4
227
+ smclarify==0.5
228
+ smdebug-rulesconfig==1.0.1
229
+ smdistributed-dataparallel==2.1.0
230
+ smmap==5.0.1
231
+ smprof==0.3.334
232
+ soupsieve==2.6
233
+ spacy-legacy==3.0.12
234
+ spacy-loggers==1.0.5
235
+ spacy==3.7.5
236
+ srsly==2.4.8
237
+ stack-data==0.6.2
238
+ sympy==1.13.0
239
+ tabulate==0.9.0
240
+ tblib==3.0.0
241
+ tenacity==9.0.0
242
+ tensor-parallel==2.0.0
243
+ termcolor==2.5.0
244
+ thinc==8.2.5
245
+ threadpoolctl==3.5.0
246
+ tinycss2==1.4.0
247
+ tokenizers==0.20.3
248
+ toml==0.10.2
249
+ tomli==2.0.1
250
+ toolz==0.12.0
251
+ torch==2.2.0
252
+ torchaudio==2.2.0
253
+ torchdata==0.7.0
254
+ torchnet==0.0.4
255
+ torchtext==0.16.0
256
+ torchvision==0.17.0
257
+ tornado==6.4.1
258
+ tqdm==4.66.5
259
+ traitlets==5.14.3
260
+ transformer-engine==0.12.0+170797
261
+ transformers==4.46.0
262
+ triton==2.2.0
263
+ trl==0.12.0
264
+ typeguard==4.3.0
265
+ typer-slim==0.12.3
266
+ typer==0.12.3
267
+ typing-extensions==4.12.2
268
+ tyro==0.8.5
269
+ tzdata==2024.1
270
+ unicodedata2==15.1.0
271
+ urllib3==1.26.19
272
+ visdom==0.2.4
273
+ wandb==0.16.3
274
+ wasabi==1.1.2
275
+ wcwidth==0.2.13
276
+ weasel==0.4.1
277
+ webencodings==0.5.1
278
+ websocket-client==1.8.0
279
+ werkzeug==3.0.3
280
+ wheel==0.41.2
281
+ wrapt==1.16.0
282
+ xxhash==3.5.0
283
+ xyzservices==2024.6.0
284
+ yarg==0.1.9
285
+ yarl==1.17.1
286
+ zipp==3.19.2
287
+ zope.event==5.0
288
+ zope.interface==7.0.1
289
+ zstandard==0.19.0
wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/wandb-metadata.json ADDED
@@ -0,0 +1,1069 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.10.226-214.879.amzn2.x86_64-x86_64-with-glibc2.31",
3
+ "python": "3.10.12",
4
+ "heartbeatAt": "2024-11-13T19:02:47.010671",
5
+ "startedAt": "2024-11-13T19:02:46.426230",
6
+ "docker": null,
7
+ "cuda": null,
8
+ "args": [
9
+ "--batch_size",
10
+ "64",
11
+ "--cache_dir",
12
+ "/opt/ml/data/input/.cache",
13
+ "--dataset",
14
+ "code-contests",
15
+ "--gradient_accumulation_steps",
16
+ "8",
17
+ "--logging_steps",
18
+ "5",
19
+ "--lr",
20
+ "1e-6",
21
+ "--max_seq_length",
22
+ "2048",
23
+ "--model_name",
24
+ "Qwen/Qwen2.5-Coder-7B-Instruct",
25
+ "--num_train_epochs",
26
+ "2",
27
+ "--output_dir",
28
+ "/opt/ml/model/",
29
+ "--push_to_hub",
30
+ "True",
31
+ "--run_name",
32
+ "sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048",
33
+ "--save_steps",
34
+ "100",
35
+ "--schedule",
36
+ "cosine",
37
+ "--wandb_project",
38
+ "sft-codecontests-1112",
39
+ "--weight_decay",
40
+ "0.0"
41
+ ],
42
+ "state": "running",
43
+ "program": "/opt/ml/code/qlearning_reasoning/training/sft.py",
44
+ "codePathLocal": "qlearning_reasoning/training/sft.py",
45
+ "codePath": "qlearning_reasoning/training/sft.py",
46
+ "git": {
47
+ "remote": "git@github.com:TRI-ML/reasoning-value-verifiers.git",
48
+ "commit": "cbfd05147f15c55be83f4996eebdb305c17f119d"
49
+ },
50
+ "email": null,
51
+ "root": "/opt/ml/code",
52
+ "host": "algo-1",
53
+ "username": "root",
54
+ "executable": "/opt/conda/bin/python3.10",
55
+ "cpu_count": 96,
56
+ "cpu_count_logical": 192,
57
+ "cpu_freq": {
58
+ "current": 2829.910838541667,
59
+ "min": 0.0,
60
+ "max": 0.0
61
+ },
62
+ "cpu_freq_per_core": [
63
+ {
64
+ "current": 2850.788,
65
+ "min": 0.0,
66
+ "max": 0.0
67
+ },
68
+ {
69
+ "current": 3406.159,
70
+ "min": 0.0,
71
+ "max": 0.0
72
+ },
73
+ {
74
+ "current": 2605.944,
75
+ "min": 0.0,
76
+ "max": 0.0
77
+ },
78
+ {
79
+ "current": 2608.162,
80
+ "min": 0.0,
81
+ "max": 0.0
82
+ },
83
+ {
84
+ "current": 2818.169,
85
+ "min": 0.0,
86
+ "max": 0.0
87
+ },
88
+ {
89
+ "current": 2591.551,
90
+ "min": 0.0,
91
+ "max": 0.0
92
+ },
93
+ {
94
+ "current": 2621.486,
95
+ "min": 0.0,
96
+ "max": 0.0
97
+ },
98
+ {
99
+ "current": 2607.504,
100
+ "min": 0.0,
101
+ "max": 0.0
102
+ },
103
+ {
104
+ "current": 2999.734,
105
+ "min": 0.0,
106
+ "max": 0.0
107
+ },
108
+ {
109
+ "current": 2991.449,
110
+ "min": 0.0,
111
+ "max": 0.0
112
+ },
113
+ {
114
+ "current": 3123.177,
115
+ "min": 0.0,
116
+ "max": 0.0
117
+ },
118
+ {
119
+ "current": 3053.641,
120
+ "min": 0.0,
121
+ "max": 0.0
122
+ },
123
+ {
124
+ "current": 3003.014,
125
+ "min": 0.0,
126
+ "max": 0.0
127
+ },
128
+ {
129
+ "current": 3000.178,
130
+ "min": 0.0,
131
+ "max": 0.0
132
+ },
133
+ {
134
+ "current": 3432.256,
135
+ "min": 0.0,
136
+ "max": 0.0
137
+ },
138
+ {
139
+ "current": 2999.802,
140
+ "min": 0.0,
141
+ "max": 0.0
142
+ },
143
+ {
144
+ "current": 3005.725,
145
+ "min": 0.0,
146
+ "max": 0.0
147
+ },
148
+ {
149
+ "current": 3565.885,
150
+ "min": 0.0,
151
+ "max": 0.0
152
+ },
153
+ {
154
+ "current": 2992.339,
155
+ "min": 0.0,
156
+ "max": 0.0
157
+ },
158
+ {
159
+ "current": 3104.468,
160
+ "min": 0.0,
161
+ "max": 0.0
162
+ },
163
+ {
164
+ "current": 3182.411,
165
+ "min": 0.0,
166
+ "max": 0.0
167
+ },
168
+ {
169
+ "current": 3013.217,
170
+ "min": 0.0,
171
+ "max": 0.0
172
+ },
173
+ {
174
+ "current": 3131.904,
175
+ "min": 0.0,
176
+ "max": 0.0
177
+ },
178
+ {
179
+ "current": 3007.314,
180
+ "min": 0.0,
181
+ "max": 0.0
182
+ },
183
+ {
184
+ "current": 2625.175,
185
+ "min": 0.0,
186
+ "max": 0.0
187
+ },
188
+ {
189
+ "current": 2653.407,
190
+ "min": 0.0,
191
+ "max": 0.0
192
+ },
193
+ {
194
+ "current": 2610.871,
195
+ "min": 0.0,
196
+ "max": 0.0
197
+ },
198
+ {
199
+ "current": 2793.855,
200
+ "min": 0.0,
201
+ "max": 0.0
202
+ },
203
+ {
204
+ "current": 2619.232,
205
+ "min": 0.0,
206
+ "max": 0.0
207
+ },
208
+ {
209
+ "current": 2627.886,
210
+ "min": 0.0,
211
+ "max": 0.0
212
+ },
213
+ {
214
+ "current": 2613.748,
215
+ "min": 0.0,
216
+ "max": 0.0
217
+ },
218
+ {
219
+ "current": 2318.039,
220
+ "min": 0.0,
221
+ "max": 0.0
222
+ },
223
+ {
224
+ "current": 3546.661,
225
+ "min": 0.0,
226
+ "max": 0.0
227
+ },
228
+ {
229
+ "current": 3022.56,
230
+ "min": 0.0,
231
+ "max": 0.0
232
+ },
233
+ {
234
+ "current": 3035.28,
235
+ "min": 0.0,
236
+ "max": 0.0
237
+ },
238
+ {
239
+ "current": 3541.443,
240
+ "min": 0.0,
241
+ "max": 0.0
242
+ },
243
+ {
244
+ "current": 3163.55,
245
+ "min": 0.0,
246
+ "max": 0.0
247
+ },
248
+ {
249
+ "current": 3023.79,
250
+ "min": 0.0,
251
+ "max": 0.0
252
+ },
253
+ {
254
+ "current": 3033.204,
255
+ "min": 0.0,
256
+ "max": 0.0
257
+ },
258
+ {
259
+ "current": 3031.695,
260
+ "min": 0.0,
261
+ "max": 0.0
262
+ },
263
+ {
264
+ "current": 3034.587,
265
+ "min": 0.0,
266
+ "max": 0.0
267
+ },
268
+ {
269
+ "current": 3031.71,
270
+ "min": 0.0,
271
+ "max": 0.0
272
+ },
273
+ {
274
+ "current": 3029.116,
275
+ "min": 0.0,
276
+ "max": 0.0
277
+ },
278
+ {
279
+ "current": 3293.701,
280
+ "min": 0.0,
281
+ "max": 0.0
282
+ },
283
+ {
284
+ "current": 3033.78,
285
+ "min": 0.0,
286
+ "max": 0.0
287
+ },
288
+ {
289
+ "current": 3032.034,
290
+ "min": 0.0,
291
+ "max": 0.0
292
+ },
293
+ {
294
+ "current": 3057.365,
295
+ "min": 0.0,
296
+ "max": 0.0
297
+ },
298
+ {
299
+ "current": 3028.183,
300
+ "min": 0.0,
301
+ "max": 0.0
302
+ },
303
+ {
304
+ "current": 3596.439,
305
+ "min": 0.0,
306
+ "max": 0.0
307
+ },
308
+ {
309
+ "current": 3327.853,
310
+ "min": 0.0,
311
+ "max": 0.0
312
+ },
313
+ {
314
+ "current": 3226.463,
315
+ "min": 0.0,
316
+ "max": 0.0
317
+ },
318
+ {
319
+ "current": 3259.361,
320
+ "min": 0.0,
321
+ "max": 0.0
322
+ },
323
+ {
324
+ "current": 3284.076,
325
+ "min": 0.0,
326
+ "max": 0.0
327
+ },
328
+ {
329
+ "current": 3257.698,
330
+ "min": 0.0,
331
+ "max": 0.0
332
+ },
333
+ {
334
+ "current": 3280.289,
335
+ "min": 0.0,
336
+ "max": 0.0
337
+ },
338
+ {
339
+ "current": 3258.834,
340
+ "min": 0.0,
341
+ "max": 0.0
342
+ },
343
+ {
344
+ "current": 2517.346,
345
+ "min": 0.0,
346
+ "max": 0.0
347
+ },
348
+ {
349
+ "current": 2553.091,
350
+ "min": 0.0,
351
+ "max": 0.0
352
+ },
353
+ {
354
+ "current": 2552.693,
355
+ "min": 0.0,
356
+ "max": 0.0
357
+ },
358
+ {
359
+ "current": 2689.091,
360
+ "min": 0.0,
361
+ "max": 0.0
362
+ },
363
+ {
364
+ "current": 2503.301,
365
+ "min": 0.0,
366
+ "max": 0.0
367
+ },
368
+ {
369
+ "current": 2524.4,
370
+ "min": 0.0,
371
+ "max": 0.0
372
+ },
373
+ {
374
+ "current": 2521.768,
375
+ "min": 0.0,
376
+ "max": 0.0
377
+ },
378
+ {
379
+ "current": 2530.364,
380
+ "min": 0.0,
381
+ "max": 0.0
382
+ },
383
+ {
384
+ "current": 3258.225,
385
+ "min": 0.0,
386
+ "max": 0.0
387
+ },
388
+ {
389
+ "current": 3597.406,
390
+ "min": 0.0,
391
+ "max": 0.0
392
+ },
393
+ {
394
+ "current": 3279.658,
395
+ "min": 0.0,
396
+ "max": 0.0
397
+ },
398
+ {
399
+ "current": 3292.257,
400
+ "min": 0.0,
401
+ "max": 0.0
402
+ },
403
+ {
404
+ "current": 3265.351,
405
+ "min": 0.0,
406
+ "max": 0.0
407
+ },
408
+ {
409
+ "current": 3286.715,
410
+ "min": 0.0,
411
+ "max": 0.0
412
+ },
413
+ {
414
+ "current": 3268.777,
415
+ "min": 0.0,
416
+ "max": 0.0
417
+ },
418
+ {
419
+ "current": 3242.068,
420
+ "min": 0.0,
421
+ "max": 0.0
422
+ },
423
+ {
424
+ "current": 3288.908,
425
+ "min": 0.0,
426
+ "max": 0.0
427
+ },
428
+ {
429
+ "current": 3278.645,
430
+ "min": 0.0,
431
+ "max": 0.0
432
+ },
433
+ {
434
+ "current": 3215.578,
435
+ "min": 0.0,
436
+ "max": 0.0
437
+ },
438
+ {
439
+ "current": 3327.101,
440
+ "min": 0.0,
441
+ "max": 0.0
442
+ },
443
+ {
444
+ "current": 3357.075,
445
+ "min": 0.0,
446
+ "max": 0.0
447
+ },
448
+ {
449
+ "current": 3144.838,
450
+ "min": 0.0,
451
+ "max": 0.0
452
+ },
453
+ {
454
+ "current": 3005.517,
455
+ "min": 0.0,
456
+ "max": 0.0
457
+ },
458
+ {
459
+ "current": 3597.381,
460
+ "min": 0.0,
461
+ "max": 0.0
462
+ },
463
+ {
464
+ "current": 3596.175,
465
+ "min": 0.0,
466
+ "max": 0.0
467
+ },
468
+ {
469
+ "current": 3317.258,
470
+ "min": 0.0,
471
+ "max": 0.0
472
+ },
473
+ {
474
+ "current": 3310.249,
475
+ "min": 0.0,
476
+ "max": 0.0
477
+ },
478
+ {
479
+ "current": 3140.687,
480
+ "min": 0.0,
481
+ "max": 0.0
482
+ },
483
+ {
484
+ "current": 3398.954,
485
+ "min": 0.0,
486
+ "max": 0.0
487
+ },
488
+ {
489
+ "current": 3318.433,
490
+ "min": 0.0,
491
+ "max": 0.0
492
+ },
493
+ {
494
+ "current": 3342.35,
495
+ "min": 0.0,
496
+ "max": 0.0
497
+ },
498
+ {
499
+ "current": 3328.575,
500
+ "min": 0.0,
501
+ "max": 0.0
502
+ },
503
+ {
504
+ "current": 3499.892,
505
+ "min": 0.0,
506
+ "max": 0.0
507
+ },
508
+ {
509
+ "current": 3344.864,
510
+ "min": 0.0,
511
+ "max": 0.0
512
+ },
513
+ {
514
+ "current": 3355.802,
515
+ "min": 0.0,
516
+ "max": 0.0
517
+ },
518
+ {
519
+ "current": 3305.032,
520
+ "min": 0.0,
521
+ "max": 0.0
522
+ },
523
+ {
524
+ "current": 3318.07,
525
+ "min": 0.0,
526
+ "max": 0.0
527
+ },
528
+ {
529
+ "current": 3336.546,
530
+ "min": 0.0,
531
+ "max": 0.0
532
+ },
533
+ {
534
+ "current": 3335.458,
535
+ "min": 0.0,
536
+ "max": 0.0
537
+ },
538
+ {
539
+ "current": 3596.561,
540
+ "min": 0.0,
541
+ "max": 0.0
542
+ },
543
+ {
544
+ "current": 2884.473,
545
+ "min": 0.0,
546
+ "max": 0.0
547
+ },
548
+ {
549
+ "current": 2934.098,
550
+ "min": 0.0,
551
+ "max": 0.0
552
+ },
553
+ {
554
+ "current": 2621.598,
555
+ "min": 0.0,
556
+ "max": 0.0
557
+ },
558
+ {
559
+ "current": 2645.184,
560
+ "min": 0.0,
561
+ "max": 0.0
562
+ },
563
+ {
564
+ "current": 2480.726,
565
+ "min": 0.0,
566
+ "max": 0.0
567
+ },
568
+ {
569
+ "current": 2515.912,
570
+ "min": 0.0,
571
+ "max": 0.0
572
+ },
573
+ {
574
+ "current": 2702.57,
575
+ "min": 0.0,
576
+ "max": 0.0
577
+ },
578
+ {
579
+ "current": 2671.473,
580
+ "min": 0.0,
581
+ "max": 0.0
582
+ },
583
+ {
584
+ "current": 3092.321,
585
+ "min": 0.0,
586
+ "max": 0.0
587
+ },
588
+ {
589
+ "current": 3092.204,
590
+ "min": 0.0,
591
+ "max": 0.0
592
+ },
593
+ {
594
+ "current": 3217.287,
595
+ "min": 0.0,
596
+ "max": 0.0
597
+ },
598
+ {
599
+ "current": 3161.646,
600
+ "min": 0.0,
601
+ "max": 0.0
602
+ },
603
+ {
604
+ "current": 3089.273,
605
+ "min": 0.0,
606
+ "max": 0.0
607
+ },
608
+ {
609
+ "current": 3088.447,
610
+ "min": 0.0,
611
+ "max": 0.0
612
+ },
613
+ {
614
+ "current": 3559.214,
615
+ "min": 0.0,
616
+ "max": 0.0
617
+ },
618
+ {
619
+ "current": 3218.611,
620
+ "min": 0.0,
621
+ "max": 0.0
622
+ },
623
+ {
624
+ "current": 3021.622,
625
+ "min": 0.0,
626
+ "max": 0.0
627
+ },
628
+ {
629
+ "current": 3332.084,
630
+ "min": 0.0,
631
+ "max": 0.0
632
+ },
633
+ {
634
+ "current": 3019.854,
635
+ "min": 0.0,
636
+ "max": 0.0
637
+ },
638
+ {
639
+ "current": 3213.831,
640
+ "min": 0.0,
641
+ "max": 0.0
642
+ },
643
+ {
644
+ "current": 3092.774,
645
+ "min": 0.0,
646
+ "max": 0.0
647
+ },
648
+ {
649
+ "current": 3015.868,
650
+ "min": 0.0,
651
+ "max": 0.0
652
+ },
653
+ {
654
+ "current": 3082.752,
655
+ "min": 0.0,
656
+ "max": 0.0
657
+ },
658
+ {
659
+ "current": 3029.559,
660
+ "min": 0.0,
661
+ "max": 0.0
662
+ },
663
+ {
664
+ "current": 2873.142,
665
+ "min": 0.0,
666
+ "max": 0.0
667
+ },
668
+ {
669
+ "current": 2215.288,
670
+ "min": 0.0,
671
+ "max": 0.0
672
+ },
673
+ {
674
+ "current": 3193.443,
675
+ "min": 0.0,
676
+ "max": 0.0
677
+ },
678
+ {
679
+ "current": 2928.717,
680
+ "min": 0.0,
681
+ "max": 0.0
682
+ },
683
+ {
684
+ "current": 2990.366,
685
+ "min": 0.0,
686
+ "max": 0.0
687
+ },
688
+ {
689
+ "current": 2923.847,
690
+ "min": 0.0,
691
+ "max": 0.0
692
+ },
693
+ {
694
+ "current": 2869.364,
695
+ "min": 0.0,
696
+ "max": 0.0
697
+ },
698
+ {
699
+ "current": 2859.189,
700
+ "min": 0.0,
701
+ "max": 0.0
702
+ },
703
+ {
704
+ "current": 3049.001,
705
+ "min": 0.0,
706
+ "max": 0.0
707
+ },
708
+ {
709
+ "current": 2805.206,
710
+ "min": 0.0,
711
+ "max": 0.0
712
+ },
713
+ {
714
+ "current": 3094.081,
715
+ "min": 0.0,
716
+ "max": 0.0
717
+ },
718
+ {
719
+ "current": 3187.222,
720
+ "min": 0.0,
721
+ "max": 0.0
722
+ },
723
+ {
724
+ "current": 2853.945,
725
+ "min": 0.0,
726
+ "max": 0.0
727
+ },
728
+ {
729
+ "current": 2944.73,
730
+ "min": 0.0,
731
+ "max": 0.0
732
+ },
733
+ {
734
+ "current": 2809.161,
735
+ "min": 0.0,
736
+ "max": 0.0
737
+ },
738
+ {
739
+ "current": 2806.787,
740
+ "min": 0.0,
741
+ "max": 0.0
742
+ },
743
+ {
744
+ "current": 3024.225,
745
+ "min": 0.0,
746
+ "max": 0.0
747
+ },
748
+ {
749
+ "current": 3032.173,
750
+ "min": 0.0,
751
+ "max": 0.0
752
+ },
753
+ {
754
+ "current": 3033.195,
755
+ "min": 0.0,
756
+ "max": 0.0
757
+ },
758
+ {
759
+ "current": 3561.287,
760
+ "min": 0.0,
761
+ "max": 0.0
762
+ },
763
+ {
764
+ "current": 2987.872,
765
+ "min": 0.0,
766
+ "max": 0.0
767
+ },
768
+ {
769
+ "current": 3051.252,
770
+ "min": 0.0,
771
+ "max": 0.0
772
+ },
773
+ {
774
+ "current": 3379.017,
775
+ "min": 0.0,
776
+ "max": 0.0
777
+ },
778
+ {
779
+ "current": 3044.319,
780
+ "min": 0.0,
781
+ "max": 0.0
782
+ },
783
+ {
784
+ "current": 3595.62,
785
+ "min": 0.0,
786
+ "max": 0.0
787
+ },
788
+ {
789
+ "current": 3261.737,
790
+ "min": 0.0,
791
+ "max": 0.0
792
+ },
793
+ {
794
+ "current": 3257.013,
795
+ "min": 0.0,
796
+ "max": 0.0
797
+ },
798
+ {
799
+ "current": 3288.317,
800
+ "min": 0.0,
801
+ "max": 0.0
802
+ },
803
+ {
804
+ "current": 3268.675,
805
+ "min": 0.0,
806
+ "max": 0.0
807
+ },
808
+ {
809
+ "current": 3269.579,
810
+ "min": 0.0,
811
+ "max": 0.0
812
+ },
813
+ {
814
+ "current": 3266.213,
815
+ "min": 0.0,
816
+ "max": 0.0
817
+ },
818
+ {
819
+ "current": 3265.148,
820
+ "min": 0.0,
821
+ "max": 0.0
822
+ },
823
+ {
824
+ "current": 2554.152,
825
+ "min": 0.0,
826
+ "max": 0.0
827
+ },
828
+ {
829
+ "current": 2556.63,
830
+ "min": 0.0,
831
+ "max": 0.0
832
+ },
833
+ {
834
+ "current": 2540.571,
835
+ "min": 0.0,
836
+ "max": 0.0
837
+ },
838
+ {
839
+ "current": 2512.005,
840
+ "min": 0.0,
841
+ "max": 0.0
842
+ },
843
+ {
844
+ "current": 2560.992,
845
+ "min": 0.0,
846
+ "max": 0.0
847
+ },
848
+ {
849
+ "current": 2526.641,
850
+ "min": 0.0,
851
+ "max": 0.0
852
+ },
853
+ {
854
+ "current": 2528.267,
855
+ "min": 0.0,
856
+ "max": 0.0
857
+ },
858
+ {
859
+ "current": 2583.061,
860
+ "min": 0.0,
861
+ "max": 0.0
862
+ },
863
+ {
864
+ "current": 3257.652,
865
+ "min": 0.0,
866
+ "max": 0.0
867
+ },
868
+ {
869
+ "current": 3598.217,
870
+ "min": 0.0,
871
+ "max": 0.0
872
+ },
873
+ {
874
+ "current": 3282.578,
875
+ "min": 0.0,
876
+ "max": 0.0
877
+ },
878
+ {
879
+ "current": 3261.664,
880
+ "min": 0.0,
881
+ "max": 0.0
882
+ },
883
+ {
884
+ "current": 3268.246,
885
+ "min": 0.0,
886
+ "max": 0.0
887
+ },
888
+ {
889
+ "current": 3331.937,
890
+ "min": 0.0,
891
+ "max": 0.0
892
+ },
893
+ {
894
+ "current": 3265.603,
895
+ "min": 0.0,
896
+ "max": 0.0
897
+ },
898
+ {
899
+ "current": 3258.501,
900
+ "min": 0.0,
901
+ "max": 0.0
902
+ },
903
+ {
904
+ "current": 3337.045,
905
+ "min": 0.0,
906
+ "max": 0.0
907
+ },
908
+ {
909
+ "current": 3259.949,
910
+ "min": 0.0,
911
+ "max": 0.0
912
+ },
913
+ {
914
+ "current": 3226.058,
915
+ "min": 0.0,
916
+ "max": 0.0
917
+ },
918
+ {
919
+ "current": 3250.308,
920
+ "min": 0.0,
921
+ "max": 0.0
922
+ },
923
+ {
924
+ "current": 3208.026,
925
+ "min": 0.0,
926
+ "max": 0.0
927
+ },
928
+ {
929
+ "current": 2974.043,
930
+ "min": 0.0,
931
+ "max": 0.0
932
+ },
933
+ {
934
+ "current": 3216.706,
935
+ "min": 0.0,
936
+ "max": 0.0
937
+ },
938
+ {
939
+ "current": 3598.55,
940
+ "min": 0.0,
941
+ "max": 0.0
942
+ },
943
+ {
944
+ "current": 3598.261,
945
+ "min": 0.0,
946
+ "max": 0.0
947
+ },
948
+ {
949
+ "current": 3222.637,
950
+ "min": 0.0,
951
+ "max": 0.0
952
+ },
953
+ {
954
+ "current": 3227.662,
955
+ "min": 0.0,
956
+ "max": 0.0
957
+ },
958
+ {
959
+ "current": 3143.781,
960
+ "min": 0.0,
961
+ "max": 0.0
962
+ },
963
+ {
964
+ "current": 3215.567,
965
+ "min": 0.0,
966
+ "max": 0.0
967
+ },
968
+ {
969
+ "current": 3250.612,
970
+ "min": 0.0,
971
+ "max": 0.0
972
+ },
973
+ {
974
+ "current": 3195.784,
975
+ "min": 0.0,
976
+ "max": 0.0
977
+ },
978
+ {
979
+ "current": 3219.289,
980
+ "min": 0.0,
981
+ "max": 0.0
982
+ },
983
+ {
984
+ "current": 3289.163,
985
+ "min": 0.0,
986
+ "max": 0.0
987
+ },
988
+ {
989
+ "current": 3247.875,
990
+ "min": 0.0,
991
+ "max": 0.0
992
+ },
993
+ {
994
+ "current": 3249.112,
995
+ "min": 0.0,
996
+ "max": 0.0
997
+ },
998
+ {
999
+ "current": 3209.3,
1000
+ "min": 0.0,
1001
+ "max": 0.0
1002
+ },
1003
+ {
1004
+ "current": 3187.212,
1005
+ "min": 0.0,
1006
+ "max": 0.0
1007
+ },
1008
+ {
1009
+ "current": 3288.003,
1010
+ "min": 0.0,
1011
+ "max": 0.0
1012
+ },
1013
+ {
1014
+ "current": 3169.417,
1015
+ "min": 0.0,
1016
+ "max": 0.0
1017
+ },
1018
+ {
1019
+ "current": 3596.071,
1020
+ "min": 0.0,
1021
+ "max": 0.0
1022
+ }
1023
+ ],
1024
+ "disk": {
1025
+ "/": {
1026
+ "total": 119.94140625,
1027
+ "used": 56.4015998840332
1028
+ }
1029
+ },
1030
+ "gpu": "NVIDIA H100 80GB HBM3",
1031
+ "gpu_count": 8,
1032
+ "gpu_devices": [
1033
+ {
1034
+ "name": "NVIDIA H100 80GB HBM3",
1035
+ "memory_total": 85520809984
1036
+ },
1037
+ {
1038
+ "name": "NVIDIA H100 80GB HBM3",
1039
+ "memory_total": 85520809984
1040
+ },
1041
+ {
1042
+ "name": "NVIDIA H100 80GB HBM3",
1043
+ "memory_total": 85520809984
1044
+ },
1045
+ {
1046
+ "name": "NVIDIA H100 80GB HBM3",
1047
+ "memory_total": 85520809984
1048
+ },
1049
+ {
1050
+ "name": "NVIDIA H100 80GB HBM3",
1051
+ "memory_total": 85520809984
1052
+ },
1053
+ {
1054
+ "name": "NVIDIA H100 80GB HBM3",
1055
+ "memory_total": 85520809984
1056
+ },
1057
+ {
1058
+ "name": "NVIDIA H100 80GB HBM3",
1059
+ "memory_total": 85520809984
1060
+ },
1061
+ {
1062
+ "name": "NVIDIA H100 80GB HBM3",
1063
+ "memory_total": 85520809984
1064
+ }
1065
+ ],
1066
+ "memory": {
1067
+ "total": 1999.9661331176758
1068
+ }
1069
+ }
wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"train/loss": 0.856, "train/grad_norm": 2.515625, "train/learning_rate": 3.086582838174551e-07, "train/epoch": 1.22, "train/global_step": 15, "_timestamp": 1731525055.2001407, "_runtime": 488.7689197063446, "_step": 6, "eval/loss": 0.7827465534210205, "eval/runtime": 1.6563, "eval/samples_per_second": 61.584, "eval/steps_per_second": 1.208}
wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/logs/debug-internal.log ADDED
@@ -0,0 +1,373 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-11-13 19:02:46,431 INFO StreamThr :1939 [internal.py:wandb_internal():86] W&B internal server running at pid: 1939, started at: 2024-11-13 19:02:46.431351
2
+ 2024-11-13 19:02:46,433 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status
3
+ 2024-11-13 19:02:46,434 INFO WriterThread:1939 [datastore.py:open_for_write():87] open: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/run-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1.wandb
4
+ 2024-11-13 19:02:46,435 DEBUG SenderThread:1939 [sender.py:send():382] send: header
5
+ 2024-11-13 19:02:46,442 DEBUG SenderThread:1939 [sender.py:send():382] send: run
6
+ 2024-11-13 19:02:46,772 INFO SenderThread:1939 [dir_watcher.py:__init__():211] watching files in: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files
7
+ 2024-11-13 19:02:46,772 INFO SenderThread:1939 [sender.py:_start_run_threads():1136] run started: kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1 with start time 1731524566.431221
8
+ 2024-11-13 19:02:46,782 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: check_version
9
+ 2024-11-13 19:02:46,782 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: check_version
10
+ 2024-11-13 19:02:46,859 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: run_start
11
+ 2024-11-13 19:02:46,947 DEBUG HandlerThread:1939 [system_info.py:__init__():27] System info init
12
+ 2024-11-13 19:02:46,947 DEBUG HandlerThread:1939 [system_info.py:__init__():42] System info init done
13
+ 2024-11-13 19:02:46,947 INFO HandlerThread:1939 [system_monitor.py:start():194] Starting system monitor
14
+ 2024-11-13 19:02:46,947 INFO SystemMonitor:1939 [system_monitor.py:_start():158] Starting system asset monitoring threads
15
+ 2024-11-13 19:02:46,948 INFO HandlerThread:1939 [system_monitor.py:probe():214] Collecting system info
16
+ 2024-11-13 19:02:46,948 INFO SystemMonitor:1939 [interfaces.py:start():190] Started cpu monitoring
17
+ 2024-11-13 19:02:46,949 INFO SystemMonitor:1939 [interfaces.py:start():190] Started disk monitoring
18
+ 2024-11-13 19:02:46,949 INFO SystemMonitor:1939 [interfaces.py:start():190] Started gpu monitoring
19
+ 2024-11-13 19:02:46,951 INFO SystemMonitor:1939 [interfaces.py:start():190] Started memory monitoring
20
+ 2024-11-13 19:02:46,952 INFO SystemMonitor:1939 [interfaces.py:start():190] Started network monitoring
21
+ 2024-11-13 19:02:47,010 DEBUG HandlerThread:1939 [system_info.py:probe():151] Probing system
22
+ 2024-11-13 19:02:47,013 DEBUG HandlerThread:1939 [system_info.py:_probe_git():136] Probing git
23
+ 2024-11-13 19:02:47,021 DEBUG HandlerThread:1939 [system_info.py:_probe_git():144] Probing git done
24
+ 2024-11-13 19:02:47,021 DEBUG HandlerThread:1939 [system_info.py:probe():199] Probing system done
25
+ 2024-11-13 19:02:47,022 DEBUG HandlerThread:1939 [system_monitor.py:probe():223] {'os': 'Linux-5.10.226-214.879.amzn2.x86_64-x86_64-with-glibc2.31', 'python': '3.10.12', 'heartbeatAt': '2024-11-13T19:02:47.010671', 'startedAt': '2024-11-13T19:02:46.426230', 'docker': None, 'cuda': None, 'args': ('--batch_size', '64', '--cache_dir', '/opt/ml/data/input/.cache', '--dataset', 'code-contests', '--gradient_accumulation_steps', '8', '--logging_steps', '5', '--lr', '1e-6', '--max_seq_length', '2048', '--model_name', 'Qwen/Qwen2.5-Coder-7B-Instruct', '--num_train_epochs', '2', '--output_dir', '/opt/ml/model/', '--push_to_hub', 'True', '--run_name', 'sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048', '--save_steps', '100', '--schedule', 'cosine', '--wandb_project', 'sft-codecontests-1112', '--weight_decay', '0.0'), 'state': 'running', 'program': '/opt/ml/code/qlearning_reasoning/training/sft.py', 'codePathLocal': 'qlearning_reasoning/training/sft.py', 'codePath': 'qlearning_reasoning/training/sft.py', 'git': {'remote': 'git@github.com:TRI-ML/reasoning-value-verifiers.git', 'commit': 'cbfd05147f15c55be83f4996eebdb305c17f119d'}, 'email': None, 'root': '/opt/ml/code', 'host': 'algo-1', 'username': 'root', 'executable': '/opt/conda/bin/python3.10', 'cpu_count': 96, 'cpu_count_logical': 192, 'cpu_freq': {'current': 2829.910838541667, 'min': 0.0, 'max': 0.0}, 'cpu_freq_per_core': [{'current': 2850.788, 'min': 0.0, 'max': 0.0}, {'current': 3406.159, 'min': 0.0, 'max': 0.0}, {'current': 2605.944, 'min': 0.0, 'max': 0.0}, {'current': 2608.162, 'min': 0.0, 'max': 0.0}, {'current': 2818.169, 'min': 0.0, 'max': 0.0}, {'current': 2591.551, 'min': 0.0, 'max': 0.0}, {'current': 2621.486, 'min': 0.0, 'max': 0.0}, {'current': 2607.504, 'min': 0.0, 'max': 0.0}, {'current': 2999.734, 'min': 0.0, 'max': 0.0}, {'current': 2991.449, 'min': 0.0, 'max': 0.0}, {'current': 3123.177, 'min': 0.0, 'max': 0.0}, {'current': 3053.641, 'min': 0.0, 'max': 0.0}, {'current': 3003.014, 'min': 0.0, 'max': 0.0}, {'current': 3000.178, 'min': 0.0, 'max': 0.0}, {'current': 3432.256, 'min': 0.0, 'max': 0.0}, {'current': 2999.802, 'min': 0.0, 'max': 0.0}, {'current': 3005.725, 'min': 0.0, 'max': 0.0}, {'current': 3565.885, 'min': 0.0, 'max': 0.0}, {'current': 2992.339, 'min': 0.0, 'max': 0.0}, {'current': 3104.468, 'min': 0.0, 'max': 0.0}, {'current': 3182.411, 'min': 0.0, 'max': 0.0}, {'current': 3013.217, 'min': 0.0, 'max': 0.0}, {'current': 3131.904, 'min': 0.0, 'max': 0.0}, {'current': 3007.314, 'min': 0.0, 'max': 0.0}, {'current': 2625.175, 'min': 0.0, 'max': 0.0}, {'current': 2653.407, 'min': 0.0, 'max': 0.0}, {'current': 2610.871, 'min': 0.0, 'max': 0.0}, {'current': 2793.855, 'min': 0.0, 'max': 0.0}, {'current': 2619.232, 'min': 0.0, 'max': 0.0}, {'current': 2627.886, 'min': 0.0, 'max': 0.0}, {'current': 2613.748, 'min': 0.0, 'max': 0.0}, {'current': 2318.039, 'min': 0.0, 'max': 0.0}, {'current': 3546.661, 'min': 0.0, 'max': 0.0}, {'current': 3022.56, 'min': 0.0, 'max': 0.0}, {'current': 3035.28, 'min': 0.0, 'max': 0.0}, {'current': 3541.443, 'min': 0.0, 'max': 0.0}, {'current': 3163.55, 'min': 0.0, 'max': 0.0}, {'current': 3023.79, 'min': 0.0, 'max': 0.0}, {'current': 3033.204, 'min': 0.0, 'max': 0.0}, {'current': 3031.695, 'min': 0.0, 'max': 0.0}, {'current': 3034.587, 'min': 0.0, 'max': 0.0}, {'current': 3031.71, 'min': 0.0, 'max': 0.0}, {'current': 3029.116, 'min': 0.0, 'max': 0.0}, {'current': 3293.701, 'min': 0.0, 'max': 0.0}, {'current': 3033.78, 'min': 0.0, 'max': 0.0}, {'current': 3032.034, 'min': 0.0, 'max': 0.0}, {'current': 3057.365, 'min': 0.0, 'max': 0.0}, {'current': 3028.183, 'min': 0.0, 'max': 0.0}, {'current': 3596.439, 'min': 0.0, 'max': 0.0}, {'current': 3327.853, 'min': 0.0, 'max': 0.0}, {'current': 3226.463, 'min': 0.0, 'max': 0.0}, {'current': 3259.361, 'min': 0.0, 'max': 0.0}, {'current': 3284.076, 'min': 0.0, 'max': 0.0}, {'current': 3257.698, 'min': 0.0, 'max': 0.0}, {'current': 3280.289, 'min': 0.0, 'max': 0.0}, {'current': 3258.834, 'min': 0.0, 'max': 0.0}, {'current': 2517.346, 'min': 0.0, 'max': 0.0}, {'current': 2553.091, 'min': 0.0, 'max': 0.0}, {'current': 2552.693, 'min': 0.0, 'max': 0.0}, {'current': 2689.091, 'min': 0.0, 'max': 0.0}, {'current': 2503.301, 'min': 0.0, 'max': 0.0}, {'current': 2524.4, 'min': 0.0, 'max': 0.0}, {'current': 2521.768, 'min': 0.0, 'max': 0.0}, {'current': 2530.364, 'min': 0.0, 'max': 0.0}, {'current': 3258.225, 'min': 0.0, 'max': 0.0}, {'current': 3597.406, 'min': 0.0, 'max': 0.0}, {'current': 3279.658, 'min': 0.0, 'max': 0.0}, {'current': 3292.257, 'min': 0.0, 'max': 0.0}, {'current': 3265.351, 'min': 0.0, 'max': 0.0}, {'current': 3286.715, 'min': 0.0, 'max': 0.0}, {'current': 3268.777, 'min': 0.0, 'max': 0.0}, {'current': 3242.068, 'min': 0.0, 'max': 0.0}, {'current': 3288.908, 'min': 0.0, 'max': 0.0}, {'current': 3278.645, 'min': 0.0, 'max': 0.0}, {'current': 3215.578, 'min': 0.0, 'max': 0.0}, {'current': 3327.101, 'min': 0.0, 'max': 0.0}, {'current': 3357.075, 'min': 0.0, 'max': 0.0}, {'current': 3144.838, 'min': 0.0, 'max': 0.0}, {'current': 3005.517, 'min': 0.0, 'max': 0.0}, {'current': 3597.381, 'min': 0.0, 'max': 0.0}, {'current': 3596.175, 'min': 0.0, 'max': 0.0}, {'current': 3317.258, 'min': 0.0, 'max': 0.0}, {'current': 3310.249, 'min': 0.0, 'max': 0.0}, {'current': 3140.687, 'min': 0.0, 'max': 0.0}, {'current': 3398.954, 'min': 0.0, 'max': 0.0}, {'current': 3318.433, 'min': 0.0, 'max': 0.0}, {'current': 3342.35, 'min': 0.0, 'max': 0.0}, {'current': 3328.575, 'min': 0.0, 'max': 0.0}, {'current': 3499.892, 'min': 0.0, 'max': 0.0}, {'current': 3344.864, 'min': 0.0, 'max': 0.0}, {'current': 3355.802, 'min': 0.0, 'max': 0.0}, {'current': 3305.032, 'min': 0.0, 'max': 0.0}, {'current': 3318.07, 'min': 0.0, 'max': 0.0}, {'current': 3336.546, 'min': 0.0, 'max': 0.0}, {'current': 3335.458, 'min': 0.0, 'max': 0.0}, {'current': 3596.561, 'min': 0.0, 'max': 0.0}, {'current': 2884.473, 'min': 0.0, 'max': 0.0}, {'current': 2934.098, 'min': 0.0, 'max': 0.0}, {'current': 2621.598, 'min': 0.0, 'max': 0.0}, {'current': 2645.184, 'min': 0.0, 'max': 0.0}, {'current': 2480.726, 'min': 0.0, 'max': 0.0}, {'current': 2515.912, 'min': 0.0, 'max': 0.0}, {'current': 2702.57, 'min': 0.0, 'max': 0.0}, {'current': 2671.473, 'min': 0.0, 'max': 0.0}, {'current': 3092.321, 'min': 0.0, 'max': 0.0}, {'current': 3092.204, 'min': 0.0, 'max': 0.0}, {'current': 3217.287, 'min': 0.0, 'max': 0.0}, {'current': 3161.646, 'min': 0.0, 'max': 0.0}, {'current': 3089.273, 'min': 0.0, 'max': 0.0}, {'current': 3088.447, 'min': 0.0, 'max': 0.0}, {'current': 3559.214, 'min': 0.0, 'max': 0.0}, {'current': 3218.611, 'min': 0.0, 'max': 0.0}, {'current': 3021.622, 'min': 0.0, 'max': 0.0}, {'current': 3332.084, 'min': 0.0, 'max': 0.0}, {'current': 3019.854, 'min': 0.0, 'max': 0.0}, {'current': 3213.831, 'min': 0.0, 'max': 0.0}, {'current': 3092.774, 'min': 0.0, 'max': 0.0}, {'current': 3015.868, 'min': 0.0, 'max': 0.0}, {'current': 3082.752, 'min': 0.0, 'max': 0.0}, {'current': 3029.559, 'min': 0.0, 'max': 0.0}, {'current': 2873.142, 'min': 0.0, 'max': 0.0}, {'current': 2215.288, 'min': 0.0, 'max': 0.0}, {'current': 3193.443, 'min': 0.0, 'max': 0.0}, {'current': 2928.717, 'min': 0.0, 'max': 0.0}, {'current': 2990.366, 'min': 0.0, 'max': 0.0}, {'current': 2923.847, 'min': 0.0, 'max': 0.0}, {'current': 2869.364, 'min': 0.0, 'max': 0.0}, {'current': 2859.189, 'min': 0.0, 'max': 0.0}, {'current': 3049.001, 'min': 0.0, 'max': 0.0}, {'current': 2805.206, 'min': 0.0, 'max': 0.0}, {'current': 3094.081, 'min': 0.0, 'max': 0.0}, {'current': 3187.222, 'min': 0.0, 'max': 0.0}, {'current': 2853.945, 'min': 0.0, 'max': 0.0}, {'current': 2944.73, 'min': 0.0, 'max': 0.0}, {'current': 2809.161, 'min': 0.0, 'max': 0.0}, {'current': 2806.787, 'min': 0.0, 'max': 0.0}, {'current': 3024.225, 'min': 0.0, 'max': 0.0}, {'current': 3032.173, 'min': 0.0, 'max': 0.0}, {'current': 3033.195, 'min': 0.0, 'max': 0.0}, {'current': 3561.287, 'min': 0.0, 'max': 0.0}, {'current': 2987.872, 'min': 0.0, 'max': 0.0}, {'current': 3051.252, 'min': 0.0, 'max': 0.0}, {'current': 3379.017, 'min': 0.0, 'max': 0.0}, {'current': 3044.319, 'min': 0.0, 'max': 0.0}, {'current': 3595.62, 'min': 0.0, 'max': 0.0}, {'current': 3261.737, 'min': 0.0, 'max': 0.0}, {'current': 3257.013, 'min': 0.0, 'max': 0.0}, {'current': 3288.317, 'min': 0.0, 'max': 0.0}, {'current': 3268.675, 'min': 0.0, 'max': 0.0}, {'current': 3269.579, 'min': 0.0, 'max': 0.0}, {'current': 3266.213, 'min': 0.0, 'max': 0.0}, {'current': 3265.148, 'min': 0.0, 'max': 0.0}, {'current': 2554.152, 'min': 0.0, 'max': 0.0}, {'current': 2556.63, 'min': 0.0, 'max': 0.0}, {'current': 2540.571, 'min': 0.0, 'max': 0.0}, {'current': 2512.005, 'min': 0.0, 'max': 0.0}, {'current': 2560.992, 'min': 0.0, 'max': 0.0}, {'current': 2526.641, 'min': 0.0, 'max': 0.0}, {'current': 2528.267, 'min': 0.0, 'max': 0.0}, {'current': 2583.061, 'min': 0.0, 'max': 0.0}, {'current': 3257.652, 'min': 0.0, 'max': 0.0}, {'current': 3598.217, 'min': 0.0, 'max': 0.0}, {'current': 3282.578, 'min': 0.0, 'max': 0.0}, {'current': 3261.664, 'min': 0.0, 'max': 0.0}, {'current': 3268.246, 'min': 0.0, 'max': 0.0}, {'current': 3331.937, 'min': 0.0, 'max': 0.0}, {'current': 3265.603, 'min': 0.0, 'max': 0.0}, {'current': 3258.501, 'min': 0.0, 'max': 0.0}, {'current': 3337.045, 'min': 0.0, 'max': 0.0}, {'current': 3259.949, 'min': 0.0, 'max': 0.0}, {'current': 3226.058, 'min': 0.0, 'max': 0.0}, {'current': 3250.308, 'min': 0.0, 'max': 0.0}, {'current': 3208.026, 'min': 0.0, 'max': 0.0}, {'current': 2974.043, 'min': 0.0, 'max': 0.0}, {'current': 3216.706, 'min': 0.0, 'max': 0.0}, {'current': 3598.55, 'min': 0.0, 'max': 0.0}, {'current': 3598.261, 'min': 0.0, 'max': 0.0}, {'current': 3222.637, 'min': 0.0, 'max': 0.0}, {'current': 3227.662, 'min': 0.0, 'max': 0.0}, {'current': 3143.781, 'min': 0.0, 'max': 0.0}, {'current': 3215.567, 'min': 0.0, 'max': 0.0}, {'current': 3250.612, 'min': 0.0, 'max': 0.0}, {'current': 3195.784, 'min': 0.0, 'max': 0.0}, {'current': 3219.289, 'min': 0.0, 'max': 0.0}, {'current': 3289.163, 'min': 0.0, 'max': 0.0}, {'current': 3247.875, 'min': 0.0, 'max': 0.0}, {'current': 3249.112, 'min': 0.0, 'max': 0.0}, {'current': 3209.3, 'min': 0.0, 'max': 0.0}, {'current': 3187.212, 'min': 0.0, 'max': 0.0}, {'current': 3288.003, 'min': 0.0, 'max': 0.0}, {'current': 3169.417, 'min': 0.0, 'max': 0.0}, {'current': 3596.071, 'min': 0.0, 'max': 0.0}], 'disk': {'/': {'total': 119.94140625, 'used': 56.4015998840332}}, 'gpu': 'NVIDIA H100 80GB HBM3', 'gpu_count': 8, 'gpu_devices': [{'name': 'NVIDIA H100 80GB HBM3', 'memory_total': 85520809984}, {'name': 'NVIDIA H100 80GB HBM3', 'memory_total': 85520809984}, {'name': 'NVIDIA H100 80GB HBM3', 'memory_total': 85520809984}, {'name': 'NVIDIA H100 80GB HBM3', 'memory_total': 85520809984}, {'name': 'NVIDIA H100 80GB HBM3', 'memory_total': 85520809984}, {'name': 'NVIDIA H100 80GB HBM3', 'memory_total': 85520809984}, {'name': 'NVIDIA H100 80GB HBM3', 'memory_total': 85520809984}, {'name': 'NVIDIA H100 80GB HBM3', 'memory_total': 85520809984}], 'memory': {'total': 1999.9661331176758}}
26
+ 2024-11-13 19:02:47,022 INFO HandlerThread:1939 [system_monitor.py:probe():224] Finished collecting system info
27
+ 2024-11-13 19:02:47,022 INFO HandlerThread:1939 [system_monitor.py:probe():227] Publishing system info
28
+ 2024-11-13 19:02:47,022 DEBUG HandlerThread:1939 [system_info.py:_save_conda():208] Saving list of conda packages installed into the current environment
29
+ 2024-11-13 19:02:47,775 INFO Thread-12 :1939 [dir_watcher.py:_on_file_created():271] file/dir created: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/conda-environment.yaml
30
+ 2024-11-13 19:03:01,345 DEBUG HandlerThread:1939 [system_info.py:_save_conda():220] Saving conda packages done
31
+ 2024-11-13 19:03:01,345 DEBUG HandlerThread:1939 [system_info.py:_save_code():45] Saving code
32
+ 2024-11-13 19:03:01,349 DEBUG HandlerThread:1939 [system_info.py:_save_code():66] Saving code done
33
+ 2024-11-13 19:03:01,349 DEBUG HandlerThread:1939 [system_info.py:_save_patches():83] Saving git patches
34
+ 2024-11-13 19:03:01,776 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/conda-environment.yaml
35
+ 2024-11-13 19:03:01,776 INFO Thread-12 :1939 [dir_watcher.py:_on_file_created():271] file/dir created: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/code/qlearning_reasoning/training/sft.py
36
+ 2024-11-13 19:03:01,776 INFO Thread-12 :1939 [dir_watcher.py:_on_file_created():271] file/dir created: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/code/qlearning_reasoning
37
+ 2024-11-13 19:03:01,776 INFO Thread-12 :1939 [dir_watcher.py:_on_file_created():271] file/dir created: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/code/qlearning_reasoning/training
38
+ 2024-11-13 19:03:01,776 INFO Thread-12 :1939 [dir_watcher.py:_on_file_created():271] file/dir created: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/code
39
+ 2024-11-13 19:03:01,989 DEBUG HandlerThread:1939 [system_info.py:_save_patches():125] Saving git patches done
40
+ 2024-11-13 19:03:01,991 INFO HandlerThread:1939 [system_monitor.py:probe():229] Finished publishing system info
41
+ 2024-11-13 19:03:01,993 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
42
+ 2024-11-13 19:03:01,993 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: keepalive
43
+ 2024-11-13 19:03:01,994 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
44
+ 2024-11-13 19:03:01,994 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: keepalive
45
+ 2024-11-13 19:03:01,994 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
46
+ 2024-11-13 19:03:01,994 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: keepalive
47
+ 2024-11-13 19:03:01,994 DEBUG SenderThread:1939 [sender.py:send():382] send: files
48
+ 2024-11-13 19:03:01,995 INFO SenderThread:1939 [sender.py:_save_file():1403] saving file wandb-metadata.json with policy now
49
+ 2024-11-13 19:03:01,995 INFO SenderThread:1939 [sender.py:_save_file():1403] saving file code/qlearning_reasoning/training/sft.py with policy now
50
+ 2024-11-13 19:03:01,995 INFO SenderThread:1939 [sender.py:_save_file():1403] saving file diff.patch with policy now
51
+ 2024-11-13 19:03:01,999 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: python_packages
52
+ 2024-11-13 19:03:01,999 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: python_packages
53
+ 2024-11-13 19:03:02,001 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
54
+ 2024-11-13 19:03:02,003 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
55
+ 2024-11-13 19:03:02,003 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
56
+ 2024-11-13 19:03:02,111 DEBUG SenderThread:1939 [sender.py:send():382] send: telemetry
57
+ 2024-11-13 19:03:02,111 DEBUG SenderThread:1939 [sender.py:send():382] send: config
58
+ 2024-11-13 19:03:02,112 DEBUG SenderThread:1939 [sender.py:send():382] send: metric
59
+ 2024-11-13 19:03:02,112 DEBUG SenderThread:1939 [sender.py:send():382] send: telemetry
60
+ 2024-11-13 19:03:02,112 DEBUG SenderThread:1939 [sender.py:send():382] send: metric
61
+ 2024-11-13 19:03:02,113 WARNING SenderThread:1939 [sender.py:send_metric():1354] Seen metric with glob (shouldn't happen)
62
+ 2024-11-13 19:03:02,113 DEBUG SenderThread:1939 [sender.py:send():382] send: telemetry
63
+ 2024-11-13 19:03:02,113 DEBUG SenderThread:1939 [sender.py:send():382] send: telemetry
64
+ 2024-11-13 19:03:02,113 DEBUG SenderThread:1939 [sender.py:send():382] send: config
65
+ 2024-11-13 19:03:02,551 INFO wandb-upload_1:1939 [upload_job.py:push():131] Uploaded file /tmp/tmpm_h7h9v0wandb/pjhigh2e-code/qlearning_reasoning/training/sft.py
66
+ 2024-11-13 19:03:02,554 INFO wandb-upload_0:1939 [upload_job.py:push():131] Uploaded file /tmp/tmpm_h7h9v0wandb/f98paeu8-wandb-metadata.json
67
+ 2024-11-13 19:03:02,556 INFO wandb-upload_2:1939 [upload_job.py:push():131] Uploaded file /tmp/tmpm_h7h9v0wandb/cc8zuviy-diff.patch
68
+ 2024-11-13 19:03:02,776 INFO Thread-12 :1939 [dir_watcher.py:_on_file_created():271] file/dir created: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/diff.patch
69
+ 2024-11-13 19:03:02,776 INFO Thread-12 :1939 [dir_watcher.py:_on_file_created():271] file/dir created: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
70
+ 2024-11-13 19:03:02,777 INFO Thread-12 :1939 [dir_watcher.py:_on_file_created():271] file/dir created: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/requirements.txt
71
+ 2024-11-13 19:03:02,777 INFO Thread-12 :1939 [dir_watcher.py:_on_file_created():271] file/dir created: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/wandb-metadata.json
72
+ 2024-11-13 19:03:04,777 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
73
+ 2024-11-13 19:03:07,114 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
74
+ 2024-11-13 19:03:12,115 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
75
+ 2024-11-13 19:03:16,999 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
76
+ 2024-11-13 19:03:17,000 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
77
+ 2024-11-13 19:03:17,002 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
78
+ 2024-11-13 19:03:18,090 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
79
+ 2024-11-13 19:03:18,782 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/config.yaml
80
+ 2024-11-13 19:03:23,250 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
81
+ 2024-11-13 19:03:26,095 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: partial_history
82
+ 2024-11-13 19:03:26,097 DEBUG SenderThread:1939 [sender.py:send():382] send: metric
83
+ 2024-11-13 19:03:26,097 DEBUG SenderThread:1939 [sender.py:send():382] send: metric
84
+ 2024-11-13 19:03:26,097 DEBUG SenderThread:1939 [sender.py:send():382] send: metric
85
+ 2024-11-13 19:03:26,097 DEBUG SenderThread:1939 [sender.py:send():382] send: metric
86
+ 2024-11-13 19:03:26,097 DEBUG SenderThread:1939 [sender.py:send():382] send: history
87
+ 2024-11-13 19:03:26,098 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: summary_record
88
+ 2024-11-13 19:03:26,099 INFO SenderThread:1939 [sender.py:_save_file():1403] saving file wandb-summary.json with policy end
89
+ 2024-11-13 19:03:26,785 INFO Thread-12 :1939 [dir_watcher.py:_on_file_created():271] file/dir created: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/wandb-summary.json
90
+ 2024-11-13 19:03:28,785 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
91
+ 2024-11-13 19:03:29,100 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
92
+ 2024-11-13 19:03:31,999 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
93
+ 2024-11-13 19:03:32,000 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
94
+ 2024-11-13 19:03:32,002 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
95
+ 2024-11-13 19:03:34,147 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
96
+ 2024-11-13 19:03:39,147 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
97
+ 2024-11-13 19:03:44,148 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
98
+ 2024-11-13 19:03:46,955 DEBUG SystemMonitor:1939 [system_monitor.py:_start():172] Starting system metrics aggregation loop
99
+ 2024-11-13 19:03:46,959 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
100
+ 2024-11-13 19:03:47,000 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
101
+ 2024-11-13 19:03:47,000 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
102
+ 2024-11-13 19:03:47,043 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
103
+ 2024-11-13 19:03:49,636 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
104
+ 2024-11-13 19:03:50,813 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
105
+ 2024-11-13 19:03:50,814 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/config.yaml
106
+ 2024-11-13 19:03:54,814 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
107
+ 2024-11-13 19:03:59,814 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
108
+ 2024-11-13 19:04:02,000 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
109
+ 2024-11-13 19:04:02,000 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
110
+ 2024-11-13 19:04:02,002 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
111
+ 2024-11-13 19:04:05,061 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
112
+ 2024-11-13 19:04:10,062 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
113
+ 2024-11-13 19:04:14,822 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
114
+ 2024-11-13 19:04:15,419 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
115
+ 2024-11-13 19:04:16,962 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
116
+ 2024-11-13 19:04:17,000 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
117
+ 2024-11-13 19:04:17,000 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
118
+ 2024-11-13 19:04:17,002 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
119
+ 2024-11-13 19:04:21,087 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
120
+ 2024-11-13 19:04:26,088 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
121
+ 2024-11-13 19:04:31,089 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
122
+ 2024-11-13 19:04:32,000 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
123
+ 2024-11-13 19:04:32,000 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
124
+ 2024-11-13 19:04:32,043 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
125
+ 2024-11-13 19:04:36,587 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
126
+ 2024-11-13 19:04:36,829 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
127
+ 2024-11-13 19:04:41,588 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
128
+ 2024-11-13 19:04:46,588 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
129
+ 2024-11-13 19:04:46,965 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
130
+ 2024-11-13 19:04:47,000 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
131
+ 2024-11-13 19:04:47,000 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
132
+ 2024-11-13 19:04:47,002 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
133
+ 2024-11-13 19:04:52,111 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
134
+ 2024-11-13 19:04:57,111 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
135
+ 2024-11-13 19:04:57,383 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: partial_history
136
+ 2024-11-13 19:04:57,384 DEBUG SenderThread:1939 [sender.py:send():382] send: history
137
+ 2024-11-13 19:04:57,385 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: summary_record
138
+ 2024-11-13 19:04:57,385 INFO SenderThread:1939 [sender.py:_save_file():1403] saving file wandb-summary.json with policy end
139
+ 2024-11-13 19:04:57,837 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/wandb-summary.json
140
+ 2024-11-13 19:04:58,837 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
141
+ 2024-11-13 19:04:59,002 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: partial_history
142
+ 2024-11-13 19:04:59,004 DEBUG SenderThread:1939 [sender.py:send():382] send: metric
143
+ 2024-11-13 19:04:59,005 DEBUG SenderThread:1939 [sender.py:send():382] send: metric
144
+ 2024-11-13 19:04:59,005 DEBUG SenderThread:1939 [sender.py:send():382] send: metric
145
+ 2024-11-13 19:04:59,005 DEBUG SenderThread:1939 [sender.py:send():382] send: metric
146
+ 2024-11-13 19:04:59,006 DEBUG SenderThread:1939 [sender.py:send():382] send: history
147
+ 2024-11-13 19:04:59,006 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: summary_record
148
+ 2024-11-13 19:04:59,006 INFO SenderThread:1939 [sender.py:_save_file():1403] saving file wandb-summary.json with policy end
149
+ 2024-11-13 19:04:59,838 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/wandb-summary.json
150
+ 2024-11-13 19:05:00,838 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
151
+ 2024-11-13 19:05:02,000 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
152
+ 2024-11-13 19:05:02,001 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
153
+ 2024-11-13 19:05:02,002 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
154
+ 2024-11-13 19:05:03,111 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
155
+ 2024-11-13 19:05:08,112 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
156
+ 2024-11-13 19:05:13,113 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
157
+ 2024-11-13 19:05:16,967 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
158
+ 2024-11-13 19:05:17,000 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
159
+ 2024-11-13 19:05:17,001 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
160
+ 2024-11-13 19:05:17,043 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
161
+ 2024-11-13 19:05:18,123 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
162
+ 2024-11-13 19:05:22,846 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
163
+ 2024-11-13 19:05:23,982 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
164
+ 2024-11-13 19:05:25,847 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/config.yaml
165
+ 2024-11-13 19:05:30,486 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
166
+ 2024-11-13 19:05:32,000 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
167
+ 2024-11-13 19:05:32,001 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
168
+ 2024-11-13 19:05:32,002 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
169
+ 2024-11-13 19:05:36,117 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
170
+ 2024-11-13 19:05:41,118 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
171
+ 2024-11-13 19:05:46,513 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
172
+ 2024-11-13 19:05:46,854 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
173
+ 2024-11-13 19:05:46,970 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
174
+ 2024-11-13 19:05:47,000 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
175
+ 2024-11-13 19:05:47,001 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
176
+ 2024-11-13 19:05:47,003 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
177
+ 2024-11-13 19:05:52,097 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
178
+ 2024-11-13 19:05:57,098 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
179
+ 2024-11-13 19:06:02,001 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
180
+ 2024-11-13 19:06:02,001 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
181
+ 2024-11-13 19:06:02,043 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
182
+ 2024-11-13 19:06:02,152 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
183
+ 2024-11-13 19:06:07,153 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
184
+ 2024-11-13 19:06:08,862 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
185
+ 2024-11-13 19:06:12,639 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
186
+ 2024-11-13 19:06:16,973 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
187
+ 2024-11-13 19:06:17,001 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
188
+ 2024-11-13 19:06:17,002 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
189
+ 2024-11-13 19:06:17,004 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
190
+ 2024-11-13 19:06:18,127 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
191
+ 2024-11-13 19:06:23,127 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
192
+ 2024-11-13 19:06:28,128 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
193
+ 2024-11-13 19:06:30,870 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
194
+ 2024-11-13 19:06:32,001 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
195
+ 2024-11-13 19:06:32,001 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
196
+ 2024-11-13 19:06:32,003 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
197
+ 2024-11-13 19:06:33,143 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
198
+ 2024-11-13 19:06:38,144 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
199
+ 2024-11-13 19:06:43,144 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
200
+ 2024-11-13 19:06:46,975 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
201
+ 2024-11-13 19:06:47,001 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
202
+ 2024-11-13 19:06:47,001 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
203
+ 2024-11-13 19:06:47,043 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
204
+ 2024-11-13 19:06:49,137 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
205
+ 2024-11-13 19:06:51,931 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: partial_history
206
+ 2024-11-13 19:06:51,932 DEBUG SenderThread:1939 [sender.py:send():382] send: history
207
+ 2024-11-13 19:06:51,933 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: summary_record
208
+ 2024-11-13 19:06:51,934 INFO SenderThread:1939 [sender.py:_save_file():1403] saving file wandb-summary.json with policy end
209
+ 2024-11-13 19:06:52,878 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
210
+ 2024-11-13 19:06:52,878 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/wandb-summary.json
211
+ 2024-11-13 19:06:53,541 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: partial_history
212
+ 2024-11-13 19:06:53,543 DEBUG SenderThread:1939 [sender.py:send():382] send: history
213
+ 2024-11-13 19:06:53,543 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: summary_record
214
+ 2024-11-13 19:06:53,544 INFO SenderThread:1939 [sender.py:_save_file():1403] saving file wandb-summary.json with policy end
215
+ 2024-11-13 19:06:53,878 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/wandb-summary.json
216
+ 2024-11-13 19:06:54,545 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
217
+ 2024-11-13 19:06:54,878 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
218
+ 2024-11-13 19:06:59,546 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
219
+ 2024-11-13 19:07:02,001 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
220
+ 2024-11-13 19:07:02,002 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
221
+ 2024-11-13 19:07:02,004 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
222
+ 2024-11-13 19:07:05,164 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
223
+ 2024-11-13 19:07:10,165 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
224
+ 2024-11-13 19:07:15,165 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
225
+ 2024-11-13 19:07:16,886 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
226
+ 2024-11-13 19:07:16,978 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
227
+ 2024-11-13 19:07:17,001 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
228
+ 2024-11-13 19:07:17,002 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
229
+ 2024-11-13 19:07:17,003 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
230
+ 2024-11-13 19:07:21,108 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
231
+ 2024-11-13 19:07:26,108 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
232
+ 2024-11-13 19:07:31,109 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
233
+ 2024-11-13 19:07:32,001 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
234
+ 2024-11-13 19:07:32,002 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
235
+ 2024-11-13 19:07:32,043 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
236
+ 2024-11-13 19:07:36,934 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
237
+ 2024-11-13 19:07:38,894 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
238
+ 2024-11-13 19:07:41,935 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
239
+ 2024-11-13 19:07:46,935 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
240
+ 2024-11-13 19:07:46,981 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
241
+ 2024-11-13 19:07:47,002 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
242
+ 2024-11-13 19:07:47,002 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
243
+ 2024-11-13 19:07:47,004 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
244
+ 2024-11-13 19:07:52,135 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
245
+ 2024-11-13 19:07:57,136 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
246
+ 2024-11-13 19:08:02,136 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
247
+ 2024-11-13 19:08:02,856 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
248
+ 2024-11-13 19:08:02,923 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
249
+ 2024-11-13 19:08:02,924 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
250
+ 2024-11-13 19:08:08,076 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
251
+ 2024-11-13 19:08:13,077 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
252
+ 2024-11-13 19:08:16,985 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
253
+ 2024-11-13 19:08:17,407 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
254
+ 2024-11-13 19:08:17,407 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
255
+ 2024-11-13 19:08:17,447 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
256
+ 2024-11-13 19:08:18,545 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
257
+ 2024-11-13 19:08:23,546 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
258
+ 2024-11-13 19:08:28,546 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
259
+ 2024-11-13 19:08:32,914 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
260
+ 2024-11-13 19:08:33,843 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
261
+ 2024-11-13 19:08:34,729 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
262
+ 2024-11-13 19:08:34,729 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
263
+ 2024-11-13 19:08:34,771 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
264
+ 2024-11-13 19:08:39,802 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
265
+ 2024-11-13 19:08:44,803 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
266
+ 2024-11-13 19:08:46,987 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
267
+ 2024-11-13 19:08:48,007 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
268
+ 2024-11-13 19:08:48,007 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
269
+ 2024-11-13 19:08:48,008 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
270
+ 2024-11-13 19:08:48,921 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
271
+ 2024-11-13 19:08:49,945 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
272
+ 2024-11-13 19:08:50,922 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
273
+ 2024-11-13 19:08:52,923 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
274
+ 2024-11-13 19:08:54,924 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
275
+ 2024-11-13 19:08:55,154 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
276
+ 2024-11-13 19:08:56,925 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
277
+ 2024-11-13 19:08:58,926 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
278
+ 2024-11-13 19:09:00,371 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
279
+ 2024-11-13 19:09:00,948 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
280
+ 2024-11-13 19:09:03,950 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
281
+ 2024-11-13 19:09:05,541 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
282
+ 2024-11-13 19:09:05,782 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
283
+ 2024-11-13 19:09:06,102 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
284
+ 2024-11-13 19:09:06,102 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
285
+ 2024-11-13 19:09:11,263 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
286
+ 2024-11-13 19:09:16,264 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
287
+ 2024-11-13 19:09:16,990 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
288
+ 2024-11-13 19:09:20,961 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
289
+ 2024-11-13 19:09:22,073 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
290
+ 2024-11-13 19:09:22,272 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
291
+ 2024-11-13 19:09:22,272 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
292
+ 2024-11-13 19:09:27,401 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
293
+ 2024-11-13 19:09:32,401 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
294
+ 2024-11-13 19:09:36,126 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
295
+ 2024-11-13 19:09:36,666 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
296
+ 2024-11-13 19:09:36,666 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
297
+ 2024-11-13 19:09:38,052 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
298
+ 2024-11-13 19:09:39,964 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
299
+ 2024-11-13 19:09:43,769 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
300
+ 2024-11-13 19:09:46,992 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
301
+ 2024-11-13 19:09:47,846 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
302
+ 2024-11-13 19:09:47,847 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
303
+ 2024-11-13 19:09:47,847 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
304
+ 2024-11-13 19:09:49,230 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
305
+ 2024-11-13 19:09:54,231 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
306
+ 2024-11-13 19:09:59,231 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
307
+ 2024-11-13 19:10:02,847 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
308
+ 2024-11-13 19:10:02,848 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
309
+ 2024-11-13 19:10:02,848 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
310
+ 2024-11-13 19:10:05,230 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
311
+ 2024-11-13 19:10:10,231 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
312
+ 2024-11-13 19:10:11,977 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
313
+ 2024-11-13 19:10:15,248 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
314
+ 2024-11-13 19:10:16,996 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
315
+ 2024-11-13 19:10:17,847 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
316
+ 2024-11-13 19:10:17,848 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
317
+ 2024-11-13 19:10:17,848 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
318
+ 2024-11-13 19:10:21,230 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
319
+ 2024-11-13 19:10:26,231 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
320
+ 2024-11-13 19:10:31,231 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
321
+ 2024-11-13 19:10:32,847 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
322
+ 2024-11-13 19:10:32,847 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
323
+ 2024-11-13 19:10:32,847 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
324
+ 2024-11-13 19:10:33,985 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
325
+ 2024-11-13 19:10:37,230 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
326
+ 2024-11-13 19:10:42,231 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
327
+ 2024-11-13 19:10:46,998 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
328
+ 2024-11-13 19:10:47,847 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
329
+ 2024-11-13 19:10:47,847 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
330
+ 2024-11-13 19:10:47,887 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
331
+ 2024-11-13 19:10:47,921 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
332
+ 2024-11-13 19:10:53,230 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
333
+ 2024-11-13 19:10:53,575 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: partial_history
334
+ 2024-11-13 19:10:53,576 DEBUG SenderThread:1939 [sender.py:send():382] send: history
335
+ 2024-11-13 19:10:53,576 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: summary_record
336
+ 2024-11-13 19:10:53,577 INFO SenderThread:1939 [sender.py:_save_file():1403] saving file wandb-summary.json with policy end
337
+ 2024-11-13 19:10:53,992 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/wandb-summary.json
338
+ 2024-11-13 19:10:54,993 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
339
+ 2024-11-13 19:10:55,200 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: partial_history
340
+ 2024-11-13 19:10:55,201 DEBUG SenderThread:1939 [sender.py:send():382] send: history
341
+ 2024-11-13 19:10:55,201 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: summary_record
342
+ 2024-11-13 19:10:55,203 INFO SenderThread:1939 [sender.py:_save_file():1403] saving file wandb-summary.json with policy end
343
+ 2024-11-13 19:10:55,993 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
344
+ 2024-11-13 19:10:55,994 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/wandb-summary.json
345
+ 2024-11-13 19:10:56,994 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
346
+ 2024-11-13 19:10:57,994 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
347
+ 2024-11-13 19:10:59,204 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
348
+ 2024-11-13 19:11:02,848 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
349
+ 2024-11-13 19:11:02,848 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
350
+ 2024-11-13 19:11:02,849 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
351
+ 2024-11-13 19:11:04,230 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
352
+ 2024-11-13 19:11:09,231 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
353
+ 2024-11-13 19:11:14,231 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
354
+ 2024-11-13 19:11:17,000 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
355
+ 2024-11-13 19:11:17,847 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
356
+ 2024-11-13 19:11:17,848 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
357
+ 2024-11-13 19:11:17,848 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
358
+ 2024-11-13 19:11:19,311 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
359
+ 2024-11-13 19:11:20,004 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
360
+ 2024-11-13 19:11:24,312 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
361
+ 2024-11-13 19:11:29,312 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
362
+ 2024-11-13 19:11:32,848 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
363
+ 2024-11-13 19:11:32,848 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
364
+ 2024-11-13 19:11:32,849 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
365
+ 2024-11-13 19:11:35,231 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
366
+ 2024-11-13 19:11:40,232 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
367
+ 2024-11-13 19:11:44,014 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
368
+ 2024-11-13 19:11:46,173 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
369
+ 2024-11-13 19:11:47,003 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
370
+ 2024-11-13 19:11:47,847 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
371
+ 2024-11-13 19:11:47,847 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
372
+ 2024-11-13 19:11:47,848 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
373
+ 2024-11-13 19:11:51,230 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/logs/debug.log ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-11-13 19:02:46,428 INFO MainThread:173 [wandb_setup.py:_flush():76] Current SDK version is 0.16.3
2
+ 2024-11-13 19:02:46,428 INFO MainThread:173 [wandb_setup.py:_flush():76] Configure stats pid to 173
3
+ 2024-11-13 19:02:46,428 INFO MainThread:173 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
4
+ 2024-11-13 19:02:46,428 INFO MainThread:173 [wandb_setup.py:_flush():76] Loading settings from /opt/ml/code/wandb/settings
5
+ 2024-11-13 19:02:46,428 INFO MainThread:173 [wandb_setup.py:_flush():76] Loading settings from environment variables: {'root_dir': '/opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048', 'api_key': '***REDACTED***', 'project': 'sft-codecontests-1112', 'run_name': 'sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048'}
6
+ 2024-11-13 19:02:46,428 INFO MainThread:173 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
7
+ 2024-11-13 19:02:46,428 INFO MainThread:173 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': 'qlearning_reasoning/training/sft.py', 'program_abspath': '/opt/ml/code/qlearning_reasoning/training/sft.py', 'program': '/opt/ml/code/qlearning_reasoning/training/sft.py'}
8
+ 2024-11-13 19:02:46,428 INFO MainThread:173 [wandb_init.py:_log_setup():526] Logging user logs to /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/logs/debug.log
9
+ 2024-11-13 19:02:46,428 INFO MainThread:173 [wandb_init.py:_log_setup():527] Logging internal logs to /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/logs/debug-internal.log
10
+ 2024-11-13 19:02:46,428 INFO MainThread:173 [wandb_init.py:init():566] calling init triggers
11
+ 2024-11-13 19:02:46,428 INFO MainThread:173 [wandb_init.py:init():573] wandb.init called with sweep_config: {}
12
+ config: {'sagemaker_training_job_name': 'kushalarora-rvv-main-2024-11-13-17-55-42-664', 'batch_size': 64, 'cache_dir': '/opt/ml/data/input/.cache', 'dataset': 'code-contests', 'gradient_accumulation_steps': 8, 'logging_steps': 5, 'lr': '1e-6', 'max_seq_length': 2048, 'model_name': 'Qwen/Qwen2.5-Coder-7B-Instruct', 'num_train_epochs': 2, 'output_dir': '/opt/ml/output', 'push_to_hub': 'true', 'run_name': 'sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048', 'sagemaker_container_log_level': 20, 'sagemaker_instance_type': 'ml.p5.48xlarge', 'sagemaker_job_name': 'kushalarora-rvv-main-2024-11-13-17-55-42-664', 'sagemaker_program': 'sft.py', 'sagemaker_region': 'us-east-1', 'sagemaker_submit_directory': 's3://tri-ml-datasets/kushalarora-rvv-main-2024-11-13-17-55-42-664/source/sourcedir.tar.gz', 'sagemaker_torch_distributed_enabled': 'true', 'save_steps': 100, 'schedule': 'cosine', 'wandb_project': 'sft-codecontests-1112', 'weight_decay': 0.0, 'additional_framework_parameters': {'sagemaker_instance_type': 'ml.p5.48xlarge', 'sagemaker_torch_distributed_enabled': True}, 'channel_input_dirs': {}, 'current_host': 'algo-1', 'current_instance_group': 'homogeneousCluster', 'current_instance_group_hosts': ['algo-1'], 'current_instance_type': 'ml.p5.48xlarge', 'distribution_hosts': ['algo-1'], 'distribution_instance_groups': ['homogeneousCluster'], 'framework_module': 'sagemaker_pytorch_container.training:main', 'hosts': ['algo-1'], 'hyperparameters': {'batch_size': 64, 'cache_dir': '/opt/ml/data/input/.cache', 'dataset': 'code-contests', 'gradient_accumulation_steps': 8, 'logging_steps': 5, 'lr': '1e-6', 'max_seq_length': 2048, 'model_name': 'Qwen/Qwen2.5-Coder-7B-Instruct', 'num_train_epochs': 2, 'output_dir': '/opt/ml/model/', 'push_to_hub': True, 'run_name': 'sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048', 'save_steps': 100, 'schedule': 'cosine', 'wandb_project': 'sft-codecontests-1112', 'weight_decay': 0.0}, 'input_config_dir': '/opt/ml/input/config', 'input_data_config': {}, 'input_dir': '/opt/ml/input', 'instance_groups': ['homogeneousCluster'], 'instance_groups_dict': {'homogeneousCluster': {'hosts': ['algo-1'], 'instance_group_name': 'homogeneousCluster', 'instance_type': 'ml.p5.48xlarge'}}, 'is_hetero': False, 'is_master': True, 'is_modelparallel_enabled': None, 'is_smddpmprun_installed': False, 'is_smddprun_installed': True, 'job_name': 'kushalarora-rvv-main-2024-11-13-17-55-42-664', 'log_level': 20, 'master_hostname': 'algo-1', 'model_dir': '/opt/ml/model', 'module_dir': 's3://tri-ml-datasets/kushalarora-rvv-main-2024-11-13-17-55-42-664/source/sourcedir.tar.gz', 'module_name': '/opt/ml/code/qlearning_reasoning/training/sft', 'network_interface_name': 'eth0', 'num_cpus': 192, 'num_gpus': 8, 'num_neurons': 0, 'output_data_dir': '/opt/ml/output/data', 'output_intermediate_dir': '/opt/ml/output/intermediate', 'resource_config': {'current_group_name': 'homogeneousCluster', 'current_host': 'algo-1', 'current_instance_type': 'ml.p5.48xlarge', 'hosts': ['algo-1'], 'instance_groups': [{'hosts': ['algo-1'], 'instance_group_name': 'homogeneousCluster', 'instance_type': 'ml.p5.48xlarge'}], 'network_interface_name': 'eth0'}, 'user_entry_point': '/opt/ml/code/qlearning_reasoning/training/sft.py'}
13
+ 2024-11-13 19:02:46,428 INFO MainThread:173 [wandb_init.py:init():616] starting backend
14
+ 2024-11-13 19:02:46,428 INFO MainThread:173 [wandb_init.py:init():620] setting up manager
15
+ 2024-11-13 19:02:46,429 INFO MainThread:173 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
16
+ 2024-11-13 19:02:46,431 INFO MainThread:173 [wandb_init.py:init():628] backend started and connected
17
+ 2024-11-13 19:02:46,434 INFO MainThread:173 [wandb_init.py:init():720] updated telemetry
18
+ 2024-11-13 19:02:46,440 INFO MainThread:173 [wandb_init.py:init():753] communicating run to backend with 90.0 second timeout
19
+ 2024-11-13 19:02:46,781 INFO MainThread:173 [wandb_run.py:_on_init():2262] communicating current version
20
+ 2024-11-13 19:02:46,812 INFO MainThread:173 [wandb_run.py:_on_init():2271] got version response upgrade_message: "wandb version 0.18.6 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
21
+
22
+ 2024-11-13 19:02:46,813 INFO MainThread:173 [wandb_init.py:init():804] starting run threads in backend
23
+ 2024-11-13 19:03:02,000 INFO MainThread:173 [wandb_run.py:_console_start():2241] atexit reg
24
+ 2024-11-13 19:03:02,000 INFO MainThread:173 [wandb_run.py:_redirect():2096] redirect: wrap_raw
25
+ 2024-11-13 19:03:02,000 INFO MainThread:173 [wandb_run.py:_redirect():2161] Wrapping output streams.
26
+ 2024-11-13 19:03:02,001 INFO MainThread:173 [wandb_run.py:_redirect():2186] Redirects installed.
27
+ 2024-11-13 19:03:02,001 INFO MainThread:173 [wandb_init.py:init():847] run started, returning control to user process
28
+ 2024-11-13 19:03:02,004 INFO MainThread:173 [wandb_run.py:_config_callback():1343] config_cb None None {'vocab_size': 151671, 'max_position_embeddings': 32768, 'hidden_size': 3584, 'intermediate_size': 18944, 'num_hidden_layers': 28, 'num_attention_heads': 28, 'use_sliding_window': False, 'sliding_window': None, 'max_window_layers': 28, 'num_key_value_heads': 4, 'hidden_act': 'silu', 'initializer_range': 0.02, 'rms_norm_eps': 1e-06, 'use_cache': False, 'rope_theta': 1000000.0, 'rope_scaling': None, 'attention_dropout': 0.0, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'bfloat16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['Qwen2ForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': 151643, 'eos_token_id': 151645, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'Qwen/Qwen2.5-Coder-7B-Instruct', '_attn_implementation_autoset': True, 'transformers_version': '4.46.0', 'model_type': 'qwen2', 'output_dir': '/opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'eval_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 8, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 8, 'eval_accumulation_steps': None, 'eval_delay': 0, 'torch_empty_cache_steps': None, 'learning_rate': 1e-06, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 2, 'max_steps': -1, 'lr_scheduler_type': 'cosine', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/runs/Nov13_19-02-15_algo-1', 'logging_strategy': 'steps', 'logging_first_step': True, 'logging_steps': 5, 'logging_nan_inf_filter': True, 'save_strategy': 'epoch', 'save_steps': 100, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'restore_callback_states_from_checkpoint': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': 5, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': 'sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': ['full_shard', 'auto_wrap'], 'fsdp_min_num_params': 0, 'fsdp_config': {'activation_checkpointing': False, 'auto_wrap_policy': 'TRANSFORMER_BASED_WRAP', 'backward_prefetch': 'BACKWARD_PRE', 'cpu_ram_efficient_loading': True, 'forward_prefetch': False, 'offload_params': False, 'sharding_strategy': 'FULL_SHARD', 'state_dict_type': 'FULL_STATE_DICT', 'sync_module_states': True, 'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': 'sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_', 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': {'use_reentrant': False}, 'include_inputs_for_metrics': False, 'include_for_metrics': [], 'eval_do_concat_batches': True, 'fp16_backend': 'auto', 'evaluation_strategy': None, 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None, 'batch_eval_metrics': False, 'eval_on_start': False, 'use_liger_kernel': False, 'eval_use_gather_object': False, 'dataset_text_field': 'text', 'packing': False, 'max_seq_length': 2048, 'dataset_num_proc': None, 'dataset_batch_size': 64, 'model_init_kwargs': None, 'dataset_kwargs': {}, 'eval_packing': None, 'num_of_sequences': 1024, 'chars_per_token': '<CHARS_PER_TOKEN>', 'use_liger': False}
29
+ 2024-11-13 19:03:02,008 INFO MainThread:173 [wandb_config.py:__setitem__():151] config set model/num_parameters = 951599936 - <bound method Run._config_callback of <wandb.sdk.wandb_run.Run object at 0x7fb464957310>>
30
+ 2024-11-13 19:03:02,009 INFO MainThread:173 [wandb_run.py:_config_callback():1343] config_cb model/num_parameters 951599936 None
wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/run-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1.wandb ADDED
Binary file (392 kB). View file