kushal-tri
commited on
Commit
•
5a60e84
1
Parent(s):
2fe0f98
Training in progress, epoch 0
Browse files- .gitattributes +1 -0
- added_tokens.json +30 -0
- config.json +29 -0
- merges.txt +0 -0
- model-00001-of-00004.safetensors +3 -0
- model-00002-of-00004.safetensors +3 -0
- model-00003-of-00004.safetensors +3 -0
- model-00004-of-00004.safetensors +3 -0
- model.safetensors.index.json +346 -0
- special_tokens_map.json +60 -0
- tokenizer.json +3 -0
- tokenizer_config.json +248 -0
- training_args.bin +3 -0
- vocab.json +0 -0
- wandb/debug-internal.log +373 -0
- wandb/debug.log +30 -0
- wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/code/qlearning_reasoning/training/sft.py +249 -0
- wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/conda-environment.yaml +515 -0
- wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/config.yaml +948 -0
- wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/diff.patch +49 -0
- wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log +0 -0
- wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/requirements.txt +289 -0
- wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/wandb-metadata.json +1069 -0
- wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/wandb-summary.json +1 -0
- wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/logs/debug-internal.log +373 -0
- wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/logs/debug.log +30 -0
- wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/run-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1.wandb +0 -0
.gitattributes
CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
36 |
+
tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
added_tokens.json
ADDED
@@ -0,0 +1,30 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"</tool_call>": 151658,
|
3 |
+
"<tool_call>": 151657,
|
4 |
+
"<|box_end|>": 151649,
|
5 |
+
"<|box_start|>": 151648,
|
6 |
+
"<|endoftext|>": 151643,
|
7 |
+
"<|file_sep|>": 151664,
|
8 |
+
"<|fim_middle|>": 151660,
|
9 |
+
"<|fim_pad|>": 151662,
|
10 |
+
"<|fim_prefix|>": 151659,
|
11 |
+
"<|fim_suffix|>": 151661,
|
12 |
+
"<|im_end|>": 151645,
|
13 |
+
"<|im_start|>": 151644,
|
14 |
+
"<|image_pad|>": 151655,
|
15 |
+
"<|object_ref_end|>": 151647,
|
16 |
+
"<|object_ref_start|>": 151646,
|
17 |
+
"<|quad_end|>": 151651,
|
18 |
+
"<|quad_start|>": 151650,
|
19 |
+
"<|repo_name|>": 151663,
|
20 |
+
"<|video_pad|>": 151656,
|
21 |
+
"<|vision_end|>": 151653,
|
22 |
+
"<|vision_pad|>": 151654,
|
23 |
+
"<|vision_start|>": 151652,
|
24 |
+
"[/REVISION]": 151670,
|
25 |
+
"[/STEP]": 151666,
|
26 |
+
"[/TURN]": 151668,
|
27 |
+
"[REVISION]": 151669,
|
28 |
+
"[STEP]": 151665,
|
29 |
+
"[TURN]": 151667
|
30 |
+
}
|
config.json
ADDED
@@ -0,0 +1,29 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "Qwen/Qwen2.5-Coder-7B-Instruct",
|
3 |
+
"architectures": [
|
4 |
+
"Qwen2ForCausalLM"
|
5 |
+
],
|
6 |
+
"attention_dropout": 0.0,
|
7 |
+
"eos_token_id": 151645,
|
8 |
+
"hidden_act": "silu",
|
9 |
+
"hidden_size": 3584,
|
10 |
+
"initializer_range": 0.02,
|
11 |
+
"intermediate_size": 18944,
|
12 |
+
"max_position_embeddings": 32768,
|
13 |
+
"max_window_layers": 28,
|
14 |
+
"model_type": "qwen2",
|
15 |
+
"num_attention_heads": 28,
|
16 |
+
"num_hidden_layers": 28,
|
17 |
+
"num_key_value_heads": 4,
|
18 |
+
"pad_token_id": 151643,
|
19 |
+
"rms_norm_eps": 1e-06,
|
20 |
+
"rope_scaling": null,
|
21 |
+
"rope_theta": 1000000.0,
|
22 |
+
"sliding_window": null,
|
23 |
+
"tie_word_embeddings": false,
|
24 |
+
"torch_dtype": "bfloat16",
|
25 |
+
"transformers_version": "4.46.0",
|
26 |
+
"use_cache": false,
|
27 |
+
"use_sliding_window": false,
|
28 |
+
"vocab_size": 151671
|
29 |
+
}
|
merges.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|
model-00001-of-00004.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a98b4e3194379ba7bec05ef5af0baa479391fb72c51b261749bdd1d251fb9b81
|
3 |
+
size 4874843752
|
model-00002-of-00004.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ab9e2130ff05c8efc6ce715675f96987ccecebb141684c969aa92ece6777278a
|
3 |
+
size 4932751008
|
model-00003-of-00004.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:181225151db82d7c2a4b0d6bf1b800f2aa0b9e19ebcbc3f666e96ff341cd3a0b
|
3 |
+
size 4330865200
|
model-00004-of-00004.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e9fd93165bb07b6dbb7f816180d4f65ad089d8c4efe19e5171a47c0c751f84b0
|
3 |
+
size 1087177856
|
model.safetensors.index.json
ADDED
@@ -0,0 +1,346 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"metadata": {
|
3 |
+
"total_size": 15225598976
|
4 |
+
},
|
5 |
+
"weight_map": {
|
6 |
+
"lm_head.weight": "model-00004-of-00004.safetensors",
|
7 |
+
"model.embed_tokens.weight": "model-00001-of-00004.safetensors",
|
8 |
+
"model.layers.0.input_layernorm.weight": "model-00001-of-00004.safetensors",
|
9 |
+
"model.layers.0.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
|
10 |
+
"model.layers.0.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
|
11 |
+
"model.layers.0.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
|
12 |
+
"model.layers.0.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
|
13 |
+
"model.layers.0.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
|
14 |
+
"model.layers.0.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
|
15 |
+
"model.layers.0.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
|
16 |
+
"model.layers.0.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
|
17 |
+
"model.layers.0.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
|
18 |
+
"model.layers.0.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
|
19 |
+
"model.layers.0.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
|
20 |
+
"model.layers.1.input_layernorm.weight": "model-00001-of-00004.safetensors",
|
21 |
+
"model.layers.1.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
|
22 |
+
"model.layers.1.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
|
23 |
+
"model.layers.1.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
|
24 |
+
"model.layers.1.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
|
25 |
+
"model.layers.1.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
|
26 |
+
"model.layers.1.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
|
27 |
+
"model.layers.1.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
|
28 |
+
"model.layers.1.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
|
29 |
+
"model.layers.1.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
|
30 |
+
"model.layers.1.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
|
31 |
+
"model.layers.1.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
|
32 |
+
"model.layers.10.input_layernorm.weight": "model-00002-of-00004.safetensors",
|
33 |
+
"model.layers.10.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
|
34 |
+
"model.layers.10.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
|
35 |
+
"model.layers.10.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
|
36 |
+
"model.layers.10.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
|
37 |
+
"model.layers.10.self_attn.k_proj.bias": "model-00002-of-00004.safetensors",
|
38 |
+
"model.layers.10.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
|
39 |
+
"model.layers.10.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
|
40 |
+
"model.layers.10.self_attn.q_proj.bias": "model-00002-of-00004.safetensors",
|
41 |
+
"model.layers.10.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
|
42 |
+
"model.layers.10.self_attn.v_proj.bias": "model-00002-of-00004.safetensors",
|
43 |
+
"model.layers.10.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
|
44 |
+
"model.layers.11.input_layernorm.weight": "model-00002-of-00004.safetensors",
|
45 |
+
"model.layers.11.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
|
46 |
+
"model.layers.11.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
|
47 |
+
"model.layers.11.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
|
48 |
+
"model.layers.11.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
|
49 |
+
"model.layers.11.self_attn.k_proj.bias": "model-00002-of-00004.safetensors",
|
50 |
+
"model.layers.11.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
|
51 |
+
"model.layers.11.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
|
52 |
+
"model.layers.11.self_attn.q_proj.bias": "model-00002-of-00004.safetensors",
|
53 |
+
"model.layers.11.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
|
54 |
+
"model.layers.11.self_attn.v_proj.bias": "model-00002-of-00004.safetensors",
|
55 |
+
"model.layers.11.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
|
56 |
+
"model.layers.12.input_layernorm.weight": "model-00002-of-00004.safetensors",
|
57 |
+
"model.layers.12.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
|
58 |
+
"model.layers.12.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
|
59 |
+
"model.layers.12.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
|
60 |
+
"model.layers.12.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
|
61 |
+
"model.layers.12.self_attn.k_proj.bias": "model-00002-of-00004.safetensors",
|
62 |
+
"model.layers.12.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
|
63 |
+
"model.layers.12.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
|
64 |
+
"model.layers.12.self_attn.q_proj.bias": "model-00002-of-00004.safetensors",
|
65 |
+
"model.layers.12.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
|
66 |
+
"model.layers.12.self_attn.v_proj.bias": "model-00002-of-00004.safetensors",
|
67 |
+
"model.layers.12.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
|
68 |
+
"model.layers.13.input_layernorm.weight": "model-00002-of-00004.safetensors",
|
69 |
+
"model.layers.13.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
|
70 |
+
"model.layers.13.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
|
71 |
+
"model.layers.13.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
|
72 |
+
"model.layers.13.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
|
73 |
+
"model.layers.13.self_attn.k_proj.bias": "model-00002-of-00004.safetensors",
|
74 |
+
"model.layers.13.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
|
75 |
+
"model.layers.13.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
|
76 |
+
"model.layers.13.self_attn.q_proj.bias": "model-00002-of-00004.safetensors",
|
77 |
+
"model.layers.13.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
|
78 |
+
"model.layers.13.self_attn.v_proj.bias": "model-00002-of-00004.safetensors",
|
79 |
+
"model.layers.13.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
|
80 |
+
"model.layers.14.input_layernorm.weight": "model-00002-of-00004.safetensors",
|
81 |
+
"model.layers.14.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
|
82 |
+
"model.layers.14.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
|
83 |
+
"model.layers.14.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
|
84 |
+
"model.layers.14.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
|
85 |
+
"model.layers.14.self_attn.k_proj.bias": "model-00002-of-00004.safetensors",
|
86 |
+
"model.layers.14.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
|
87 |
+
"model.layers.14.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
|
88 |
+
"model.layers.14.self_attn.q_proj.bias": "model-00002-of-00004.safetensors",
|
89 |
+
"model.layers.14.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
|
90 |
+
"model.layers.14.self_attn.v_proj.bias": "model-00002-of-00004.safetensors",
|
91 |
+
"model.layers.14.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
|
92 |
+
"model.layers.15.input_layernorm.weight": "model-00002-of-00004.safetensors",
|
93 |
+
"model.layers.15.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
|
94 |
+
"model.layers.15.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
|
95 |
+
"model.layers.15.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
|
96 |
+
"model.layers.15.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
|
97 |
+
"model.layers.15.self_attn.k_proj.bias": "model-00002-of-00004.safetensors",
|
98 |
+
"model.layers.15.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
|
99 |
+
"model.layers.15.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
|
100 |
+
"model.layers.15.self_attn.q_proj.bias": "model-00002-of-00004.safetensors",
|
101 |
+
"model.layers.15.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
|
102 |
+
"model.layers.15.self_attn.v_proj.bias": "model-00002-of-00004.safetensors",
|
103 |
+
"model.layers.15.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
|
104 |
+
"model.layers.16.input_layernorm.weight": "model-00002-of-00004.safetensors",
|
105 |
+
"model.layers.16.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
|
106 |
+
"model.layers.16.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
|
107 |
+
"model.layers.16.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
|
108 |
+
"model.layers.16.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
|
109 |
+
"model.layers.16.self_attn.k_proj.bias": "model-00002-of-00004.safetensors",
|
110 |
+
"model.layers.16.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
|
111 |
+
"model.layers.16.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
|
112 |
+
"model.layers.16.self_attn.q_proj.bias": "model-00002-of-00004.safetensors",
|
113 |
+
"model.layers.16.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
|
114 |
+
"model.layers.16.self_attn.v_proj.bias": "model-00002-of-00004.safetensors",
|
115 |
+
"model.layers.16.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
|
116 |
+
"model.layers.17.input_layernorm.weight": "model-00002-of-00004.safetensors",
|
117 |
+
"model.layers.17.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
|
118 |
+
"model.layers.17.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
|
119 |
+
"model.layers.17.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
|
120 |
+
"model.layers.17.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
|
121 |
+
"model.layers.17.self_attn.k_proj.bias": "model-00002-of-00004.safetensors",
|
122 |
+
"model.layers.17.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
|
123 |
+
"model.layers.17.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
|
124 |
+
"model.layers.17.self_attn.q_proj.bias": "model-00002-of-00004.safetensors",
|
125 |
+
"model.layers.17.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
|
126 |
+
"model.layers.17.self_attn.v_proj.bias": "model-00002-of-00004.safetensors",
|
127 |
+
"model.layers.17.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
|
128 |
+
"model.layers.18.input_layernorm.weight": "model-00003-of-00004.safetensors",
|
129 |
+
"model.layers.18.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
|
130 |
+
"model.layers.18.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
|
131 |
+
"model.layers.18.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
|
132 |
+
"model.layers.18.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
|
133 |
+
"model.layers.18.self_attn.k_proj.bias": "model-00002-of-00004.safetensors",
|
134 |
+
"model.layers.18.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
|
135 |
+
"model.layers.18.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
|
136 |
+
"model.layers.18.self_attn.q_proj.bias": "model-00002-of-00004.safetensors",
|
137 |
+
"model.layers.18.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
|
138 |
+
"model.layers.18.self_attn.v_proj.bias": "model-00002-of-00004.safetensors",
|
139 |
+
"model.layers.18.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
|
140 |
+
"model.layers.19.input_layernorm.weight": "model-00003-of-00004.safetensors",
|
141 |
+
"model.layers.19.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
|
142 |
+
"model.layers.19.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
|
143 |
+
"model.layers.19.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
|
144 |
+
"model.layers.19.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
|
145 |
+
"model.layers.19.self_attn.k_proj.bias": "model-00003-of-00004.safetensors",
|
146 |
+
"model.layers.19.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
|
147 |
+
"model.layers.19.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
|
148 |
+
"model.layers.19.self_attn.q_proj.bias": "model-00003-of-00004.safetensors",
|
149 |
+
"model.layers.19.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
|
150 |
+
"model.layers.19.self_attn.v_proj.bias": "model-00003-of-00004.safetensors",
|
151 |
+
"model.layers.19.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
|
152 |
+
"model.layers.2.input_layernorm.weight": "model-00001-of-00004.safetensors",
|
153 |
+
"model.layers.2.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
|
154 |
+
"model.layers.2.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
|
155 |
+
"model.layers.2.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
|
156 |
+
"model.layers.2.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
|
157 |
+
"model.layers.2.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
|
158 |
+
"model.layers.2.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
|
159 |
+
"model.layers.2.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
|
160 |
+
"model.layers.2.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
|
161 |
+
"model.layers.2.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
|
162 |
+
"model.layers.2.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
|
163 |
+
"model.layers.2.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
|
164 |
+
"model.layers.20.input_layernorm.weight": "model-00003-of-00004.safetensors",
|
165 |
+
"model.layers.20.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
|
166 |
+
"model.layers.20.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
|
167 |
+
"model.layers.20.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
|
168 |
+
"model.layers.20.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
|
169 |
+
"model.layers.20.self_attn.k_proj.bias": "model-00003-of-00004.safetensors",
|
170 |
+
"model.layers.20.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
|
171 |
+
"model.layers.20.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
|
172 |
+
"model.layers.20.self_attn.q_proj.bias": "model-00003-of-00004.safetensors",
|
173 |
+
"model.layers.20.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
|
174 |
+
"model.layers.20.self_attn.v_proj.bias": "model-00003-of-00004.safetensors",
|
175 |
+
"model.layers.20.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
|
176 |
+
"model.layers.21.input_layernorm.weight": "model-00003-of-00004.safetensors",
|
177 |
+
"model.layers.21.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
|
178 |
+
"model.layers.21.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
|
179 |
+
"model.layers.21.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
|
180 |
+
"model.layers.21.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
|
181 |
+
"model.layers.21.self_attn.k_proj.bias": "model-00003-of-00004.safetensors",
|
182 |
+
"model.layers.21.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
|
183 |
+
"model.layers.21.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
|
184 |
+
"model.layers.21.self_attn.q_proj.bias": "model-00003-of-00004.safetensors",
|
185 |
+
"model.layers.21.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
|
186 |
+
"model.layers.21.self_attn.v_proj.bias": "model-00003-of-00004.safetensors",
|
187 |
+
"model.layers.21.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
|
188 |
+
"model.layers.22.input_layernorm.weight": "model-00003-of-00004.safetensors",
|
189 |
+
"model.layers.22.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
|
190 |
+
"model.layers.22.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
|
191 |
+
"model.layers.22.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
|
192 |
+
"model.layers.22.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
|
193 |
+
"model.layers.22.self_attn.k_proj.bias": "model-00003-of-00004.safetensors",
|
194 |
+
"model.layers.22.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
|
195 |
+
"model.layers.22.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
|
196 |
+
"model.layers.22.self_attn.q_proj.bias": "model-00003-of-00004.safetensors",
|
197 |
+
"model.layers.22.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
|
198 |
+
"model.layers.22.self_attn.v_proj.bias": "model-00003-of-00004.safetensors",
|
199 |
+
"model.layers.22.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
|
200 |
+
"model.layers.23.input_layernorm.weight": "model-00003-of-00004.safetensors",
|
201 |
+
"model.layers.23.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
|
202 |
+
"model.layers.23.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
|
203 |
+
"model.layers.23.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
|
204 |
+
"model.layers.23.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
|
205 |
+
"model.layers.23.self_attn.k_proj.bias": "model-00003-of-00004.safetensors",
|
206 |
+
"model.layers.23.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
|
207 |
+
"model.layers.23.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
|
208 |
+
"model.layers.23.self_attn.q_proj.bias": "model-00003-of-00004.safetensors",
|
209 |
+
"model.layers.23.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
|
210 |
+
"model.layers.23.self_attn.v_proj.bias": "model-00003-of-00004.safetensors",
|
211 |
+
"model.layers.23.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
|
212 |
+
"model.layers.24.input_layernorm.weight": "model-00003-of-00004.safetensors",
|
213 |
+
"model.layers.24.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
|
214 |
+
"model.layers.24.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
|
215 |
+
"model.layers.24.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
|
216 |
+
"model.layers.24.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
|
217 |
+
"model.layers.24.self_attn.k_proj.bias": "model-00003-of-00004.safetensors",
|
218 |
+
"model.layers.24.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
|
219 |
+
"model.layers.24.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
|
220 |
+
"model.layers.24.self_attn.q_proj.bias": "model-00003-of-00004.safetensors",
|
221 |
+
"model.layers.24.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
|
222 |
+
"model.layers.24.self_attn.v_proj.bias": "model-00003-of-00004.safetensors",
|
223 |
+
"model.layers.24.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
|
224 |
+
"model.layers.25.input_layernorm.weight": "model-00003-of-00004.safetensors",
|
225 |
+
"model.layers.25.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
|
226 |
+
"model.layers.25.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
|
227 |
+
"model.layers.25.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
|
228 |
+
"model.layers.25.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
|
229 |
+
"model.layers.25.self_attn.k_proj.bias": "model-00003-of-00004.safetensors",
|
230 |
+
"model.layers.25.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
|
231 |
+
"model.layers.25.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
|
232 |
+
"model.layers.25.self_attn.q_proj.bias": "model-00003-of-00004.safetensors",
|
233 |
+
"model.layers.25.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
|
234 |
+
"model.layers.25.self_attn.v_proj.bias": "model-00003-of-00004.safetensors",
|
235 |
+
"model.layers.25.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
|
236 |
+
"model.layers.26.input_layernorm.weight": "model-00003-of-00004.safetensors",
|
237 |
+
"model.layers.26.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
|
238 |
+
"model.layers.26.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
|
239 |
+
"model.layers.26.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
|
240 |
+
"model.layers.26.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
|
241 |
+
"model.layers.26.self_attn.k_proj.bias": "model-00003-of-00004.safetensors",
|
242 |
+
"model.layers.26.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
|
243 |
+
"model.layers.26.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
|
244 |
+
"model.layers.26.self_attn.q_proj.bias": "model-00003-of-00004.safetensors",
|
245 |
+
"model.layers.26.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
|
246 |
+
"model.layers.26.self_attn.v_proj.bias": "model-00003-of-00004.safetensors",
|
247 |
+
"model.layers.26.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
|
248 |
+
"model.layers.27.input_layernorm.weight": "model-00003-of-00004.safetensors",
|
249 |
+
"model.layers.27.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
|
250 |
+
"model.layers.27.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
|
251 |
+
"model.layers.27.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
|
252 |
+
"model.layers.27.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
|
253 |
+
"model.layers.27.self_attn.k_proj.bias": "model-00003-of-00004.safetensors",
|
254 |
+
"model.layers.27.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
|
255 |
+
"model.layers.27.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
|
256 |
+
"model.layers.27.self_attn.q_proj.bias": "model-00003-of-00004.safetensors",
|
257 |
+
"model.layers.27.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
|
258 |
+
"model.layers.27.self_attn.v_proj.bias": "model-00003-of-00004.safetensors",
|
259 |
+
"model.layers.27.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
|
260 |
+
"model.layers.3.input_layernorm.weight": "model-00001-of-00004.safetensors",
|
261 |
+
"model.layers.3.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
|
262 |
+
"model.layers.3.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
|
263 |
+
"model.layers.3.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
|
264 |
+
"model.layers.3.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
|
265 |
+
"model.layers.3.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
|
266 |
+
"model.layers.3.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
|
267 |
+
"model.layers.3.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
|
268 |
+
"model.layers.3.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
|
269 |
+
"model.layers.3.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
|
270 |
+
"model.layers.3.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
|
271 |
+
"model.layers.3.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
|
272 |
+
"model.layers.4.input_layernorm.weight": "model-00001-of-00004.safetensors",
|
273 |
+
"model.layers.4.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
|
274 |
+
"model.layers.4.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
|
275 |
+
"model.layers.4.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
|
276 |
+
"model.layers.4.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
|
277 |
+
"model.layers.4.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
|
278 |
+
"model.layers.4.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
|
279 |
+
"model.layers.4.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
|
280 |
+
"model.layers.4.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
|
281 |
+
"model.layers.4.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
|
282 |
+
"model.layers.4.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
|
283 |
+
"model.layers.4.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
|
284 |
+
"model.layers.5.input_layernorm.weight": "model-00001-of-00004.safetensors",
|
285 |
+
"model.layers.5.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
|
286 |
+
"model.layers.5.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
|
287 |
+
"model.layers.5.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
|
288 |
+
"model.layers.5.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
|
289 |
+
"model.layers.5.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
|
290 |
+
"model.layers.5.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
|
291 |
+
"model.layers.5.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
|
292 |
+
"model.layers.5.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
|
293 |
+
"model.layers.5.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
|
294 |
+
"model.layers.5.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
|
295 |
+
"model.layers.5.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
|
296 |
+
"model.layers.6.input_layernorm.weight": "model-00001-of-00004.safetensors",
|
297 |
+
"model.layers.6.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
|
298 |
+
"model.layers.6.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
|
299 |
+
"model.layers.6.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
|
300 |
+
"model.layers.6.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
|
301 |
+
"model.layers.6.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
|
302 |
+
"model.layers.6.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
|
303 |
+
"model.layers.6.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
|
304 |
+
"model.layers.6.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
|
305 |
+
"model.layers.6.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
|
306 |
+
"model.layers.6.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
|
307 |
+
"model.layers.6.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
|
308 |
+
"model.layers.7.input_layernorm.weight": "model-00001-of-00004.safetensors",
|
309 |
+
"model.layers.7.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
|
310 |
+
"model.layers.7.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
|
311 |
+
"model.layers.7.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
|
312 |
+
"model.layers.7.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
|
313 |
+
"model.layers.7.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
|
314 |
+
"model.layers.7.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
|
315 |
+
"model.layers.7.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
|
316 |
+
"model.layers.7.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
|
317 |
+
"model.layers.7.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
|
318 |
+
"model.layers.7.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
|
319 |
+
"model.layers.7.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
|
320 |
+
"model.layers.8.input_layernorm.weight": "model-00002-of-00004.safetensors",
|
321 |
+
"model.layers.8.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
|
322 |
+
"model.layers.8.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
|
323 |
+
"model.layers.8.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
|
324 |
+
"model.layers.8.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
|
325 |
+
"model.layers.8.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
|
326 |
+
"model.layers.8.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
|
327 |
+
"model.layers.8.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
|
328 |
+
"model.layers.8.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
|
329 |
+
"model.layers.8.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
|
330 |
+
"model.layers.8.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
|
331 |
+
"model.layers.8.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
|
332 |
+
"model.layers.9.input_layernorm.weight": "model-00002-of-00004.safetensors",
|
333 |
+
"model.layers.9.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
|
334 |
+
"model.layers.9.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
|
335 |
+
"model.layers.9.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
|
336 |
+
"model.layers.9.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
|
337 |
+
"model.layers.9.self_attn.k_proj.bias": "model-00002-of-00004.safetensors",
|
338 |
+
"model.layers.9.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
|
339 |
+
"model.layers.9.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
|
340 |
+
"model.layers.9.self_attn.q_proj.bias": "model-00002-of-00004.safetensors",
|
341 |
+
"model.layers.9.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
|
342 |
+
"model.layers.9.self_attn.v_proj.bias": "model-00002-of-00004.safetensors",
|
343 |
+
"model.layers.9.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
|
344 |
+
"model.norm.weight": "model-00003-of-00004.safetensors"
|
345 |
+
}
|
346 |
+
}
|
special_tokens_map.json
ADDED
@@ -0,0 +1,60 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"additional_special_tokens": [
|
3 |
+
{
|
4 |
+
"content": "[STEP]",
|
5 |
+
"lstrip": false,
|
6 |
+
"normalized": false,
|
7 |
+
"rstrip": false,
|
8 |
+
"single_word": false
|
9 |
+
},
|
10 |
+
{
|
11 |
+
"content": "[/STEP]",
|
12 |
+
"lstrip": false,
|
13 |
+
"normalized": false,
|
14 |
+
"rstrip": false,
|
15 |
+
"single_word": false
|
16 |
+
},
|
17 |
+
{
|
18 |
+
"content": "[TURN]",
|
19 |
+
"lstrip": false,
|
20 |
+
"normalized": false,
|
21 |
+
"rstrip": false,
|
22 |
+
"single_word": false
|
23 |
+
},
|
24 |
+
{
|
25 |
+
"content": "[/TURN]",
|
26 |
+
"lstrip": false,
|
27 |
+
"normalized": false,
|
28 |
+
"rstrip": false,
|
29 |
+
"single_word": false
|
30 |
+
},
|
31 |
+
{
|
32 |
+
"content": "[REVISION]",
|
33 |
+
"lstrip": false,
|
34 |
+
"normalized": false,
|
35 |
+
"rstrip": false,
|
36 |
+
"single_word": false
|
37 |
+
},
|
38 |
+
{
|
39 |
+
"content": "[/REVISION]",
|
40 |
+
"lstrip": false,
|
41 |
+
"normalized": false,
|
42 |
+
"rstrip": false,
|
43 |
+
"single_word": false
|
44 |
+
}
|
45 |
+
],
|
46 |
+
"eos_token": {
|
47 |
+
"content": "<|im_end|>",
|
48 |
+
"lstrip": false,
|
49 |
+
"normalized": false,
|
50 |
+
"rstrip": false,
|
51 |
+
"single_word": false
|
52 |
+
},
|
53 |
+
"pad_token": {
|
54 |
+
"content": "<|endoftext|>",
|
55 |
+
"lstrip": false,
|
56 |
+
"normalized": false,
|
57 |
+
"rstrip": false,
|
58 |
+
"single_word": false
|
59 |
+
}
|
60 |
+
}
|
tokenizer.json
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:119474abe8593f47402f00c90e22eac0436e64f8c84c8270309a27e152901418
|
3 |
+
size 11423104
|
tokenizer_config.json
ADDED
@@ -0,0 +1,248 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"add_bos_token": false,
|
3 |
+
"add_prefix_space": false,
|
4 |
+
"added_tokens_decoder": {
|
5 |
+
"151643": {
|
6 |
+
"content": "<|endoftext|>",
|
7 |
+
"lstrip": false,
|
8 |
+
"normalized": false,
|
9 |
+
"rstrip": false,
|
10 |
+
"single_word": false,
|
11 |
+
"special": true
|
12 |
+
},
|
13 |
+
"151644": {
|
14 |
+
"content": "<|im_start|>",
|
15 |
+
"lstrip": false,
|
16 |
+
"normalized": false,
|
17 |
+
"rstrip": false,
|
18 |
+
"single_word": false,
|
19 |
+
"special": true
|
20 |
+
},
|
21 |
+
"151645": {
|
22 |
+
"content": "<|im_end|>",
|
23 |
+
"lstrip": false,
|
24 |
+
"normalized": false,
|
25 |
+
"rstrip": false,
|
26 |
+
"single_word": false,
|
27 |
+
"special": true
|
28 |
+
},
|
29 |
+
"151646": {
|
30 |
+
"content": "<|object_ref_start|>",
|
31 |
+
"lstrip": false,
|
32 |
+
"normalized": false,
|
33 |
+
"rstrip": false,
|
34 |
+
"single_word": false,
|
35 |
+
"special": true
|
36 |
+
},
|
37 |
+
"151647": {
|
38 |
+
"content": "<|object_ref_end|>",
|
39 |
+
"lstrip": false,
|
40 |
+
"normalized": false,
|
41 |
+
"rstrip": false,
|
42 |
+
"single_word": false,
|
43 |
+
"special": true
|
44 |
+
},
|
45 |
+
"151648": {
|
46 |
+
"content": "<|box_start|>",
|
47 |
+
"lstrip": false,
|
48 |
+
"normalized": false,
|
49 |
+
"rstrip": false,
|
50 |
+
"single_word": false,
|
51 |
+
"special": true
|
52 |
+
},
|
53 |
+
"151649": {
|
54 |
+
"content": "<|box_end|>",
|
55 |
+
"lstrip": false,
|
56 |
+
"normalized": false,
|
57 |
+
"rstrip": false,
|
58 |
+
"single_word": false,
|
59 |
+
"special": true
|
60 |
+
},
|
61 |
+
"151650": {
|
62 |
+
"content": "<|quad_start|>",
|
63 |
+
"lstrip": false,
|
64 |
+
"normalized": false,
|
65 |
+
"rstrip": false,
|
66 |
+
"single_word": false,
|
67 |
+
"special": true
|
68 |
+
},
|
69 |
+
"151651": {
|
70 |
+
"content": "<|quad_end|>",
|
71 |
+
"lstrip": false,
|
72 |
+
"normalized": false,
|
73 |
+
"rstrip": false,
|
74 |
+
"single_word": false,
|
75 |
+
"special": true
|
76 |
+
},
|
77 |
+
"151652": {
|
78 |
+
"content": "<|vision_start|>",
|
79 |
+
"lstrip": false,
|
80 |
+
"normalized": false,
|
81 |
+
"rstrip": false,
|
82 |
+
"single_word": false,
|
83 |
+
"special": true
|
84 |
+
},
|
85 |
+
"151653": {
|
86 |
+
"content": "<|vision_end|>",
|
87 |
+
"lstrip": false,
|
88 |
+
"normalized": false,
|
89 |
+
"rstrip": false,
|
90 |
+
"single_word": false,
|
91 |
+
"special": true
|
92 |
+
},
|
93 |
+
"151654": {
|
94 |
+
"content": "<|vision_pad|>",
|
95 |
+
"lstrip": false,
|
96 |
+
"normalized": false,
|
97 |
+
"rstrip": false,
|
98 |
+
"single_word": false,
|
99 |
+
"special": true
|
100 |
+
},
|
101 |
+
"151655": {
|
102 |
+
"content": "<|image_pad|>",
|
103 |
+
"lstrip": false,
|
104 |
+
"normalized": false,
|
105 |
+
"rstrip": false,
|
106 |
+
"single_word": false,
|
107 |
+
"special": true
|
108 |
+
},
|
109 |
+
"151656": {
|
110 |
+
"content": "<|video_pad|>",
|
111 |
+
"lstrip": false,
|
112 |
+
"normalized": false,
|
113 |
+
"rstrip": false,
|
114 |
+
"single_word": false,
|
115 |
+
"special": true
|
116 |
+
},
|
117 |
+
"151657": {
|
118 |
+
"content": "<tool_call>",
|
119 |
+
"lstrip": false,
|
120 |
+
"normalized": false,
|
121 |
+
"rstrip": false,
|
122 |
+
"single_word": false,
|
123 |
+
"special": false
|
124 |
+
},
|
125 |
+
"151658": {
|
126 |
+
"content": "</tool_call>",
|
127 |
+
"lstrip": false,
|
128 |
+
"normalized": false,
|
129 |
+
"rstrip": false,
|
130 |
+
"single_word": false,
|
131 |
+
"special": false
|
132 |
+
},
|
133 |
+
"151659": {
|
134 |
+
"content": "<|fim_prefix|>",
|
135 |
+
"lstrip": false,
|
136 |
+
"normalized": false,
|
137 |
+
"rstrip": false,
|
138 |
+
"single_word": false,
|
139 |
+
"special": false
|
140 |
+
},
|
141 |
+
"151660": {
|
142 |
+
"content": "<|fim_middle|>",
|
143 |
+
"lstrip": false,
|
144 |
+
"normalized": false,
|
145 |
+
"rstrip": false,
|
146 |
+
"single_word": false,
|
147 |
+
"special": false
|
148 |
+
},
|
149 |
+
"151661": {
|
150 |
+
"content": "<|fim_suffix|>",
|
151 |
+
"lstrip": false,
|
152 |
+
"normalized": false,
|
153 |
+
"rstrip": false,
|
154 |
+
"single_word": false,
|
155 |
+
"special": false
|
156 |
+
},
|
157 |
+
"151662": {
|
158 |
+
"content": "<|fim_pad|>",
|
159 |
+
"lstrip": false,
|
160 |
+
"normalized": false,
|
161 |
+
"rstrip": false,
|
162 |
+
"single_word": false,
|
163 |
+
"special": false
|
164 |
+
},
|
165 |
+
"151663": {
|
166 |
+
"content": "<|repo_name|>",
|
167 |
+
"lstrip": false,
|
168 |
+
"normalized": false,
|
169 |
+
"rstrip": false,
|
170 |
+
"single_word": false,
|
171 |
+
"special": false
|
172 |
+
},
|
173 |
+
"151664": {
|
174 |
+
"content": "<|file_sep|>",
|
175 |
+
"lstrip": false,
|
176 |
+
"normalized": false,
|
177 |
+
"rstrip": false,
|
178 |
+
"single_word": false,
|
179 |
+
"special": false
|
180 |
+
},
|
181 |
+
"151665": {
|
182 |
+
"content": "[STEP]",
|
183 |
+
"lstrip": false,
|
184 |
+
"normalized": false,
|
185 |
+
"rstrip": false,
|
186 |
+
"single_word": false,
|
187 |
+
"special": true
|
188 |
+
},
|
189 |
+
"151666": {
|
190 |
+
"content": "[/STEP]",
|
191 |
+
"lstrip": false,
|
192 |
+
"normalized": false,
|
193 |
+
"rstrip": false,
|
194 |
+
"single_word": false,
|
195 |
+
"special": true
|
196 |
+
},
|
197 |
+
"151667": {
|
198 |
+
"content": "[TURN]",
|
199 |
+
"lstrip": false,
|
200 |
+
"normalized": false,
|
201 |
+
"rstrip": false,
|
202 |
+
"single_word": false,
|
203 |
+
"special": true
|
204 |
+
},
|
205 |
+
"151668": {
|
206 |
+
"content": "[/TURN]",
|
207 |
+
"lstrip": false,
|
208 |
+
"normalized": false,
|
209 |
+
"rstrip": false,
|
210 |
+
"single_word": false,
|
211 |
+
"special": true
|
212 |
+
},
|
213 |
+
"151669": {
|
214 |
+
"content": "[REVISION]",
|
215 |
+
"lstrip": false,
|
216 |
+
"normalized": false,
|
217 |
+
"rstrip": false,
|
218 |
+
"single_word": false,
|
219 |
+
"special": true
|
220 |
+
},
|
221 |
+
"151670": {
|
222 |
+
"content": "[/REVISION]",
|
223 |
+
"lstrip": false,
|
224 |
+
"normalized": false,
|
225 |
+
"rstrip": false,
|
226 |
+
"single_word": false,
|
227 |
+
"special": true
|
228 |
+
}
|
229 |
+
},
|
230 |
+
"additional_special_tokens": [
|
231 |
+
"[STEP]",
|
232 |
+
"[/STEP]",
|
233 |
+
"[TURN]",
|
234 |
+
"[/TURN]",
|
235 |
+
"[REVISION]",
|
236 |
+
"[/REVISION]"
|
237 |
+
],
|
238 |
+
"bos_token": null,
|
239 |
+
"chat_template": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0]['role'] == 'system' %}\n {{- messages[0]['content'] }}\n {%- else %}\n {{- 'You are Qwen, created by Alibaba Cloud. You are a helpful assistant.' }}\n {%- endif %}\n {{- \"\\n\\n# Tools\\n\\nYou may call one or more functions to assist with the user query.\\n\\nYou are provided with function signatures within <tools></tools> XML tags:\\n<tools>\" }}\n {%- for tool in tools %}\n {{- \"\\n\" }}\n {{- tool | tojson }}\n {%- endfor %}\n {{- \"\\n</tools>\\n\\nFor each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:\\n<tool_call>\\n{\\\"name\\\": <function-name>, \\\"arguments\\\": <args-json-object>}\\n</tool_call><|im_end|>\\n\" }}\n{%- else %}\n {%- if messages[0]['role'] == 'system' %}\n {{- '<|im_start|>system\\n' + messages[0]['content'] + '<|im_end|>\\n' }}\n {%- else %}\n {{- '<|im_start|>system\\nYou are Qwen, created by Alibaba Cloud. You are a helpful assistant.<|im_end|>\\n' }}\n {%- endif %}\n{%- endif %}\n{%- for message in messages %}\n {%- if (message.role == \"user\") or (message.role == \"system\" and not loop.first) or (message.role == \"assistant\" and not message.tool_calls) %}\n {{- '<|im_start|>' + message.role + '\\n' + message.content + '<|im_end|>' + '\\n' }}\n {%- elif message.role == \"assistant\" %}\n {{- '<|im_start|>' + message.role }}\n {%- if message.content %}\n {{- '\\n' + message.content }}\n {%- endif %}\n {%- for tool_call in message.tool_calls %}\n {%- if tool_call.function is defined %}\n {%- set tool_call = tool_call.function %}\n {%- endif %}\n {{- '\\n<tool_call>\\n{\"name\": \"' }}\n {{- tool_call.name }}\n {{- '\", \"arguments\": ' }}\n {{- tool_call.arguments | tojson }}\n {{- '}\\n</tool_call>' }}\n {%- endfor %}\n {{- '<|im_end|>\\n' }}\n {%- elif message.role == \"tool\" %}\n {%- if (loop.index0 == 0) or (messages[loop.index0 - 1].role != \"tool\") %}\n {{- '<|im_start|>user' }}\n {%- endif %}\n {{- '\\n<tool_response>\\n' }}\n {{- message.content }}\n {{- '\\n</tool_response>' }}\n {%- if loop.last or (messages[loop.index0 + 1].role != \"tool\") %}\n {{- '<|im_end|>\\n' }}\n {%- endif %}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|im_start|>assistant\\n' }}\n{%- endif %}\n",
|
240 |
+
"clean_up_tokenization_spaces": false,
|
241 |
+
"eos_token": "<|im_end|>",
|
242 |
+
"errors": "replace",
|
243 |
+
"model_max_length": 131072,
|
244 |
+
"pad_token": "<|endoftext|>",
|
245 |
+
"split_special_tokens": false,
|
246 |
+
"tokenizer_class": "Qwen2Tokenizer",
|
247 |
+
"unk_token": null
|
248 |
+
}
|
training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5d8b58b2fd3affe635bf267c95f88ed337f48f927e28276811b6aa45b9a3dd53
|
3 |
+
size 6328
|
vocab.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
wandb/debug-internal.log
ADDED
@@ -0,0 +1,373 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
2024-11-13 19:02:46,431 INFO StreamThr :1939 [internal.py:wandb_internal():86] W&B internal server running at pid: 1939, started at: 2024-11-13 19:02:46.431351
|
2 |
+
2024-11-13 19:02:46,433 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status
|
3 |
+
2024-11-13 19:02:46,434 INFO WriterThread:1939 [datastore.py:open_for_write():87] open: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/run-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1.wandb
|
4 |
+
2024-11-13 19:02:46,435 DEBUG SenderThread:1939 [sender.py:send():382] send: header
|
5 |
+
2024-11-13 19:02:46,442 DEBUG SenderThread:1939 [sender.py:send():382] send: run
|
6 |
+
2024-11-13 19:02:46,772 INFO SenderThread:1939 [dir_watcher.py:__init__():211] watching files in: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files
|
7 |
+
2024-11-13 19:02:46,772 INFO SenderThread:1939 [sender.py:_start_run_threads():1136] run started: kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1 with start time 1731524566.431221
|
8 |
+
2024-11-13 19:02:46,782 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: check_version
|
9 |
+
2024-11-13 19:02:46,782 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: check_version
|
10 |
+
2024-11-13 19:02:46,859 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: run_start
|
11 |
+
2024-11-13 19:02:46,947 DEBUG HandlerThread:1939 [system_info.py:__init__():27] System info init
|
12 |
+
2024-11-13 19:02:46,947 DEBUG HandlerThread:1939 [system_info.py:__init__():42] System info init done
|
13 |
+
2024-11-13 19:02:46,947 INFO HandlerThread:1939 [system_monitor.py:start():194] Starting system monitor
|
14 |
+
2024-11-13 19:02:46,947 INFO SystemMonitor:1939 [system_monitor.py:_start():158] Starting system asset monitoring threads
|
15 |
+
2024-11-13 19:02:46,948 INFO HandlerThread:1939 [system_monitor.py:probe():214] Collecting system info
|
16 |
+
2024-11-13 19:02:46,948 INFO SystemMonitor:1939 [interfaces.py:start():190] Started cpu monitoring
|
17 |
+
2024-11-13 19:02:46,949 INFO SystemMonitor:1939 [interfaces.py:start():190] Started disk monitoring
|
18 |
+
2024-11-13 19:02:46,949 INFO SystemMonitor:1939 [interfaces.py:start():190] Started gpu monitoring
|
19 |
+
2024-11-13 19:02:46,951 INFO SystemMonitor:1939 [interfaces.py:start():190] Started memory monitoring
|
20 |
+
2024-11-13 19:02:46,952 INFO SystemMonitor:1939 [interfaces.py:start():190] Started network monitoring
|
21 |
+
2024-11-13 19:02:47,010 DEBUG HandlerThread:1939 [system_info.py:probe():151] Probing system
|
22 |
+
2024-11-13 19:02:47,013 DEBUG HandlerThread:1939 [system_info.py:_probe_git():136] Probing git
|
23 |
+
2024-11-13 19:02:47,021 DEBUG HandlerThread:1939 [system_info.py:_probe_git():144] Probing git done
|
24 |
+
2024-11-13 19:02:47,021 DEBUG HandlerThread:1939 [system_info.py:probe():199] Probing system done
|
25 |
+
2024-11-13 19:02:47,022 DEBUG HandlerThread:1939 [system_monitor.py:probe():223] {'os': 'Linux-5.10.226-214.879.amzn2.x86_64-x86_64-with-glibc2.31', 'python': '3.10.12', 'heartbeatAt': '2024-11-13T19:02:47.010671', 'startedAt': '2024-11-13T19:02:46.426230', 'docker': None, 'cuda': None, 'args': ('--batch_size', '64', '--cache_dir', '/opt/ml/data/input/.cache', '--dataset', 'code-contests', '--gradient_accumulation_steps', '8', '--logging_steps', '5', '--lr', '1e-6', '--max_seq_length', '2048', '--model_name', 'Qwen/Qwen2.5-Coder-7B-Instruct', '--num_train_epochs', '2', '--output_dir', '/opt/ml/model/', '--push_to_hub', 'True', '--run_name', 'sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048', '--save_steps', '100', '--schedule', 'cosine', '--wandb_project', 'sft-codecontests-1112', '--weight_decay', '0.0'), 'state': 'running', 'program': '/opt/ml/code/qlearning_reasoning/training/sft.py', 'codePathLocal': 'qlearning_reasoning/training/sft.py', 'codePath': 'qlearning_reasoning/training/sft.py', 'git': {'remote': 'git@github.com:TRI-ML/reasoning-value-verifiers.git', 'commit': 'cbfd05147f15c55be83f4996eebdb305c17f119d'}, 'email': None, 'root': '/opt/ml/code', 'host': 'algo-1', 'username': 'root', 'executable': '/opt/conda/bin/python3.10', 'cpu_count': 96, 'cpu_count_logical': 192, 'cpu_freq': {'current': 2829.910838541667, 'min': 0.0, 'max': 0.0}, 'cpu_freq_per_core': [{'current': 2850.788, 'min': 0.0, 'max': 0.0}, {'current': 3406.159, 'min': 0.0, 'max': 0.0}, {'current': 2605.944, 'min': 0.0, 'max': 0.0}, {'current': 2608.162, 'min': 0.0, 'max': 0.0}, {'current': 2818.169, 'min': 0.0, 'max': 0.0}, {'current': 2591.551, 'min': 0.0, 'max': 0.0}, {'current': 2621.486, 'min': 0.0, 'max': 0.0}, {'current': 2607.504, 'min': 0.0, 'max': 0.0}, {'current': 2999.734, 'min': 0.0, 'max': 0.0}, {'current': 2991.449, 'min': 0.0, 'max': 0.0}, {'current': 3123.177, 'min': 0.0, 'max': 0.0}, {'current': 3053.641, 'min': 0.0, 'max': 0.0}, {'current': 3003.014, 'min': 0.0, 'max': 0.0}, {'current': 3000.178, 'min': 0.0, 'max': 0.0}, {'current': 3432.256, 'min': 0.0, 'max': 0.0}, {'current': 2999.802, 'min': 0.0, 'max': 0.0}, {'current': 3005.725, 'min': 0.0, 'max': 0.0}, {'current': 3565.885, 'min': 0.0, 'max': 0.0}, {'current': 2992.339, 'min': 0.0, 'max': 0.0}, {'current': 3104.468, 'min': 0.0, 'max': 0.0}, {'current': 3182.411, 'min': 0.0, 'max': 0.0}, {'current': 3013.217, 'min': 0.0, 'max': 0.0}, {'current': 3131.904, 'min': 0.0, 'max': 0.0}, {'current': 3007.314, 'min': 0.0, 'max': 0.0}, {'current': 2625.175, 'min': 0.0, 'max': 0.0}, {'current': 2653.407, 'min': 0.0, 'max': 0.0}, {'current': 2610.871, 'min': 0.0, 'max': 0.0}, {'current': 2793.855, 'min': 0.0, 'max': 0.0}, {'current': 2619.232, 'min': 0.0, 'max': 0.0}, {'current': 2627.886, 'min': 0.0, 'max': 0.0}, {'current': 2613.748, 'min': 0.0, 'max': 0.0}, {'current': 2318.039, 'min': 0.0, 'max': 0.0}, {'current': 3546.661, 'min': 0.0, 'max': 0.0}, {'current': 3022.56, 'min': 0.0, 'max': 0.0}, {'current': 3035.28, 'min': 0.0, 'max': 0.0}, {'current': 3541.443, 'min': 0.0, 'max': 0.0}, {'current': 3163.55, 'min': 0.0, 'max': 0.0}, {'current': 3023.79, 'min': 0.0, 'max': 0.0}, {'current': 3033.204, 'min': 0.0, 'max': 0.0}, {'current': 3031.695, 'min': 0.0, 'max': 0.0}, {'current': 3034.587, 'min': 0.0, 'max': 0.0}, {'current': 3031.71, 'min': 0.0, 'max': 0.0}, {'current': 3029.116, 'min': 0.0, 'max': 0.0}, {'current': 3293.701, 'min': 0.0, 'max': 0.0}, {'current': 3033.78, 'min': 0.0, 'max': 0.0}, {'current': 3032.034, 'min': 0.0, 'max': 0.0}, {'current': 3057.365, 'min': 0.0, 'max': 0.0}, {'current': 3028.183, 'min': 0.0, 'max': 0.0}, {'current': 3596.439, 'min': 0.0, 'max': 0.0}, {'current': 3327.853, 'min': 0.0, 'max': 0.0}, {'current': 3226.463, 'min': 0.0, 'max': 0.0}, {'current': 3259.361, 'min': 0.0, 'max': 0.0}, {'current': 3284.076, 'min': 0.0, 'max': 0.0}, {'current': 3257.698, 'min': 0.0, 'max': 0.0}, {'current': 3280.289, 'min': 0.0, 'max': 0.0}, {'current': 3258.834, 'min': 0.0, 'max': 0.0}, {'current': 2517.346, 'min': 0.0, 'max': 0.0}, {'current': 2553.091, 'min': 0.0, 'max': 0.0}, {'current': 2552.693, 'min': 0.0, 'max': 0.0}, {'current': 2689.091, 'min': 0.0, 'max': 0.0}, {'current': 2503.301, 'min': 0.0, 'max': 0.0}, {'current': 2524.4, 'min': 0.0, 'max': 0.0}, {'current': 2521.768, 'min': 0.0, 'max': 0.0}, {'current': 2530.364, 'min': 0.0, 'max': 0.0}, {'current': 3258.225, 'min': 0.0, 'max': 0.0}, {'current': 3597.406, 'min': 0.0, 'max': 0.0}, {'current': 3279.658, 'min': 0.0, 'max': 0.0}, {'current': 3292.257, 'min': 0.0, 'max': 0.0}, {'current': 3265.351, 'min': 0.0, 'max': 0.0}, {'current': 3286.715, 'min': 0.0, 'max': 0.0}, {'current': 3268.777, 'min': 0.0, 'max': 0.0}, {'current': 3242.068, 'min': 0.0, 'max': 0.0}, {'current': 3288.908, 'min': 0.0, 'max': 0.0}, {'current': 3278.645, 'min': 0.0, 'max': 0.0}, {'current': 3215.578, 'min': 0.0, 'max': 0.0}, {'current': 3327.101, 'min': 0.0, 'max': 0.0}, {'current': 3357.075, 'min': 0.0, 'max': 0.0}, {'current': 3144.838, 'min': 0.0, 'max': 0.0}, {'current': 3005.517, 'min': 0.0, 'max': 0.0}, {'current': 3597.381, 'min': 0.0, 'max': 0.0}, {'current': 3596.175, 'min': 0.0, 'max': 0.0}, {'current': 3317.258, 'min': 0.0, 'max': 0.0}, {'current': 3310.249, 'min': 0.0, 'max': 0.0}, {'current': 3140.687, 'min': 0.0, 'max': 0.0}, {'current': 3398.954, 'min': 0.0, 'max': 0.0}, {'current': 3318.433, 'min': 0.0, 'max': 0.0}, {'current': 3342.35, 'min': 0.0, 'max': 0.0}, {'current': 3328.575, 'min': 0.0, 'max': 0.0}, {'current': 3499.892, 'min': 0.0, 'max': 0.0}, {'current': 3344.864, 'min': 0.0, 'max': 0.0}, {'current': 3355.802, 'min': 0.0, 'max': 0.0}, {'current': 3305.032, 'min': 0.0, 'max': 0.0}, {'current': 3318.07, 'min': 0.0, 'max': 0.0}, {'current': 3336.546, 'min': 0.0, 'max': 0.0}, {'current': 3335.458, 'min': 0.0, 'max': 0.0}, {'current': 3596.561, 'min': 0.0, 'max': 0.0}, {'current': 2884.473, 'min': 0.0, 'max': 0.0}, {'current': 2934.098, 'min': 0.0, 'max': 0.0}, {'current': 2621.598, 'min': 0.0, 'max': 0.0}, {'current': 2645.184, 'min': 0.0, 'max': 0.0}, {'current': 2480.726, 'min': 0.0, 'max': 0.0}, {'current': 2515.912, 'min': 0.0, 'max': 0.0}, {'current': 2702.57, 'min': 0.0, 'max': 0.0}, {'current': 2671.473, 'min': 0.0, 'max': 0.0}, {'current': 3092.321, 'min': 0.0, 'max': 0.0}, {'current': 3092.204, 'min': 0.0, 'max': 0.0}, {'current': 3217.287, 'min': 0.0, 'max': 0.0}, {'current': 3161.646, 'min': 0.0, 'max': 0.0}, {'current': 3089.273, 'min': 0.0, 'max': 0.0}, {'current': 3088.447, 'min': 0.0, 'max': 0.0}, {'current': 3559.214, 'min': 0.0, 'max': 0.0}, {'current': 3218.611, 'min': 0.0, 'max': 0.0}, {'current': 3021.622, 'min': 0.0, 'max': 0.0}, {'current': 3332.084, 'min': 0.0, 'max': 0.0}, {'current': 3019.854, 'min': 0.0, 'max': 0.0}, {'current': 3213.831, 'min': 0.0, 'max': 0.0}, {'current': 3092.774, 'min': 0.0, 'max': 0.0}, {'current': 3015.868, 'min': 0.0, 'max': 0.0}, {'current': 3082.752, 'min': 0.0, 'max': 0.0}, {'current': 3029.559, 'min': 0.0, 'max': 0.0}, {'current': 2873.142, 'min': 0.0, 'max': 0.0}, {'current': 2215.288, 'min': 0.0, 'max': 0.0}, {'current': 3193.443, 'min': 0.0, 'max': 0.0}, {'current': 2928.717, 'min': 0.0, 'max': 0.0}, {'current': 2990.366, 'min': 0.0, 'max': 0.0}, {'current': 2923.847, 'min': 0.0, 'max': 0.0}, {'current': 2869.364, 'min': 0.0, 'max': 0.0}, {'current': 2859.189, 'min': 0.0, 'max': 0.0}, {'current': 3049.001, 'min': 0.0, 'max': 0.0}, {'current': 2805.206, 'min': 0.0, 'max': 0.0}, {'current': 3094.081, 'min': 0.0, 'max': 0.0}, {'current': 3187.222, 'min': 0.0, 'max': 0.0}, {'current': 2853.945, 'min': 0.0, 'max': 0.0}, {'current': 2944.73, 'min': 0.0, 'max': 0.0}, {'current': 2809.161, 'min': 0.0, 'max': 0.0}, {'current': 2806.787, 'min': 0.0, 'max': 0.0}, {'current': 3024.225, 'min': 0.0, 'max': 0.0}, {'current': 3032.173, 'min': 0.0, 'max': 0.0}, {'current': 3033.195, 'min': 0.0, 'max': 0.0}, {'current': 3561.287, 'min': 0.0, 'max': 0.0}, {'current': 2987.872, 'min': 0.0, 'max': 0.0}, {'current': 3051.252, 'min': 0.0, 'max': 0.0}, {'current': 3379.017, 'min': 0.0, 'max': 0.0}, {'current': 3044.319, 'min': 0.0, 'max': 0.0}, {'current': 3595.62, 'min': 0.0, 'max': 0.0}, {'current': 3261.737, 'min': 0.0, 'max': 0.0}, {'current': 3257.013, 'min': 0.0, 'max': 0.0}, {'current': 3288.317, 'min': 0.0, 'max': 0.0}, {'current': 3268.675, 'min': 0.0, 'max': 0.0}, {'current': 3269.579, 'min': 0.0, 'max': 0.0}, {'current': 3266.213, 'min': 0.0, 'max': 0.0}, {'current': 3265.148, 'min': 0.0, 'max': 0.0}, {'current': 2554.152, 'min': 0.0, 'max': 0.0}, {'current': 2556.63, 'min': 0.0, 'max': 0.0}, {'current': 2540.571, 'min': 0.0, 'max': 0.0}, {'current': 2512.005, 'min': 0.0, 'max': 0.0}, {'current': 2560.992, 'min': 0.0, 'max': 0.0}, {'current': 2526.641, 'min': 0.0, 'max': 0.0}, {'current': 2528.267, 'min': 0.0, 'max': 0.0}, {'current': 2583.061, 'min': 0.0, 'max': 0.0}, {'current': 3257.652, 'min': 0.0, 'max': 0.0}, {'current': 3598.217, 'min': 0.0, 'max': 0.0}, {'current': 3282.578, 'min': 0.0, 'max': 0.0}, {'current': 3261.664, 'min': 0.0, 'max': 0.0}, {'current': 3268.246, 'min': 0.0, 'max': 0.0}, {'current': 3331.937, 'min': 0.0, 'max': 0.0}, {'current': 3265.603, 'min': 0.0, 'max': 0.0}, {'current': 3258.501, 'min': 0.0, 'max': 0.0}, {'current': 3337.045, 'min': 0.0, 'max': 0.0}, {'current': 3259.949, 'min': 0.0, 'max': 0.0}, {'current': 3226.058, 'min': 0.0, 'max': 0.0}, {'current': 3250.308, 'min': 0.0, 'max': 0.0}, {'current': 3208.026, 'min': 0.0, 'max': 0.0}, {'current': 2974.043, 'min': 0.0, 'max': 0.0}, {'current': 3216.706, 'min': 0.0, 'max': 0.0}, {'current': 3598.55, 'min': 0.0, 'max': 0.0}, {'current': 3598.261, 'min': 0.0, 'max': 0.0}, {'current': 3222.637, 'min': 0.0, 'max': 0.0}, {'current': 3227.662, 'min': 0.0, 'max': 0.0}, {'current': 3143.781, 'min': 0.0, 'max': 0.0}, {'current': 3215.567, 'min': 0.0, 'max': 0.0}, {'current': 3250.612, 'min': 0.0, 'max': 0.0}, {'current': 3195.784, 'min': 0.0, 'max': 0.0}, {'current': 3219.289, 'min': 0.0, 'max': 0.0}, {'current': 3289.163, 'min': 0.0, 'max': 0.0}, {'current': 3247.875, 'min': 0.0, 'max': 0.0}, {'current': 3249.112, 'min': 0.0, 'max': 0.0}, {'current': 3209.3, 'min': 0.0, 'max': 0.0}, {'current': 3187.212, 'min': 0.0, 'max': 0.0}, {'current': 3288.003, 'min': 0.0, 'max': 0.0}, {'current': 3169.417, 'min': 0.0, 'max': 0.0}, {'current': 3596.071, 'min': 0.0, 'max': 0.0}], 'disk': {'/': {'total': 119.94140625, 'used': 56.4015998840332}}, 'gpu': 'NVIDIA H100 80GB HBM3', 'gpu_count': 8, 'gpu_devices': [{'name': 'NVIDIA H100 80GB HBM3', 'memory_total': 85520809984}, {'name': 'NVIDIA H100 80GB HBM3', 'memory_total': 85520809984}, {'name': 'NVIDIA H100 80GB HBM3', 'memory_total': 85520809984}, {'name': 'NVIDIA H100 80GB HBM3', 'memory_total': 85520809984}, {'name': 'NVIDIA H100 80GB HBM3', 'memory_total': 85520809984}, {'name': 'NVIDIA H100 80GB HBM3', 'memory_total': 85520809984}, {'name': 'NVIDIA H100 80GB HBM3', 'memory_total': 85520809984}, {'name': 'NVIDIA H100 80GB HBM3', 'memory_total': 85520809984}], 'memory': {'total': 1999.9661331176758}}
|
26 |
+
2024-11-13 19:02:47,022 INFO HandlerThread:1939 [system_monitor.py:probe():224] Finished collecting system info
|
27 |
+
2024-11-13 19:02:47,022 INFO HandlerThread:1939 [system_monitor.py:probe():227] Publishing system info
|
28 |
+
2024-11-13 19:02:47,022 DEBUG HandlerThread:1939 [system_info.py:_save_conda():208] Saving list of conda packages installed into the current environment
|
29 |
+
2024-11-13 19:02:47,775 INFO Thread-12 :1939 [dir_watcher.py:_on_file_created():271] file/dir created: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/conda-environment.yaml
|
30 |
+
2024-11-13 19:03:01,345 DEBUG HandlerThread:1939 [system_info.py:_save_conda():220] Saving conda packages done
|
31 |
+
2024-11-13 19:03:01,345 DEBUG HandlerThread:1939 [system_info.py:_save_code():45] Saving code
|
32 |
+
2024-11-13 19:03:01,349 DEBUG HandlerThread:1939 [system_info.py:_save_code():66] Saving code done
|
33 |
+
2024-11-13 19:03:01,349 DEBUG HandlerThread:1939 [system_info.py:_save_patches():83] Saving git patches
|
34 |
+
2024-11-13 19:03:01,776 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/conda-environment.yaml
|
35 |
+
2024-11-13 19:03:01,776 INFO Thread-12 :1939 [dir_watcher.py:_on_file_created():271] file/dir created: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/code/qlearning_reasoning/training/sft.py
|
36 |
+
2024-11-13 19:03:01,776 INFO Thread-12 :1939 [dir_watcher.py:_on_file_created():271] file/dir created: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/code/qlearning_reasoning
|
37 |
+
2024-11-13 19:03:01,776 INFO Thread-12 :1939 [dir_watcher.py:_on_file_created():271] file/dir created: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/code/qlearning_reasoning/training
|
38 |
+
2024-11-13 19:03:01,776 INFO Thread-12 :1939 [dir_watcher.py:_on_file_created():271] file/dir created: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/code
|
39 |
+
2024-11-13 19:03:01,989 DEBUG HandlerThread:1939 [system_info.py:_save_patches():125] Saving git patches done
|
40 |
+
2024-11-13 19:03:01,991 INFO HandlerThread:1939 [system_monitor.py:probe():229] Finished publishing system info
|
41 |
+
2024-11-13 19:03:01,993 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
42 |
+
2024-11-13 19:03:01,993 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: keepalive
|
43 |
+
2024-11-13 19:03:01,994 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
44 |
+
2024-11-13 19:03:01,994 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: keepalive
|
45 |
+
2024-11-13 19:03:01,994 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
46 |
+
2024-11-13 19:03:01,994 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: keepalive
|
47 |
+
2024-11-13 19:03:01,994 DEBUG SenderThread:1939 [sender.py:send():382] send: files
|
48 |
+
2024-11-13 19:03:01,995 INFO SenderThread:1939 [sender.py:_save_file():1403] saving file wandb-metadata.json with policy now
|
49 |
+
2024-11-13 19:03:01,995 INFO SenderThread:1939 [sender.py:_save_file():1403] saving file code/qlearning_reasoning/training/sft.py with policy now
|
50 |
+
2024-11-13 19:03:01,995 INFO SenderThread:1939 [sender.py:_save_file():1403] saving file diff.patch with policy now
|
51 |
+
2024-11-13 19:03:01,999 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: python_packages
|
52 |
+
2024-11-13 19:03:01,999 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: python_packages
|
53 |
+
2024-11-13 19:03:02,001 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
54 |
+
2024-11-13 19:03:02,003 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
55 |
+
2024-11-13 19:03:02,003 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
56 |
+
2024-11-13 19:03:02,111 DEBUG SenderThread:1939 [sender.py:send():382] send: telemetry
|
57 |
+
2024-11-13 19:03:02,111 DEBUG SenderThread:1939 [sender.py:send():382] send: config
|
58 |
+
2024-11-13 19:03:02,112 DEBUG SenderThread:1939 [sender.py:send():382] send: metric
|
59 |
+
2024-11-13 19:03:02,112 DEBUG SenderThread:1939 [sender.py:send():382] send: telemetry
|
60 |
+
2024-11-13 19:03:02,112 DEBUG SenderThread:1939 [sender.py:send():382] send: metric
|
61 |
+
2024-11-13 19:03:02,113 WARNING SenderThread:1939 [sender.py:send_metric():1354] Seen metric with glob (shouldn't happen)
|
62 |
+
2024-11-13 19:03:02,113 DEBUG SenderThread:1939 [sender.py:send():382] send: telemetry
|
63 |
+
2024-11-13 19:03:02,113 DEBUG SenderThread:1939 [sender.py:send():382] send: telemetry
|
64 |
+
2024-11-13 19:03:02,113 DEBUG SenderThread:1939 [sender.py:send():382] send: config
|
65 |
+
2024-11-13 19:03:02,551 INFO wandb-upload_1:1939 [upload_job.py:push():131] Uploaded file /tmp/tmpm_h7h9v0wandb/pjhigh2e-code/qlearning_reasoning/training/sft.py
|
66 |
+
2024-11-13 19:03:02,554 INFO wandb-upload_0:1939 [upload_job.py:push():131] Uploaded file /tmp/tmpm_h7h9v0wandb/f98paeu8-wandb-metadata.json
|
67 |
+
2024-11-13 19:03:02,556 INFO wandb-upload_2:1939 [upload_job.py:push():131] Uploaded file /tmp/tmpm_h7h9v0wandb/cc8zuviy-diff.patch
|
68 |
+
2024-11-13 19:03:02,776 INFO Thread-12 :1939 [dir_watcher.py:_on_file_created():271] file/dir created: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/diff.patch
|
69 |
+
2024-11-13 19:03:02,776 INFO Thread-12 :1939 [dir_watcher.py:_on_file_created():271] file/dir created: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
70 |
+
2024-11-13 19:03:02,777 INFO Thread-12 :1939 [dir_watcher.py:_on_file_created():271] file/dir created: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/requirements.txt
|
71 |
+
2024-11-13 19:03:02,777 INFO Thread-12 :1939 [dir_watcher.py:_on_file_created():271] file/dir created: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/wandb-metadata.json
|
72 |
+
2024-11-13 19:03:04,777 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
73 |
+
2024-11-13 19:03:07,114 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
74 |
+
2024-11-13 19:03:12,115 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
75 |
+
2024-11-13 19:03:16,999 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
76 |
+
2024-11-13 19:03:17,000 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
77 |
+
2024-11-13 19:03:17,002 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
78 |
+
2024-11-13 19:03:18,090 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
79 |
+
2024-11-13 19:03:18,782 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/config.yaml
|
80 |
+
2024-11-13 19:03:23,250 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
81 |
+
2024-11-13 19:03:26,095 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: partial_history
|
82 |
+
2024-11-13 19:03:26,097 DEBUG SenderThread:1939 [sender.py:send():382] send: metric
|
83 |
+
2024-11-13 19:03:26,097 DEBUG SenderThread:1939 [sender.py:send():382] send: metric
|
84 |
+
2024-11-13 19:03:26,097 DEBUG SenderThread:1939 [sender.py:send():382] send: metric
|
85 |
+
2024-11-13 19:03:26,097 DEBUG SenderThread:1939 [sender.py:send():382] send: metric
|
86 |
+
2024-11-13 19:03:26,097 DEBUG SenderThread:1939 [sender.py:send():382] send: history
|
87 |
+
2024-11-13 19:03:26,098 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: summary_record
|
88 |
+
2024-11-13 19:03:26,099 INFO SenderThread:1939 [sender.py:_save_file():1403] saving file wandb-summary.json with policy end
|
89 |
+
2024-11-13 19:03:26,785 INFO Thread-12 :1939 [dir_watcher.py:_on_file_created():271] file/dir created: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/wandb-summary.json
|
90 |
+
2024-11-13 19:03:28,785 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
91 |
+
2024-11-13 19:03:29,100 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
92 |
+
2024-11-13 19:03:31,999 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
93 |
+
2024-11-13 19:03:32,000 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
94 |
+
2024-11-13 19:03:32,002 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
95 |
+
2024-11-13 19:03:34,147 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
96 |
+
2024-11-13 19:03:39,147 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
97 |
+
2024-11-13 19:03:44,148 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
98 |
+
2024-11-13 19:03:46,955 DEBUG SystemMonitor:1939 [system_monitor.py:_start():172] Starting system metrics aggregation loop
|
99 |
+
2024-11-13 19:03:46,959 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
|
100 |
+
2024-11-13 19:03:47,000 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
101 |
+
2024-11-13 19:03:47,000 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
102 |
+
2024-11-13 19:03:47,043 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
103 |
+
2024-11-13 19:03:49,636 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
104 |
+
2024-11-13 19:03:50,813 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
105 |
+
2024-11-13 19:03:50,814 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/config.yaml
|
106 |
+
2024-11-13 19:03:54,814 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
107 |
+
2024-11-13 19:03:59,814 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
108 |
+
2024-11-13 19:04:02,000 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
109 |
+
2024-11-13 19:04:02,000 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
110 |
+
2024-11-13 19:04:02,002 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
111 |
+
2024-11-13 19:04:05,061 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
112 |
+
2024-11-13 19:04:10,062 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
113 |
+
2024-11-13 19:04:14,822 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
114 |
+
2024-11-13 19:04:15,419 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
115 |
+
2024-11-13 19:04:16,962 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
|
116 |
+
2024-11-13 19:04:17,000 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
117 |
+
2024-11-13 19:04:17,000 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
118 |
+
2024-11-13 19:04:17,002 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
119 |
+
2024-11-13 19:04:21,087 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
120 |
+
2024-11-13 19:04:26,088 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
121 |
+
2024-11-13 19:04:31,089 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
122 |
+
2024-11-13 19:04:32,000 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
123 |
+
2024-11-13 19:04:32,000 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
124 |
+
2024-11-13 19:04:32,043 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
125 |
+
2024-11-13 19:04:36,587 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
126 |
+
2024-11-13 19:04:36,829 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
127 |
+
2024-11-13 19:04:41,588 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
128 |
+
2024-11-13 19:04:46,588 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
129 |
+
2024-11-13 19:04:46,965 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
|
130 |
+
2024-11-13 19:04:47,000 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
131 |
+
2024-11-13 19:04:47,000 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
132 |
+
2024-11-13 19:04:47,002 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
133 |
+
2024-11-13 19:04:52,111 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
134 |
+
2024-11-13 19:04:57,111 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
135 |
+
2024-11-13 19:04:57,383 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: partial_history
|
136 |
+
2024-11-13 19:04:57,384 DEBUG SenderThread:1939 [sender.py:send():382] send: history
|
137 |
+
2024-11-13 19:04:57,385 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: summary_record
|
138 |
+
2024-11-13 19:04:57,385 INFO SenderThread:1939 [sender.py:_save_file():1403] saving file wandb-summary.json with policy end
|
139 |
+
2024-11-13 19:04:57,837 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/wandb-summary.json
|
140 |
+
2024-11-13 19:04:58,837 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
141 |
+
2024-11-13 19:04:59,002 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: partial_history
|
142 |
+
2024-11-13 19:04:59,004 DEBUG SenderThread:1939 [sender.py:send():382] send: metric
|
143 |
+
2024-11-13 19:04:59,005 DEBUG SenderThread:1939 [sender.py:send():382] send: metric
|
144 |
+
2024-11-13 19:04:59,005 DEBUG SenderThread:1939 [sender.py:send():382] send: metric
|
145 |
+
2024-11-13 19:04:59,005 DEBUG SenderThread:1939 [sender.py:send():382] send: metric
|
146 |
+
2024-11-13 19:04:59,006 DEBUG SenderThread:1939 [sender.py:send():382] send: history
|
147 |
+
2024-11-13 19:04:59,006 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: summary_record
|
148 |
+
2024-11-13 19:04:59,006 INFO SenderThread:1939 [sender.py:_save_file():1403] saving file wandb-summary.json with policy end
|
149 |
+
2024-11-13 19:04:59,838 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/wandb-summary.json
|
150 |
+
2024-11-13 19:05:00,838 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
151 |
+
2024-11-13 19:05:02,000 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
152 |
+
2024-11-13 19:05:02,001 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
153 |
+
2024-11-13 19:05:02,002 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
154 |
+
2024-11-13 19:05:03,111 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
155 |
+
2024-11-13 19:05:08,112 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
156 |
+
2024-11-13 19:05:13,113 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
157 |
+
2024-11-13 19:05:16,967 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
|
158 |
+
2024-11-13 19:05:17,000 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
159 |
+
2024-11-13 19:05:17,001 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
160 |
+
2024-11-13 19:05:17,043 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
161 |
+
2024-11-13 19:05:18,123 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
162 |
+
2024-11-13 19:05:22,846 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
163 |
+
2024-11-13 19:05:23,982 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
164 |
+
2024-11-13 19:05:25,847 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/config.yaml
|
165 |
+
2024-11-13 19:05:30,486 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
166 |
+
2024-11-13 19:05:32,000 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
167 |
+
2024-11-13 19:05:32,001 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
168 |
+
2024-11-13 19:05:32,002 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
169 |
+
2024-11-13 19:05:36,117 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
170 |
+
2024-11-13 19:05:41,118 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
171 |
+
2024-11-13 19:05:46,513 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
172 |
+
2024-11-13 19:05:46,854 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
173 |
+
2024-11-13 19:05:46,970 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
|
174 |
+
2024-11-13 19:05:47,000 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
175 |
+
2024-11-13 19:05:47,001 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
176 |
+
2024-11-13 19:05:47,003 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
177 |
+
2024-11-13 19:05:52,097 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
178 |
+
2024-11-13 19:05:57,098 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
179 |
+
2024-11-13 19:06:02,001 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
180 |
+
2024-11-13 19:06:02,001 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
181 |
+
2024-11-13 19:06:02,043 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
182 |
+
2024-11-13 19:06:02,152 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
183 |
+
2024-11-13 19:06:07,153 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
184 |
+
2024-11-13 19:06:08,862 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
185 |
+
2024-11-13 19:06:12,639 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
186 |
+
2024-11-13 19:06:16,973 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
|
187 |
+
2024-11-13 19:06:17,001 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
188 |
+
2024-11-13 19:06:17,002 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
189 |
+
2024-11-13 19:06:17,004 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
190 |
+
2024-11-13 19:06:18,127 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
191 |
+
2024-11-13 19:06:23,127 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
192 |
+
2024-11-13 19:06:28,128 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
193 |
+
2024-11-13 19:06:30,870 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
194 |
+
2024-11-13 19:06:32,001 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
195 |
+
2024-11-13 19:06:32,001 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
196 |
+
2024-11-13 19:06:32,003 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
197 |
+
2024-11-13 19:06:33,143 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
198 |
+
2024-11-13 19:06:38,144 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
199 |
+
2024-11-13 19:06:43,144 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
200 |
+
2024-11-13 19:06:46,975 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
|
201 |
+
2024-11-13 19:06:47,001 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
202 |
+
2024-11-13 19:06:47,001 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
203 |
+
2024-11-13 19:06:47,043 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
204 |
+
2024-11-13 19:06:49,137 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
205 |
+
2024-11-13 19:06:51,931 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: partial_history
|
206 |
+
2024-11-13 19:06:51,932 DEBUG SenderThread:1939 [sender.py:send():382] send: history
|
207 |
+
2024-11-13 19:06:51,933 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: summary_record
|
208 |
+
2024-11-13 19:06:51,934 INFO SenderThread:1939 [sender.py:_save_file():1403] saving file wandb-summary.json with policy end
|
209 |
+
2024-11-13 19:06:52,878 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
210 |
+
2024-11-13 19:06:52,878 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/wandb-summary.json
|
211 |
+
2024-11-13 19:06:53,541 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: partial_history
|
212 |
+
2024-11-13 19:06:53,543 DEBUG SenderThread:1939 [sender.py:send():382] send: history
|
213 |
+
2024-11-13 19:06:53,543 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: summary_record
|
214 |
+
2024-11-13 19:06:53,544 INFO SenderThread:1939 [sender.py:_save_file():1403] saving file wandb-summary.json with policy end
|
215 |
+
2024-11-13 19:06:53,878 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/wandb-summary.json
|
216 |
+
2024-11-13 19:06:54,545 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
217 |
+
2024-11-13 19:06:54,878 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
218 |
+
2024-11-13 19:06:59,546 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
219 |
+
2024-11-13 19:07:02,001 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
220 |
+
2024-11-13 19:07:02,002 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
221 |
+
2024-11-13 19:07:02,004 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
222 |
+
2024-11-13 19:07:05,164 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
223 |
+
2024-11-13 19:07:10,165 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
224 |
+
2024-11-13 19:07:15,165 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
225 |
+
2024-11-13 19:07:16,886 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
226 |
+
2024-11-13 19:07:16,978 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
|
227 |
+
2024-11-13 19:07:17,001 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
228 |
+
2024-11-13 19:07:17,002 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
229 |
+
2024-11-13 19:07:17,003 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
230 |
+
2024-11-13 19:07:21,108 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
231 |
+
2024-11-13 19:07:26,108 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
232 |
+
2024-11-13 19:07:31,109 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
233 |
+
2024-11-13 19:07:32,001 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
234 |
+
2024-11-13 19:07:32,002 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
235 |
+
2024-11-13 19:07:32,043 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
236 |
+
2024-11-13 19:07:36,934 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
237 |
+
2024-11-13 19:07:38,894 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
238 |
+
2024-11-13 19:07:41,935 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
239 |
+
2024-11-13 19:07:46,935 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
240 |
+
2024-11-13 19:07:46,981 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
|
241 |
+
2024-11-13 19:07:47,002 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
242 |
+
2024-11-13 19:07:47,002 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
243 |
+
2024-11-13 19:07:47,004 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
244 |
+
2024-11-13 19:07:52,135 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
245 |
+
2024-11-13 19:07:57,136 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
246 |
+
2024-11-13 19:08:02,136 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
247 |
+
2024-11-13 19:08:02,856 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
248 |
+
2024-11-13 19:08:02,923 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
249 |
+
2024-11-13 19:08:02,924 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
250 |
+
2024-11-13 19:08:08,076 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
251 |
+
2024-11-13 19:08:13,077 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
252 |
+
2024-11-13 19:08:16,985 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
|
253 |
+
2024-11-13 19:08:17,407 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
254 |
+
2024-11-13 19:08:17,407 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
255 |
+
2024-11-13 19:08:17,447 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
256 |
+
2024-11-13 19:08:18,545 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
257 |
+
2024-11-13 19:08:23,546 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
258 |
+
2024-11-13 19:08:28,546 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
259 |
+
2024-11-13 19:08:32,914 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
260 |
+
2024-11-13 19:08:33,843 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
261 |
+
2024-11-13 19:08:34,729 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
262 |
+
2024-11-13 19:08:34,729 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
263 |
+
2024-11-13 19:08:34,771 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
264 |
+
2024-11-13 19:08:39,802 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
265 |
+
2024-11-13 19:08:44,803 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
266 |
+
2024-11-13 19:08:46,987 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
|
267 |
+
2024-11-13 19:08:48,007 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
268 |
+
2024-11-13 19:08:48,007 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
269 |
+
2024-11-13 19:08:48,008 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
270 |
+
2024-11-13 19:08:48,921 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
271 |
+
2024-11-13 19:08:49,945 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
272 |
+
2024-11-13 19:08:50,922 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
273 |
+
2024-11-13 19:08:52,923 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
274 |
+
2024-11-13 19:08:54,924 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
275 |
+
2024-11-13 19:08:55,154 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
276 |
+
2024-11-13 19:08:56,925 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
277 |
+
2024-11-13 19:08:58,926 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
278 |
+
2024-11-13 19:09:00,371 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
279 |
+
2024-11-13 19:09:00,948 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
280 |
+
2024-11-13 19:09:03,950 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
281 |
+
2024-11-13 19:09:05,541 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
282 |
+
2024-11-13 19:09:05,782 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
283 |
+
2024-11-13 19:09:06,102 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
284 |
+
2024-11-13 19:09:06,102 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
285 |
+
2024-11-13 19:09:11,263 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
286 |
+
2024-11-13 19:09:16,264 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
287 |
+
2024-11-13 19:09:16,990 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
|
288 |
+
2024-11-13 19:09:20,961 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
289 |
+
2024-11-13 19:09:22,073 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
290 |
+
2024-11-13 19:09:22,272 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
291 |
+
2024-11-13 19:09:22,272 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
292 |
+
2024-11-13 19:09:27,401 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
293 |
+
2024-11-13 19:09:32,401 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
294 |
+
2024-11-13 19:09:36,126 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
295 |
+
2024-11-13 19:09:36,666 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
296 |
+
2024-11-13 19:09:36,666 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
297 |
+
2024-11-13 19:09:38,052 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
298 |
+
2024-11-13 19:09:39,964 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
299 |
+
2024-11-13 19:09:43,769 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
300 |
+
2024-11-13 19:09:46,992 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
|
301 |
+
2024-11-13 19:09:47,846 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
302 |
+
2024-11-13 19:09:47,847 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
303 |
+
2024-11-13 19:09:47,847 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
304 |
+
2024-11-13 19:09:49,230 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
305 |
+
2024-11-13 19:09:54,231 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
306 |
+
2024-11-13 19:09:59,231 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
307 |
+
2024-11-13 19:10:02,847 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
308 |
+
2024-11-13 19:10:02,848 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
309 |
+
2024-11-13 19:10:02,848 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
310 |
+
2024-11-13 19:10:05,230 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
311 |
+
2024-11-13 19:10:10,231 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
312 |
+
2024-11-13 19:10:11,977 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
313 |
+
2024-11-13 19:10:15,248 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
314 |
+
2024-11-13 19:10:16,996 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
|
315 |
+
2024-11-13 19:10:17,847 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
316 |
+
2024-11-13 19:10:17,848 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
317 |
+
2024-11-13 19:10:17,848 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
318 |
+
2024-11-13 19:10:21,230 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
319 |
+
2024-11-13 19:10:26,231 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
320 |
+
2024-11-13 19:10:31,231 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
321 |
+
2024-11-13 19:10:32,847 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
322 |
+
2024-11-13 19:10:32,847 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
323 |
+
2024-11-13 19:10:32,847 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
324 |
+
2024-11-13 19:10:33,985 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
325 |
+
2024-11-13 19:10:37,230 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
326 |
+
2024-11-13 19:10:42,231 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
327 |
+
2024-11-13 19:10:46,998 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
|
328 |
+
2024-11-13 19:10:47,847 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
329 |
+
2024-11-13 19:10:47,847 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
330 |
+
2024-11-13 19:10:47,887 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
331 |
+
2024-11-13 19:10:47,921 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
332 |
+
2024-11-13 19:10:53,230 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
333 |
+
2024-11-13 19:10:53,575 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: partial_history
|
334 |
+
2024-11-13 19:10:53,576 DEBUG SenderThread:1939 [sender.py:send():382] send: history
|
335 |
+
2024-11-13 19:10:53,576 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: summary_record
|
336 |
+
2024-11-13 19:10:53,577 INFO SenderThread:1939 [sender.py:_save_file():1403] saving file wandb-summary.json with policy end
|
337 |
+
2024-11-13 19:10:53,992 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/wandb-summary.json
|
338 |
+
2024-11-13 19:10:54,993 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
339 |
+
2024-11-13 19:10:55,200 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: partial_history
|
340 |
+
2024-11-13 19:10:55,201 DEBUG SenderThread:1939 [sender.py:send():382] send: history
|
341 |
+
2024-11-13 19:10:55,201 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: summary_record
|
342 |
+
2024-11-13 19:10:55,203 INFO SenderThread:1939 [sender.py:_save_file():1403] saving file wandb-summary.json with policy end
|
343 |
+
2024-11-13 19:10:55,993 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
344 |
+
2024-11-13 19:10:55,994 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/wandb-summary.json
|
345 |
+
2024-11-13 19:10:56,994 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
346 |
+
2024-11-13 19:10:57,994 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
347 |
+
2024-11-13 19:10:59,204 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
348 |
+
2024-11-13 19:11:02,848 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
349 |
+
2024-11-13 19:11:02,848 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
350 |
+
2024-11-13 19:11:02,849 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
351 |
+
2024-11-13 19:11:04,230 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
352 |
+
2024-11-13 19:11:09,231 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
353 |
+
2024-11-13 19:11:14,231 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
354 |
+
2024-11-13 19:11:17,000 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
|
355 |
+
2024-11-13 19:11:17,847 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
356 |
+
2024-11-13 19:11:17,848 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
357 |
+
2024-11-13 19:11:17,848 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
358 |
+
2024-11-13 19:11:19,311 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
359 |
+
2024-11-13 19:11:20,004 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
360 |
+
2024-11-13 19:11:24,312 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
361 |
+
2024-11-13 19:11:29,312 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
362 |
+
2024-11-13 19:11:32,848 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
363 |
+
2024-11-13 19:11:32,848 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
364 |
+
2024-11-13 19:11:32,849 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
365 |
+
2024-11-13 19:11:35,231 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
366 |
+
2024-11-13 19:11:40,232 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
367 |
+
2024-11-13 19:11:44,014 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
368 |
+
2024-11-13 19:11:46,173 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
369 |
+
2024-11-13 19:11:47,003 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
|
370 |
+
2024-11-13 19:11:47,847 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
371 |
+
2024-11-13 19:11:47,847 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
372 |
+
2024-11-13 19:11:47,848 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
373 |
+
2024-11-13 19:11:51,230 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
wandb/debug.log
ADDED
@@ -0,0 +1,30 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
2024-11-13 19:02:46,428 INFO MainThread:173 [wandb_setup.py:_flush():76] Current SDK version is 0.16.3
|
2 |
+
2024-11-13 19:02:46,428 INFO MainThread:173 [wandb_setup.py:_flush():76] Configure stats pid to 173
|
3 |
+
2024-11-13 19:02:46,428 INFO MainThread:173 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
|
4 |
+
2024-11-13 19:02:46,428 INFO MainThread:173 [wandb_setup.py:_flush():76] Loading settings from /opt/ml/code/wandb/settings
|
5 |
+
2024-11-13 19:02:46,428 INFO MainThread:173 [wandb_setup.py:_flush():76] Loading settings from environment variables: {'root_dir': '/opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048', 'api_key': '***REDACTED***', 'project': 'sft-codecontests-1112', 'run_name': 'sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048'}
|
6 |
+
2024-11-13 19:02:46,428 INFO MainThread:173 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
|
7 |
+
2024-11-13 19:02:46,428 INFO MainThread:173 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': 'qlearning_reasoning/training/sft.py', 'program_abspath': '/opt/ml/code/qlearning_reasoning/training/sft.py', 'program': '/opt/ml/code/qlearning_reasoning/training/sft.py'}
|
8 |
+
2024-11-13 19:02:46,428 INFO MainThread:173 [wandb_init.py:_log_setup():526] Logging user logs to /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/logs/debug.log
|
9 |
+
2024-11-13 19:02:46,428 INFO MainThread:173 [wandb_init.py:_log_setup():527] Logging internal logs to /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/logs/debug-internal.log
|
10 |
+
2024-11-13 19:02:46,428 INFO MainThread:173 [wandb_init.py:init():566] calling init triggers
|
11 |
+
2024-11-13 19:02:46,428 INFO MainThread:173 [wandb_init.py:init():573] wandb.init called with sweep_config: {}
|
12 |
+
config: {'sagemaker_training_job_name': 'kushalarora-rvv-main-2024-11-13-17-55-42-664', 'batch_size': 64, 'cache_dir': '/opt/ml/data/input/.cache', 'dataset': 'code-contests', 'gradient_accumulation_steps': 8, 'logging_steps': 5, 'lr': '1e-6', 'max_seq_length': 2048, 'model_name': 'Qwen/Qwen2.5-Coder-7B-Instruct', 'num_train_epochs': 2, 'output_dir': '/opt/ml/output', 'push_to_hub': 'true', 'run_name': 'sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048', 'sagemaker_container_log_level': 20, 'sagemaker_instance_type': 'ml.p5.48xlarge', 'sagemaker_job_name': 'kushalarora-rvv-main-2024-11-13-17-55-42-664', 'sagemaker_program': 'sft.py', 'sagemaker_region': 'us-east-1', 'sagemaker_submit_directory': 's3://tri-ml-datasets/kushalarora-rvv-main-2024-11-13-17-55-42-664/source/sourcedir.tar.gz', 'sagemaker_torch_distributed_enabled': 'true', 'save_steps': 100, 'schedule': 'cosine', 'wandb_project': 'sft-codecontests-1112', 'weight_decay': 0.0, 'additional_framework_parameters': {'sagemaker_instance_type': 'ml.p5.48xlarge', 'sagemaker_torch_distributed_enabled': True}, 'channel_input_dirs': {}, 'current_host': 'algo-1', 'current_instance_group': 'homogeneousCluster', 'current_instance_group_hosts': ['algo-1'], 'current_instance_type': 'ml.p5.48xlarge', 'distribution_hosts': ['algo-1'], 'distribution_instance_groups': ['homogeneousCluster'], 'framework_module': 'sagemaker_pytorch_container.training:main', 'hosts': ['algo-1'], 'hyperparameters': {'batch_size': 64, 'cache_dir': '/opt/ml/data/input/.cache', 'dataset': 'code-contests', 'gradient_accumulation_steps': 8, 'logging_steps': 5, 'lr': '1e-6', 'max_seq_length': 2048, 'model_name': 'Qwen/Qwen2.5-Coder-7B-Instruct', 'num_train_epochs': 2, 'output_dir': '/opt/ml/model/', 'push_to_hub': True, 'run_name': 'sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048', 'save_steps': 100, 'schedule': 'cosine', 'wandb_project': 'sft-codecontests-1112', 'weight_decay': 0.0}, 'input_config_dir': '/opt/ml/input/config', 'input_data_config': {}, 'input_dir': '/opt/ml/input', 'instance_groups': ['homogeneousCluster'], 'instance_groups_dict': {'homogeneousCluster': {'hosts': ['algo-1'], 'instance_group_name': 'homogeneousCluster', 'instance_type': 'ml.p5.48xlarge'}}, 'is_hetero': False, 'is_master': True, 'is_modelparallel_enabled': None, 'is_smddpmprun_installed': False, 'is_smddprun_installed': True, 'job_name': 'kushalarora-rvv-main-2024-11-13-17-55-42-664', 'log_level': 20, 'master_hostname': 'algo-1', 'model_dir': '/opt/ml/model', 'module_dir': 's3://tri-ml-datasets/kushalarora-rvv-main-2024-11-13-17-55-42-664/source/sourcedir.tar.gz', 'module_name': '/opt/ml/code/qlearning_reasoning/training/sft', 'network_interface_name': 'eth0', 'num_cpus': 192, 'num_gpus': 8, 'num_neurons': 0, 'output_data_dir': '/opt/ml/output/data', 'output_intermediate_dir': '/opt/ml/output/intermediate', 'resource_config': {'current_group_name': 'homogeneousCluster', 'current_host': 'algo-1', 'current_instance_type': 'ml.p5.48xlarge', 'hosts': ['algo-1'], 'instance_groups': [{'hosts': ['algo-1'], 'instance_group_name': 'homogeneousCluster', 'instance_type': 'ml.p5.48xlarge'}], 'network_interface_name': 'eth0'}, 'user_entry_point': '/opt/ml/code/qlearning_reasoning/training/sft.py'}
|
13 |
+
2024-11-13 19:02:46,428 INFO MainThread:173 [wandb_init.py:init():616] starting backend
|
14 |
+
2024-11-13 19:02:46,428 INFO MainThread:173 [wandb_init.py:init():620] setting up manager
|
15 |
+
2024-11-13 19:02:46,429 INFO MainThread:173 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
|
16 |
+
2024-11-13 19:02:46,431 INFO MainThread:173 [wandb_init.py:init():628] backend started and connected
|
17 |
+
2024-11-13 19:02:46,434 INFO MainThread:173 [wandb_init.py:init():720] updated telemetry
|
18 |
+
2024-11-13 19:02:46,440 INFO MainThread:173 [wandb_init.py:init():753] communicating run to backend with 90.0 second timeout
|
19 |
+
2024-11-13 19:02:46,781 INFO MainThread:173 [wandb_run.py:_on_init():2262] communicating current version
|
20 |
+
2024-11-13 19:02:46,812 INFO MainThread:173 [wandb_run.py:_on_init():2271] got version response upgrade_message: "wandb version 0.18.6 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
|
21 |
+
|
22 |
+
2024-11-13 19:02:46,813 INFO MainThread:173 [wandb_init.py:init():804] starting run threads in backend
|
23 |
+
2024-11-13 19:03:02,000 INFO MainThread:173 [wandb_run.py:_console_start():2241] atexit reg
|
24 |
+
2024-11-13 19:03:02,000 INFO MainThread:173 [wandb_run.py:_redirect():2096] redirect: wrap_raw
|
25 |
+
2024-11-13 19:03:02,000 INFO MainThread:173 [wandb_run.py:_redirect():2161] Wrapping output streams.
|
26 |
+
2024-11-13 19:03:02,001 INFO MainThread:173 [wandb_run.py:_redirect():2186] Redirects installed.
|
27 |
+
2024-11-13 19:03:02,001 INFO MainThread:173 [wandb_init.py:init():847] run started, returning control to user process
|
28 |
+
2024-11-13 19:03:02,004 INFO MainThread:173 [wandb_run.py:_config_callback():1343] config_cb None None {'vocab_size': 151671, 'max_position_embeddings': 32768, 'hidden_size': 3584, 'intermediate_size': 18944, 'num_hidden_layers': 28, 'num_attention_heads': 28, 'use_sliding_window': False, 'sliding_window': None, 'max_window_layers': 28, 'num_key_value_heads': 4, 'hidden_act': 'silu', 'initializer_range': 0.02, 'rms_norm_eps': 1e-06, 'use_cache': False, 'rope_theta': 1000000.0, 'rope_scaling': None, 'attention_dropout': 0.0, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'bfloat16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['Qwen2ForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': 151643, 'eos_token_id': 151645, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'Qwen/Qwen2.5-Coder-7B-Instruct', '_attn_implementation_autoset': True, 'transformers_version': '4.46.0', 'model_type': 'qwen2', 'output_dir': '/opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'eval_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 8, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 8, 'eval_accumulation_steps': None, 'eval_delay': 0, 'torch_empty_cache_steps': None, 'learning_rate': 1e-06, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 2, 'max_steps': -1, 'lr_scheduler_type': 'cosine', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/runs/Nov13_19-02-15_algo-1', 'logging_strategy': 'steps', 'logging_first_step': True, 'logging_steps': 5, 'logging_nan_inf_filter': True, 'save_strategy': 'epoch', 'save_steps': 100, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'restore_callback_states_from_checkpoint': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': 5, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': 'sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': ['full_shard', 'auto_wrap'], 'fsdp_min_num_params': 0, 'fsdp_config': {'activation_checkpointing': False, 'auto_wrap_policy': 'TRANSFORMER_BASED_WRAP', 'backward_prefetch': 'BACKWARD_PRE', 'cpu_ram_efficient_loading': True, 'forward_prefetch': False, 'offload_params': False, 'sharding_strategy': 'FULL_SHARD', 'state_dict_type': 'FULL_STATE_DICT', 'sync_module_states': True, 'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': 'sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_', 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': {'use_reentrant': False}, 'include_inputs_for_metrics': False, 'include_for_metrics': [], 'eval_do_concat_batches': True, 'fp16_backend': 'auto', 'evaluation_strategy': None, 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None, 'batch_eval_metrics': False, 'eval_on_start': False, 'use_liger_kernel': False, 'eval_use_gather_object': False, 'dataset_text_field': 'text', 'packing': False, 'max_seq_length': 2048, 'dataset_num_proc': None, 'dataset_batch_size': 64, 'model_init_kwargs': None, 'dataset_kwargs': {}, 'eval_packing': None, 'num_of_sequences': 1024, 'chars_per_token': '<CHARS_PER_TOKEN>', 'use_liger': False}
|
29 |
+
2024-11-13 19:03:02,008 INFO MainThread:173 [wandb_config.py:__setitem__():151] config set model/num_parameters = 951599936 - <bound method Run._config_callback of <wandb.sdk.wandb_run.Run object at 0x7fb464957310>>
|
30 |
+
2024-11-13 19:03:02,009 INFO MainThread:173 [wandb_run.py:_config_callback():1343] config_cb model/num_parameters 951599936 None
|
wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/code/qlearning_reasoning/training/sft.py
ADDED
@@ -0,0 +1,249 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
from datasets import load_dataset, concatenate_datasets
|
2 |
+
from transformers import AutoTokenizer, AutoModelForCausalLM, TrainingArguments
|
3 |
+
import torch
|
4 |
+
from transformers import get_constant_schedule_with_warmup, get_cosine_schedule_with_warmup, get_linear_schedule_with_warmup, get_inverse_sqrt_schedule
|
5 |
+
from trl import SFTTrainer, SFTConfig, DataCollatorForCompletionOnlyLM
|
6 |
+
import os
|
7 |
+
from textwrap import dedent
|
8 |
+
from absl import app, flags
|
9 |
+
FLAGS = flags.FLAGS
|
10 |
+
import warnings
|
11 |
+
warnings.filterwarnings("ignore")
|
12 |
+
from trl import setup_chat_format
|
13 |
+
|
14 |
+
ALL_DATASET_NAMES = [
|
15 |
+
'prm800k',
|
16 |
+
'prm800k-improvement',
|
17 |
+
'math',
|
18 |
+
'math-v2',
|
19 |
+
'math-prm800k',
|
20 |
+
'math-revision',
|
21 |
+
'og-math',
|
22 |
+
'leetcode',
|
23 |
+
'taco',
|
24 |
+
'math-revision-onesamp',
|
25 |
+
'math-revision-onesamp-overlap',
|
26 |
+
'math-revision-onesamp-correct-overlap',
|
27 |
+
'math-revision-onesamp-correct-overlap',
|
28 |
+
'code-contests'
|
29 |
+
]
|
30 |
+
|
31 |
+
flags.DEFINE_enum('dataset', 'prm800k', ALL_DATASET_NAMES, 'Name of the dataset')
|
32 |
+
flags.DEFINE_string('cache_dir', '/scratch/bclj/asingh15/.cache', 'Directory to store cache files')
|
33 |
+
flags.DEFINE_string('output_dir', '/scratch/bclj/asingh15/sft_checkpoints/', 'Directory to store output files')
|
34 |
+
flags.DEFINE_string('run_name', 'sft', 'Name of the run')
|
35 |
+
flags.DEFINE_string('model_name', 'deepseek-ai/DeepSeek-Coder-V2-Lite-Instruct', 'Name of the model')
|
36 |
+
flags.DEFINE_string('token', None, 'Hugging Face API token')
|
37 |
+
flags.DEFINE_string('wandb_project', 'sft-prm800k-pregen-0501', 'Wandb project name')
|
38 |
+
flags.DEFINE_string('optim', 'adamw_torch', 'Optimizer')
|
39 |
+
flags.DEFINE_string('schedule', 'constant', 'Learning rate schedule')
|
40 |
+
flags.DEFINE_float('lr', 5e-5, 'Learning rate')
|
41 |
+
flags.DEFINE_float('weight_decay', 0.0, 'Weight decay')
|
42 |
+
flags.DEFINE_integer('warmup_steps', 0, 'Number of warmup steps')
|
43 |
+
flags.DEFINE_integer('num_train_epochs', 1, 'Number of training epochs')
|
44 |
+
flags.DEFINE_integer('batch_size', 256, 'Batch size')
|
45 |
+
flags.DEFINE_integer('gradient_accumulation_steps', 64, 'Gradient accumulation steps')
|
46 |
+
flags.DEFINE_integer('max_seq_length', 1024, 'Maximum sequence length')
|
47 |
+
flags.DEFINE_integer('logging_steps', 100, 'Logging steps')
|
48 |
+
flags.DEFINE_integer('save_steps', 1000, 'Save steps')
|
49 |
+
flags.DEFINE_integer('seed', 42, 'Random seed')
|
50 |
+
flags.DEFINE_string('fsdp', "full_shard auto_wrap", 'If to use FSDP')
|
51 |
+
flags.DEFINE_string('fsdp_config', "qlearning_reasoning/accelerate_configs/fsdp_config.json", 'FSDP config to use.')
|
52 |
+
flags.DEFINE_boolean("push_to_hub", False, "Should the model be uploaded to Huggingface hub?")
|
53 |
+
flags.DEFINE_boolean('add_step_tokens', True, 'Add step tokens')
|
54 |
+
flags.DEFINE_string("hub_model_id", None, "Hub model id. If not defined it will be truncated run name.")
|
55 |
+
|
56 |
+
START_OF_STEP_TOKEN = '[STEP]'
|
57 |
+
END_OF_STEP_TOKEN = '[/STEP]'
|
58 |
+
START_OF_TURN_TOKEN = '[TURN]'
|
59 |
+
END_OF_TURN_TOKEN = '[/TURN]'
|
60 |
+
START_OF_REVISION_TOKEN = '[REVISION]'
|
61 |
+
END_OF_REVISION_TOKEN = '[/REVISION]'
|
62 |
+
|
63 |
+
def main(_):
|
64 |
+
|
65 |
+
if FLAGS.hub_model_id is None:
|
66 |
+
FLAGS.hub_model_id = FLAGS.run_name[:96]
|
67 |
+
|
68 |
+
if not os.path.exists(FLAGS.output_dir):
|
69 |
+
os.makedirs(FLAGS.output_dir, exist_ok=True)
|
70 |
+
os.makedirs(f"{FLAGS.output_dir}/{FLAGS.run_name}", exist_ok=True)
|
71 |
+
|
72 |
+
os.environ["WANDB_PROJECT"] = FLAGS.wandb_project
|
73 |
+
os.environ["WANDB_DIR"] = f"{FLAGS.output_dir}/{FLAGS.run_name}"
|
74 |
+
os.environ["WANDB_NAME"] = FLAGS.run_name
|
75 |
+
os.environ["HF_DATASETS_CACHE"] = FLAGS.cache_dir
|
76 |
+
|
77 |
+
shared_kwargs = dict(
|
78 |
+
cache_dir=FLAGS.cache_dir,
|
79 |
+
trust_remote_code=True,
|
80 |
+
)
|
81 |
+
tokenizer = AutoTokenizer.from_pretrained(
|
82 |
+
FLAGS.model_name,
|
83 |
+
**shared_kwargs,
|
84 |
+
)
|
85 |
+
if tokenizer.pad_token is None:
|
86 |
+
tokenizer.pad_token = tokenizer.eos_token
|
87 |
+
|
88 |
+
if FLAGS.add_step_tokens:
|
89 |
+
tokenizer.add_special_tokens(
|
90 |
+
{'additional_special_tokens': [START_OF_STEP_TOKEN, END_OF_STEP_TOKEN, START_OF_TURN_TOKEN, END_OF_TURN_TOKEN, START_OF_REVISION_TOKEN, END_OF_REVISION_TOKEN]}
|
91 |
+
)
|
92 |
+
|
93 |
+
print(f"Tokenizer vocab size: {len(tokenizer)}")
|
94 |
+
print('EOS token:', tokenizer.eos_token)
|
95 |
+
print('PAD token:', tokenizer.pad_token)
|
96 |
+
print('Special tokens:', tokenizer.special_tokens_map)
|
97 |
+
|
98 |
+
instruction_template = "# Question\n\n"
|
99 |
+
if FLAGS.dataset == 'prm800k':
|
100 |
+
dataset_name = 'Asap7772/sft_prm800k_processed'
|
101 |
+
response_template = "# Solution\n\n"
|
102 |
+
elif FLAGS.dataset == 'math':
|
103 |
+
dataset_name = 'Asap7772/sft_math_processed'
|
104 |
+
response_template = "# Solution\n\n"
|
105 |
+
elif FLAGS.dataset == 'math-prm800k':
|
106 |
+
dataset_name = ['Asap7772/sft_prm800k_processed', 'Asap7772/sft_math_processed']
|
107 |
+
response_template = "# Solution\n\n"
|
108 |
+
elif FLAGS.dataset == 'og-math':
|
109 |
+
dataset_name = 'Asap7772/hendrycks_math_sft_formatted'
|
110 |
+
response_template = "# Solution\n\n"
|
111 |
+
elif FLAGS.dataset == 'leetcode':
|
112 |
+
dataset_name = 'Asap7772/leetcode-rosetta-processed'
|
113 |
+
instruction_template = "# Question"
|
114 |
+
response_template = "# Solution"
|
115 |
+
elif FLAGS.dataset == 'prm800k-improvement':
|
116 |
+
dataset_name = 'Asap7772/prm800k_improvement_groundtruth_values'
|
117 |
+
instruction_template = 'Solution Steps:'
|
118 |
+
response_template = "Improved Step:"
|
119 |
+
elif FLAGS.dataset == 'taco':
|
120 |
+
dataset_name = 'Asap7772/TACO-Processed'
|
121 |
+
instruction_template = "# Question"
|
122 |
+
response_template = "# Solution"
|
123 |
+
elif FLAGS.dataset == 'math-revision':
|
124 |
+
dataset_name = 'Asap7772/hendrycks-math-mc-llama-sftnoic-revisions-filtered'
|
125 |
+
instruction_template = "# Question"
|
126 |
+
response_template = f"{END_OF_TURN_TOKEN}\n\n{START_OF_TURN_TOKEN}"
|
127 |
+
elif FLAGS.dataset == 'math-revision-ic':
|
128 |
+
dataset_name = 'Asap7772/hendrycks-math-mc-llama-revisions-filtered'
|
129 |
+
instruction_template = "# Question"
|
130 |
+
response_template = f"{END_OF_TURN_TOKEN}\n\n{START_OF_TURN_TOKEN}"
|
131 |
+
elif FLAGS.dataset == 'math-revision-onesamp':
|
132 |
+
dataset_name = 'Asap7772/hendrycks-math-onesamp-revision'
|
133 |
+
instruction_template = "# Question"
|
134 |
+
response_template = "# Solution"
|
135 |
+
elif FLAGS.dataset == 'math-revision-onesamp-overlap':
|
136 |
+
dataset_name = 'Asap7772/hendrycks-math-onesamp-revision-overlap'
|
137 |
+
instruction_template = "# Question"
|
138 |
+
response_template = "# Solution"
|
139 |
+
elif FLAGS.dataset == 'math-v2':
|
140 |
+
dataset_name = 'Asap7772/sft_math_405b'
|
141 |
+
instruction_template = "# Question"
|
142 |
+
response_template = "# Solution"
|
143 |
+
elif FLAGS.dataset == 'math-revision-onesamp-correct-overlap':
|
144 |
+
dataset_name = 'Asap7772/hendrycks-math-onesamp-revision-correct-overlap'
|
145 |
+
instruction_template = "# Question"
|
146 |
+
response_template = "# Solution"
|
147 |
+
elif FLAGS.dataset == 'code-contests':
|
148 |
+
dataset_name = 'Asap7772/code_contests'
|
149 |
+
instruction_template = "Q:"
|
150 |
+
response_template = "A:"
|
151 |
+
else:
|
152 |
+
raise ValueError(f"Invalid dataset: {FLAGS.dataset}")
|
153 |
+
|
154 |
+
if isinstance(dataset_name, list):
|
155 |
+
all_datasets = [load_dataset(name, cache_dir=FLAGS.cache_dir) for name in dataset_name]
|
156 |
+
train_dataset = concatenate_datasets([dataset['train'] for dataset in all_datasets])
|
157 |
+
eval_dataset = concatenate_datasets([dataset['test'] for dataset in all_datasets])
|
158 |
+
print(f"example: {train_dataset[0]}")
|
159 |
+
elif isinstance(dataset_name, str):
|
160 |
+
dataset_dict = load_dataset(dataset_name, cache_dir=FLAGS.cache_dir)
|
161 |
+
train_dataset, eval_dataset = dataset_dict['train'], dataset_dict['test']
|
162 |
+
else:
|
163 |
+
raise ValueError(f"Invalid dataset name: {dataset_name}")
|
164 |
+
train_dataset, eval_dataset = train_dataset.shuffle(seed=FLAGS.seed), eval_dataset.shuffle(seed=FLAGS.seed)
|
165 |
+
|
166 |
+
num_train_epochs = FLAGS.num_train_epochs
|
167 |
+
|
168 |
+
def formatting_prompts_func(example):
|
169 |
+
output_texts = []
|
170 |
+
for i in range(len(example['text'])):
|
171 |
+
if FLAGS.dataset in ALL_DATASET_NAMES:
|
172 |
+
text = f"{example['text'][i]}{tokenizer.eos_token}"
|
173 |
+
else:
|
174 |
+
raise ValueError(f"Invalid dataset: {FLAGS.dataset}")
|
175 |
+
|
176 |
+
text = dedent(text).strip()
|
177 |
+
output_texts.append(text)
|
178 |
+
return output_texts
|
179 |
+
format_example = formatting_prompts_func(train_dataset[0:1])
|
180 |
+
print(f"Example formatted prompt: \n\n{format_example[0]}\n\n")
|
181 |
+
assert instruction_template in format_example[0], f"Instruction template not found in text."
|
182 |
+
assert response_template in format_example[0], f"Response template not found in text."
|
183 |
+
assert tokenizer.eos_token in format_example[0], f"EOS token not found in text."
|
184 |
+
|
185 |
+
instruction_template_ids = tokenizer.encode(instruction_template, add_special_tokens=False)[:2]
|
186 |
+
response_template_ids = tokenizer.encode(response_template, add_special_tokens=False)[:2]
|
187 |
+
collator = DataCollatorForCompletionOnlyLM(instruction_template=instruction_template_ids, response_template=response_template_ids, tokenizer=tokenizer, mlm=False)
|
188 |
+
|
189 |
+
model = AutoModelForCausalLM.from_pretrained(
|
190 |
+
FLAGS.model_name,
|
191 |
+
low_cpu_mem_usage=True,
|
192 |
+
torch_dtype=torch.bfloat16,
|
193 |
+
**shared_kwargs,
|
194 |
+
)
|
195 |
+
model.resize_token_embeddings(len(tokenizer))
|
196 |
+
if getattr(model, "config", None) is not None:
|
197 |
+
model.config.pad_token_id = tokenizer.pad_token_id
|
198 |
+
model.config.bos_token_id = tokenizer.bos_token_id
|
199 |
+
model.config.eos_token_id = tokenizer.eos_token_id
|
200 |
+
# Update the generation config to use the new eos & bos token
|
201 |
+
if getattr(model, "generation_config", None) is not None:
|
202 |
+
model.generation_config.bos_token_id = tokenizer.bos_token_id
|
203 |
+
model.generation_config.eos_token_id = tokenizer.eos_token_id
|
204 |
+
model.generation_config.pad_token_id = tokenizer.pad_token_id
|
205 |
+
|
206 |
+
model.config.use_cache = False
|
207 |
+
model.gradient_checkpointing_enable()
|
208 |
+
|
209 |
+
transformer_args = SFTConfig(
|
210 |
+
run_name=FLAGS.run_name,
|
211 |
+
output_dir=f"{FLAGS.output_dir}/{FLAGS.run_name}",
|
212 |
+
learning_rate=FLAGS.lr,
|
213 |
+
eval_strategy="steps",
|
214 |
+
logging_strategy="steps",
|
215 |
+
logging_first_step=True,
|
216 |
+
save_strategy="epoch",
|
217 |
+
dataset_batch_size=FLAGS.batch_size,
|
218 |
+
max_seq_length=FLAGS.max_seq_length,
|
219 |
+
gradient_accumulation_steps=FLAGS.gradient_accumulation_steps,
|
220 |
+
gradient_checkpointing_kwargs={'use_reentrant':False},
|
221 |
+
logging_steps=FLAGS.logging_steps,
|
222 |
+
save_steps=FLAGS.save_steps,
|
223 |
+
lr_scheduler_type=FLAGS.schedule,
|
224 |
+
optim=FLAGS.optim,
|
225 |
+
weight_decay=FLAGS.weight_decay,
|
226 |
+
warmup_steps=FLAGS.warmup_steps,
|
227 |
+
fsdp=FLAGS.fsdp,
|
228 |
+
fsdp_config=FLAGS.fsdp_config if FLAGS.fsdp != '' else None,
|
229 |
+
push_to_hub=FLAGS.push_to_hub,
|
230 |
+
num_train_epochs=num_train_epochs,
|
231 |
+
hub_model_id=FLAGS.hub_model_id
|
232 |
+
)
|
233 |
+
|
234 |
+
flags_dict = FLAGS.flag_values_dict()
|
235 |
+
trainer = SFTTrainer(
|
236 |
+
model=model,
|
237 |
+
args=transformer_args,
|
238 |
+
tokenizer=tokenizer,
|
239 |
+
train_dataset=train_dataset,
|
240 |
+
eval_dataset=eval_dataset,
|
241 |
+
formatting_func=formatting_prompts_func,
|
242 |
+
data_collator=collator,
|
243 |
+
dataset_batch_size=FLAGS.batch_size,
|
244 |
+
)
|
245 |
+
|
246 |
+
trainer.train()
|
247 |
+
|
248 |
+
if __name__ == '__main__':
|
249 |
+
app.run(main)
|
wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/conda-environment.yaml
ADDED
@@ -0,0 +1,515 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
name: base
|
2 |
+
channels:
|
3 |
+
- fastai
|
4 |
+
- nvidia
|
5 |
+
- https://aws-ml-conda.s3.us-west-2.amazonaws.com
|
6 |
+
- conda-forge
|
7 |
+
dependencies:
|
8 |
+
- _libgcc_mutex=0.1=conda_forge
|
9 |
+
- _openmp_mutex=4.5=2_kmp_llvm
|
10 |
+
- alsa-lib=1.2.12=h4ab18f5_0
|
11 |
+
- annotated-types=0.7.0=pyhd8ed1ab_0
|
12 |
+
- asttokens=2.4.1=pyhd8ed1ab_0
|
13 |
+
- attr=2.5.1=h166bdaf_1
|
14 |
+
- aws-ofi-nccl=1.9.1=aws_efa1.26.1_0
|
15 |
+
- blas=1.0=mkl
|
16 |
+
- boltons=23.0.0=pyhd8ed1ab_0
|
17 |
+
- brotli-bin=1.1.0=hd590300_0
|
18 |
+
- bzip2=1.0.8=h7f98852_4
|
19 |
+
- c-ares=1.32.3=h4bc722e_0
|
20 |
+
- ca-certificates=2024.7.4=hbcca054_0
|
21 |
+
- cached-property=1.5.2=hd8ed1ab_1
|
22 |
+
- cached_property=1.5.2=pyha770c72_1
|
23 |
+
- cairo=1.18.0=h3faef2a_0
|
24 |
+
- catalogue=2.0.10=py310hff52083_0
|
25 |
+
- certifi=2024.7.4=pyhd8ed1ab_0
|
26 |
+
- cffi=1.15.1=py310h255011f_3
|
27 |
+
- charset-normalizer=3.2.0=pyhd8ed1ab_0
|
28 |
+
- click=8.1.7=unix_pyh707e725_0
|
29 |
+
- cloudpathlib=0.18.1=pyhd8ed1ab_0
|
30 |
+
- colorama=0.4.6=pyhd8ed1ab_0
|
31 |
+
- comm=0.2.2=pyhd8ed1ab_0
|
32 |
+
- conda=23.3.1=py310hff52083_0
|
33 |
+
- conda-libmamba-solver=23.3.0=pyhd8ed1ab_0
|
34 |
+
- conda-package-handling=2.2.0=pyh38be061_0
|
35 |
+
- conda-package-streaming=0.9.0=pyhd8ed1ab_0
|
36 |
+
- confection=0.1.4=py310h17c5347_0
|
37 |
+
- contourpy=1.2.1=py310hd41b1e2_0
|
38 |
+
- cuda-cccl=12.4.127=0
|
39 |
+
- cuda-compiler=12.1.1=0
|
40 |
+
- cuda-cudart=12.1.105=0
|
41 |
+
- cuda-cudart-dev=12.1.105=0
|
42 |
+
- cuda-cudart-static=12.1.105=0
|
43 |
+
- cuda-cuobjdump=12.4.127=0
|
44 |
+
- cuda-cupti=12.1.105=0
|
45 |
+
- cuda-cupti-static=12.1.105=0
|
46 |
+
- cuda-cuxxfilt=12.4.127=0
|
47 |
+
- cuda-driver-dev=12.4.127=0
|
48 |
+
- cuda-libraries=12.1.0=0
|
49 |
+
- cuda-libraries-dev=12.1.0=0
|
50 |
+
- cuda-libraries-static=12.1.1=0
|
51 |
+
- cuda-nvcc=12.1.105=0
|
52 |
+
- cuda-nvprune=12.4.127=0
|
53 |
+
- cuda-nvrtc=12.1.105=0
|
54 |
+
- cuda-nvrtc-dev=12.1.105=0
|
55 |
+
- cuda-nvrtc-static=12.1.105=0
|
56 |
+
- cuda-nvtx=12.1.105=0
|
57 |
+
- cuda-opencl=12.4.127=0
|
58 |
+
- cuda-opencl-dev=12.4.127=0
|
59 |
+
- cuda-profiler-api=12.4.127=0
|
60 |
+
- cuda-runtime=12.1.0=0
|
61 |
+
- cuda-version=12.1=h1d6eff3_3
|
62 |
+
- curl=8.8.0=he654da7_1
|
63 |
+
- cycler=0.12.1=pyhd8ed1ab_0
|
64 |
+
- cymem=2.0.8=py310hc6cd4ac_1
|
65 |
+
- cython=3.0.11=py310hea249c9_0
|
66 |
+
- cython-blis=0.7.10=py310h1f7b6fc_2
|
67 |
+
- dbus=1.13.6=h5008d03_3
|
68 |
+
- debugpy=1.8.5=py310hea249c9_0
|
69 |
+
- decorator=5.1.1=pyhd8ed1ab_0
|
70 |
+
- docutils=0.16=py310hff52083_4
|
71 |
+
- exceptiongroup=1.2.2=pyhd8ed1ab_0
|
72 |
+
- executing=2.0.1=pyhd8ed1ab_0
|
73 |
+
- expat=2.6.2=h59595ed_0
|
74 |
+
- fastai=2.7.16=py_0
|
75 |
+
- fastcore=1.5.55=py_0
|
76 |
+
- fastdownload=0.0.7=py_0
|
77 |
+
- fastprogress=1.0.3=py_0
|
78 |
+
- ffmpeg=4.2=h3fd9d12_1
|
79 |
+
- filelock=3.15.4=pyhd8ed1ab_0
|
80 |
+
- fmt=9.1.0=h924138e_0
|
81 |
+
- font-ttf-dejavu-sans-mono=2.37=hab24e00_0
|
82 |
+
- font-ttf-inconsolata=3.000=h77eed37_0
|
83 |
+
- font-ttf-source-code-pro=2.038=h77eed37_0
|
84 |
+
- font-ttf-ubuntu=0.83=h77eed37_2
|
85 |
+
- fontconfig=2.14.2=h14ed4e7_0
|
86 |
+
- fonts-conda-ecosystem=1=0
|
87 |
+
- fonts-conda-forge=1=0
|
88 |
+
- fonttools=4.53.1=py310h5b4e0ec_0
|
89 |
+
- freetype=2.12.1=h267a509_2
|
90 |
+
- gettext=0.22.5=h59595ed_2
|
91 |
+
- gettext-tools=0.22.5=h59595ed_2
|
92 |
+
- glib=2.80.2=hf974151_0
|
93 |
+
- glib-tools=2.80.2=hb6ce0ca_0
|
94 |
+
- gmp=6.3.0=hac33072_2
|
95 |
+
- gmpy2=2.1.5=py310hc7909c9_1
|
96 |
+
- gnutls=3.6.15=he1e5248_0
|
97 |
+
- graphite2=1.3.13=h59595ed_1003
|
98 |
+
- gst-plugins-base=1.22.9=hfa15dee_1
|
99 |
+
- gstreamer=1.22.9=h98fc4e7_1
|
100 |
+
- h5py=3.11.0=nompi_py310hf054cd7_102
|
101 |
+
- harfbuzz=8.5.0=hfac3d4d_0
|
102 |
+
- hdf5=1.14.3=nompi_hdf9ad27_105
|
103 |
+
- hwloc=2.9.2=h2bc3f7f_0
|
104 |
+
- icu=73.2=h59595ed_0
|
105 |
+
- idna=3.7=pyhd8ed1ab_0
|
106 |
+
- importlib_metadata=8.2.0=hd8ed1ab_0
|
107 |
+
- jedi=0.19.1=pyhd8ed1ab_0
|
108 |
+
- jinja2=3.1.4=pyhd8ed1ab_0
|
109 |
+
- jmespath=1.0.1=pyhd8ed1ab_0
|
110 |
+
- joblib=1.4.2=pyhd8ed1ab_0
|
111 |
+
- jsonpatch=1.32=pyhd8ed1ab_0
|
112 |
+
- jsonpointer=2.0=py_0
|
113 |
+
- jupyter_client=8.6.2=pyhd8ed1ab_0
|
114 |
+
- jupyter_core=5.7.2=py310hff52083_0
|
115 |
+
- keyutils=1.6.1=h166bdaf_0
|
116 |
+
- kiwisolver=1.4.5=py310hd41b1e2_1
|
117 |
+
- krb5=1.21.3=h659f571_0
|
118 |
+
- lame=3.100=h166bdaf_1003
|
119 |
+
- langcodes=3.4.0=pyhd8ed1ab_0
|
120 |
+
- language-data=1.2.0=pyhd8ed1ab_0
|
121 |
+
- lcms2=2.16=hb7c19ff_0
|
122 |
+
- ld_impl_linux-64=2.40=h41732ed_0
|
123 |
+
- lerc=4.0.0=h27087fc_0
|
124 |
+
- libaec=1.1.3=h59595ed_0
|
125 |
+
- libarchive=3.6.2=h039dbb9_1
|
126 |
+
- libasprintf=0.22.5=h661eb56_2
|
127 |
+
- libasprintf-devel=0.22.5=h661eb56_2
|
128 |
+
- libblas=3.9.0=21_linux64_mkl
|
129 |
+
- libbrotlicommon=1.1.0=hd590300_0
|
130 |
+
- libbrotlidec=1.1.0=hd590300_0
|
131 |
+
- libbrotlienc=1.1.0=hd590300_0
|
132 |
+
- libcap=2.69=h0f662aa_0
|
133 |
+
- libcblas=3.9.0=21_linux64_mkl
|
134 |
+
- libclang=15.0.7=default_h127d8a8_5
|
135 |
+
- libclang13=15.0.7=default_h5d6823c_5
|
136 |
+
- libcublas=12.1.0.26=0
|
137 |
+
- libcublas-dev=12.1.0.26=0
|
138 |
+
- libcublas-static=12.1.3.1=hd3aeb46_0
|
139 |
+
- libcufft=11.0.2.4=0
|
140 |
+
- libcufft-dev=11.0.2.4=0
|
141 |
+
- libcufft-static=11.0.2.54=hd3aeb46_0
|
142 |
+
- libcufile=1.9.1.3=0
|
143 |
+
- libcufile-dev=1.9.1.3=0
|
144 |
+
- libcufile-static=1.9.1.3=0
|
145 |
+
- libcups=2.3.3=h4637d8d_4
|
146 |
+
- libcurand=10.3.5.147=0
|
147 |
+
- libcurand-dev=10.3.5.147=0
|
148 |
+
- libcurand-static=10.3.5.147=0
|
149 |
+
- libcurl=8.8.0=hca28451_1
|
150 |
+
- libcusolver=11.4.4.55=0
|
151 |
+
- libcusolver-dev=11.4.4.55=0
|
152 |
+
- libcusolver-static=11.4.5.107=hd3aeb46_0
|
153 |
+
- libcusparse=12.0.2.55=0
|
154 |
+
- libcusparse-dev=12.0.2.55=0
|
155 |
+
- libcusparse-static=12.1.0.106=hd3aeb46_0
|
156 |
+
- libdeflate=1.20=hd590300_0
|
157 |
+
- libedit=3.1.20191231=he28a2e2_2
|
158 |
+
- libev=4.33=h516909a_1
|
159 |
+
- libevent=2.1.12=hf998b51_1
|
160 |
+
- libexpat=2.6.2=h59595ed_0
|
161 |
+
- libffi=3.4.2=h7f98852_5
|
162 |
+
- libflac=1.4.3=h59595ed_0
|
163 |
+
- libgcc=7.2.0=h69d50b8_2
|
164 |
+
- libgcc-ng=13.1.0=he5830b7_0
|
165 |
+
- libgcrypt=1.11.0=h4ab18f5_1
|
166 |
+
- libgettextpo=0.22.5=h59595ed_2
|
167 |
+
- libgettextpo-devel=0.22.5=h59595ed_2
|
168 |
+
- libgfortran-ng=13.2.0=h69a702a_0
|
169 |
+
- libgfortran5=13.2.0=ha4646dd_0
|
170 |
+
- libglib=2.80.2=hf974151_0
|
171 |
+
- libgomp=13.1.0=he5830b7_0
|
172 |
+
- libgpg-error=1.50=h4f305b6_0
|
173 |
+
- libhwloc=2.9.3=default_h554bfaf_1009
|
174 |
+
- libiconv=1.17=h166bdaf_0
|
175 |
+
- libidn2=2.3.7=hd590300_0
|
176 |
+
- libjpeg-turbo=3.0.0=hd590300_1
|
177 |
+
- liblapack=3.9.0=21_linux64_mkl
|
178 |
+
- libllvm15=15.0.7=h5cf9203_3
|
179 |
+
- libmamba=1.4.2=hcea66bb_0
|
180 |
+
- libmambapy=1.4.2=py310h1428755_0
|
181 |
+
- libnghttp2=1.58.0=h47da74e_1
|
182 |
+
- libnl=3.10.0=h4bc722e_0
|
183 |
+
- libnpp=12.0.2.50=0
|
184 |
+
- libnpp-dev=12.0.2.50=0
|
185 |
+
- libnpp-static=12.1.0.40=hd3aeb46_0
|
186 |
+
- libnsl=2.0.0=h7f98852_0
|
187 |
+
- libnvjitlink=12.1.105=0
|
188 |
+
- libnvjitlink-dev=12.1.105=0
|
189 |
+
- libnvjitlink-static=12.1.105=hd3aeb46_0
|
190 |
+
- libnvjpeg=12.1.1.14=0
|
191 |
+
- libnvjpeg-dev=12.1.1.14=0
|
192 |
+
- libnvjpeg-static=12.2.0.2=ha770c72_0
|
193 |
+
- libogg=1.3.5=h4ab18f5_0
|
194 |
+
- libopenblas=0.3.27=pthreads_hac2b453_1
|
195 |
+
- libopus=1.3.1=h7f98852_1
|
196 |
+
- libpng=1.6.43=h2797004_0
|
197 |
+
- libpq=16.4=h482b261_0
|
198 |
+
- libsndfile=1.2.2=hc60ed4a_1
|
199 |
+
- libsodium=1.0.18=h36c2ea0_1
|
200 |
+
- libsolv=0.7.29=ha6fb4c9_0
|
201 |
+
- libsqlite=3.46.0=hde9e2c9_0
|
202 |
+
- libssh2=1.11.0=h0841786_0
|
203 |
+
- libstdcxx-ng=13.1.0=hfd8a6a1_0
|
204 |
+
- libsystemd0=255=h3516f8a_1
|
205 |
+
- libtasn1=4.19.0=h166bdaf_0
|
206 |
+
- libtiff=4.6.0=h1dd3fc0_3
|
207 |
+
- libunistring=0.9.10=h7f98852_0
|
208 |
+
- libuuid=2.38.1=h0b41bf4_0
|
209 |
+
- libuv=1.48.0=hd590300_0
|
210 |
+
- libvorbis=1.3.7=h9c3ff4c_0
|
211 |
+
- libwebp-base=1.4.0=hd590300_0
|
212 |
+
- libxcb=1.15=h0b41bf4_0
|
213 |
+
- libxkbcommon=1.6.0=h5d7e998_0
|
214 |
+
- libxml2=2.11.6=h232c23b_0
|
215 |
+
- libzlib=1.2.13=h4ab18f5_6
|
216 |
+
- llvm-openmp=18.1.7=ha31de31_0
|
217 |
+
- lz4-c=1.9.4=hcb278e6_0
|
218 |
+
- lzo=2.10=h516909a_1000
|
219 |
+
- mamba=1.4.2=py310h51d5547_0
|
220 |
+
- marisa-trie=1.1.0=py310hc6cd4ac_1
|
221 |
+
- markdown-it-py=3.0.0=pyhd8ed1ab_0
|
222 |
+
- markupsafe=2.1.5=py310h2372a71_0
|
223 |
+
- matplotlib=3.9.1=py310hff52083_1
|
224 |
+
- matplotlib-base=3.9.1=py310hf02ac8c_2
|
225 |
+
- matplotlib-inline=0.1.7=pyhd8ed1ab_0
|
226 |
+
- mdurl=0.1.2=pyhd8ed1ab_0
|
227 |
+
- mkl=2024.0.0=ha957f24_49657
|
228 |
+
- mkl-include=2024.2.0=ha957f24_665
|
229 |
+
- mpc=1.3.1=hfe3b2da_0
|
230 |
+
- mpfr=4.2.1=h38ae2d0_2
|
231 |
+
- mpg123=1.32.6=h59595ed_0
|
232 |
+
- mpi=1.0=openmpi
|
233 |
+
- mpi4py=3.1.6=py310hb2ba3f8_1
|
234 |
+
- mpmath=1.3.0=pyhd8ed1ab_0
|
235 |
+
- munkres=1.1.4=pyh9f0ad1d_0
|
236 |
+
- murmurhash=1.0.10=py310hc6cd4ac_1
|
237 |
+
- mysql-common=8.0.33=hf1915f5_6
|
238 |
+
- mysql-libs=8.0.33=hca2cd23_6
|
239 |
+
- ncurses=6.5=h59595ed_0
|
240 |
+
- nest-asyncio=1.6.0=pyhd8ed1ab_0
|
241 |
+
- nettle=3.7.3=hbbd107a_1
|
242 |
+
- networkx=3.3=pyhd8ed1ab_1
|
243 |
+
- nspr=4.35=h27087fc_0
|
244 |
+
- nss=3.100=hca3bf56_0
|
245 |
+
- openh264=2.1.1=h780b84a_0
|
246 |
+
- openjpeg=2.5.2=h488ebb8_0
|
247 |
+
- openmpi=5.0.3=h817cd4e_101
|
248 |
+
- openssl=3.3.1=h4bc722e_2
|
249 |
+
- packaging=23.1=pyhd8ed1ab_0
|
250 |
+
- pandas=2.2.2=py310hf9f9076_1
|
251 |
+
- parso=0.8.4=pyhd8ed1ab_0
|
252 |
+
- pcre2=10.43=hcad00b1_0
|
253 |
+
- pexpect=4.9.0=pyhd8ed1ab_0
|
254 |
+
- pickleshare=0.7.5=py_1003
|
255 |
+
- pillow=10.3.0=py310hf73ecf8_0
|
256 |
+
- pixman=0.43.2=h59595ed_0
|
257 |
+
- platformdirs=4.2.2=pyhd8ed1ab_0
|
258 |
+
- pluggy=1.3.0=pyhd8ed1ab_0
|
259 |
+
- ply=3.11=pyhd8ed1ab_2
|
260 |
+
- preshed=3.0.9=py310hc6cd4ac_1
|
261 |
+
- prompt-toolkit=3.0.47=pyha770c72_0
|
262 |
+
- psutil=6.0.0=py310hc51659f_0
|
263 |
+
- pthread-stubs=0.4=h36c2ea0_1001
|
264 |
+
- ptyprocess=0.7.0=pyhd3deb0d_0
|
265 |
+
- pulseaudio-client=16.1=hb77b528_5
|
266 |
+
- pure_eval=0.2.3=pyhd8ed1ab_0
|
267 |
+
- pyasn1=0.6.0=pyhd8ed1ab_0
|
268 |
+
- pybind11=2.13.1=py310h25c7140_0
|
269 |
+
- pybind11-abi=4=hd8ed1ab_3
|
270 |
+
- pybind11-global=2.13.1=py310h25c7140_0
|
271 |
+
- pycosat=0.6.4=py310h5764c6d_1
|
272 |
+
- pycparser=2.21=pyhd8ed1ab_0
|
273 |
+
- pydantic=2.8.2=pyhd8ed1ab_0
|
274 |
+
- pydantic-core=2.20.1=py310h42e942d_0
|
275 |
+
- pygments=2.18.0=pyhd8ed1ab_0
|
276 |
+
- pyparsing=3.1.2=pyhd8ed1ab_0
|
277 |
+
- pyqt=5.15.9=py310h04931ad_5
|
278 |
+
- pyqt5-sip=12.12.2=py310hc6cd4ac_5
|
279 |
+
- pysocks=1.7.1=pyha2e5f31_6
|
280 |
+
- python=3.10.12=hd12c33a_0_cpython
|
281 |
+
- python-dateutil=2.9.0=pyhd8ed1ab_0
|
282 |
+
- python-tzdata=2024.1=pyhd8ed1ab_0
|
283 |
+
- python_abi=3.10=3_cp310
|
284 |
+
- pytorch-cuda=12.1=ha16c6d3_5
|
285 |
+
- pytorch-mutex=1.0=cuda
|
286 |
+
- pytz=2024.1=pyhd8ed1ab_0
|
287 |
+
- pyyaml=6.0.2=py310h5b4e0ec_0
|
288 |
+
- pyzmq=26.1.0=py310h7d2b5bf_0
|
289 |
+
- qhull=2020.2=h434a139_5
|
290 |
+
- qt-main=5.15.8=h82b777d_17
|
291 |
+
- readline=8.2=h8228510_1
|
292 |
+
- reproc=14.2.4=h0b41bf4_0
|
293 |
+
- reproc-cpp=14.2.4=hcb278e6_0
|
294 |
+
- requests=2.32.3=pyhd8ed1ab_0
|
295 |
+
- rhash=1.4.4=hd590300_0
|
296 |
+
- rich=13.7.1=pyhd8ed1ab_0
|
297 |
+
- rsa=4.7.2=pyh44b312d_0
|
298 |
+
- ruamel.yaml=0.17.32=py310h2372a71_0
|
299 |
+
- ruamel.yaml.clib=0.2.7=py310h1fa729e_1
|
300 |
+
- s3transfer=0.10.2=pyhd8ed1ab_0
|
301 |
+
- scipy=1.14.0=py310h93e2701_1
|
302 |
+
- shellingham=1.5.4=pyhd8ed1ab_0
|
303 |
+
- sip=6.7.12=py310hc6cd4ac_0
|
304 |
+
- six=1.16.0=pyh6c4a22f_0
|
305 |
+
- smart-open=7.0.4=hd8ed1ab_0
|
306 |
+
- smart_open=7.0.4=pyhd8ed1ab_0
|
307 |
+
- spacy=3.7.5=py310h9fd543b_1
|
308 |
+
- spacy-legacy=3.0.12=pyhd8ed1ab_0
|
309 |
+
- spacy-loggers=1.0.5=pyhd8ed1ab_0
|
310 |
+
- srsly=2.4.8=py310hc6cd4ac_1
|
311 |
+
- stack_data=0.6.2=pyhd8ed1ab_0
|
312 |
+
- sympy=1.13.0=pypyh2585a3b_103
|
313 |
+
- tbb=2021.8.0=hdb19cb5_0
|
314 |
+
- thinc=8.2.5=py310h9fd543b_0
|
315 |
+
- threadpoolctl=3.5.0=pyhc1e730c_0
|
316 |
+
- tk=8.6.13=noxft_h4845f30_101
|
317 |
+
- toml=0.10.2=pyhd8ed1ab_0
|
318 |
+
- tomli=2.0.1=pyhd8ed1ab_0
|
319 |
+
- toolz=0.12.0=pyhd8ed1ab_0
|
320 |
+
- torchdata=0.7.0=py310
|
321 |
+
- torchtext=0.16.0=py310
|
322 |
+
- tornado=6.4.1=py310hc51659f_0
|
323 |
+
- tqdm=4.66.5=pyhd8ed1ab_0
|
324 |
+
- traitlets=5.14.3=pyhd8ed1ab_0
|
325 |
+
- typer=0.12.3=pyhd8ed1ab_0
|
326 |
+
- typer-slim=0.12.3=pyhd8ed1ab_0
|
327 |
+
- typer-slim-standard=0.12.3=hd8ed1ab_0
|
328 |
+
- typing=3.10.0.0=pyhd8ed1ab_1
|
329 |
+
- typing-extensions=4.12.2=hd8ed1ab_0
|
330 |
+
- typing_extensions=4.12.2=pyha770c72_0
|
331 |
+
- tzdata=2023c=h71feb2d_0
|
332 |
+
- unicodedata2=15.1.0=py310h2372a71_0
|
333 |
+
- urllib3=1.26.19=pyhd8ed1ab_0
|
334 |
+
- wasabi=1.1.2=py310hff52083_1
|
335 |
+
- wcwidth=0.2.13=pyhd8ed1ab_0
|
336 |
+
- weasel=0.4.1=pyhd8ed1ab_1
|
337 |
+
- wheel=0.41.2=pyhd8ed1ab_0
|
338 |
+
- wrapt=1.16.0=py310h2372a71_0
|
339 |
+
- xcb-util=0.4.0=hd590300_1
|
340 |
+
- xcb-util-image=0.4.0=h8ee46fc_1
|
341 |
+
- xcb-util-keysyms=0.4.0=h8ee46fc_1
|
342 |
+
- xcb-util-renderutil=0.3.9=hd590300_1
|
343 |
+
- xcb-util-wm=0.4.1=h8ee46fc_1
|
344 |
+
- xkeyboard-config=2.42=h4ab18f5_0
|
345 |
+
- xorg-kbproto=1.0.7=h7f98852_1002
|
346 |
+
- xorg-libice=1.1.1=hd590300_0
|
347 |
+
- xorg-libsm=1.2.4=h7391055_0
|
348 |
+
- xorg-libx11=1.8.9=h8ee46fc_0
|
349 |
+
- xorg-libxau=1.0.11=hd590300_0
|
350 |
+
- xorg-libxdmcp=1.1.3=h7f98852_0
|
351 |
+
- xorg-libxext=1.3.4=h0b41bf4_2
|
352 |
+
- xorg-libxrender=0.9.11=hd590300_0
|
353 |
+
- xorg-renderproto=0.11.1=h7f98852_1002
|
354 |
+
- xorg-xextproto=7.3.0=h0b41bf4_1003
|
355 |
+
- xorg-xf86vidmodeproto=2.3.1=h7f98852_1002
|
356 |
+
- xorg-xproto=7.0.31=h7f98852_1007
|
357 |
+
- xz=5.2.6=h166bdaf_0
|
358 |
+
- yaml=0.2.5=h7f98852_2
|
359 |
+
- yaml-cpp=0.7.0=h27087fc_2
|
360 |
+
- zeromq=4.3.5=h75354e8_4
|
361 |
+
- zipp=3.19.2=pyhd8ed1ab_0
|
362 |
+
- zlib=1.2.13=h4ab18f5_6
|
363 |
+
- zstandard=0.19.0=py310h5764c6d_0
|
364 |
+
- zstd=1.5.6=ha6fb4c9_0
|
365 |
+
- pip:
|
366 |
+
- absl-py==2.1.0
|
367 |
+
- accelerate==1.1.1
|
368 |
+
- aiohappyeyeballs==2.4.3
|
369 |
+
- aiohttp==3.10.10
|
370 |
+
- aiosignal==1.3.1
|
371 |
+
- antlr4-python3-runtime==4.9.3
|
372 |
+
- apex==0.1
|
373 |
+
- appdirs==1.4.4
|
374 |
+
- argparse==1.4.0
|
375 |
+
- async-timeout==4.0.3
|
376 |
+
- attrdict==2.0.1
|
377 |
+
- attrs==23.2.0
|
378 |
+
- awscli==1.33.40
|
379 |
+
- backcall==0.2.0
|
380 |
+
- bcrypt==4.2.0
|
381 |
+
- beautifulsoup4==4.12.3
|
382 |
+
- bleach==6.2.0
|
383 |
+
- bokeh==3.5.1
|
384 |
+
- boto3==1.34.158
|
385 |
+
- botocore==1.34.158
|
386 |
+
- brotli==1.0.9
|
387 |
+
- cachetools==5.5.0
|
388 |
+
- chardet==5.2.0
|
389 |
+
- cloudpickle==2.2.1
|
390 |
+
- cmake==3.30.2
|
391 |
+
- cryptography==41.0.7
|
392 |
+
- datasets==3.1.0
|
393 |
+
- defusedxml==0.7.1
|
394 |
+
- dill==0.3.8
|
395 |
+
- docker==7.1.0
|
396 |
+
- docker-pycreds==0.4.0
|
397 |
+
- docopt==0.6.2
|
398 |
+
- docstring-parser==0.16
|
399 |
+
- einops==0.8.0
|
400 |
+
- fastjsonschema==2.20.0
|
401 |
+
- fire==0.6.0
|
402 |
+
- flash-attn==2.5.5
|
403 |
+
- frozenlist==1.5.0
|
404 |
+
- fsspec==2024.6.1
|
405 |
+
- gevent==24.2.1
|
406 |
+
- gitdb==4.0.11
|
407 |
+
- gitpython==3.1.43
|
408 |
+
- google-pasta==0.2.0
|
409 |
+
- greenlet==3.0.3
|
410 |
+
- huggingface-hub==0.26.2
|
411 |
+
- hydra-core==1.3.2
|
412 |
+
- imageio==2.34.2
|
413 |
+
- importlib-metadata==6.11.0
|
414 |
+
- importlib-resources==6.1.1
|
415 |
+
- inotify-simple==1.2.1
|
416 |
+
- ipykernel==6.29.1
|
417 |
+
- ipython==8.12.3
|
418 |
+
- jsonschema==4.23.0
|
419 |
+
- jsonschema-specifications==2023.12.1
|
420 |
+
- jupyterlab-pygments==0.3.0
|
421 |
+
- llvmlite==0.43.0
|
422 |
+
- mistune==3.0.2
|
423 |
+
- multidict==6.1.0
|
424 |
+
- multiprocess==0.70.16
|
425 |
+
- nbclient==0.10.0
|
426 |
+
- nbconvert==7.16.4
|
427 |
+
- nbformat==5.10.4
|
428 |
+
- ninja==1.11.1.1
|
429 |
+
- numba==0.60.0
|
430 |
+
- numpy==1.24.3
|
431 |
+
- nvidia-cublas-cu12==12.1.3.1
|
432 |
+
- nvidia-cuda-cupti-cu12==12.1.105
|
433 |
+
- nvidia-cuda-nvrtc-cu12==12.1.105
|
434 |
+
- nvidia-cuda-runtime-cu12==12.1.105
|
435 |
+
- nvidia-cudnn-cu12==8.9.2.26
|
436 |
+
- nvidia-cufft-cu12==11.0.2.54
|
437 |
+
- nvidia-curand-cu12==10.3.2.106
|
438 |
+
- nvidia-cusolver-cu12==11.4.5.107
|
439 |
+
- nvidia-cusparse-cu12==12.1.0.106
|
440 |
+
- nvidia-ml-py==12.535.161
|
441 |
+
- nvidia-nccl-cu12==2.19.3
|
442 |
+
- nvidia-nvjitlink-cu12==12.6.77
|
443 |
+
- nvidia-nvtx-cu12==12.1.105
|
444 |
+
- nvitop==1.3.2
|
445 |
+
- omegaconf==2.3.0
|
446 |
+
- opencv-python==4.10.0.84
|
447 |
+
- pandocfilters==1.5.1
|
448 |
+
- paramiko==3.4.0
|
449 |
+
- pathos==0.3.2
|
450 |
+
- pip==24.1.2
|
451 |
+
- pip-chill==1.0.3
|
452 |
+
- pipreqs==0.5.0
|
453 |
+
- plotly==5.23.0
|
454 |
+
- pox==0.3.4
|
455 |
+
- ppft==1.7.6.8
|
456 |
+
- propcache==0.2.0
|
457 |
+
- protobuf==3.20.3
|
458 |
+
- pyarrow==17.0.0
|
459 |
+
- pyfunctional==1.5.0
|
460 |
+
- pynacl==1.5.0
|
461 |
+
- pyopenssl==23.2.0
|
462 |
+
- referencing==0.35.1
|
463 |
+
- regex==2024.11.6
|
464 |
+
- retrying==1.3.4
|
465 |
+
- rpds-py==0.20.0
|
466 |
+
- s3fs==0.4.2
|
467 |
+
- safetensors==0.4.5
|
468 |
+
- sagemaker==2.224.2
|
469 |
+
- sagemaker-experiments==0.1.45
|
470 |
+
- sagemaker-pytorch-training==2.8.1
|
471 |
+
- sagemaker-training==4.8.1
|
472 |
+
- schema==0.7.7
|
473 |
+
- scikit-learn==1.3.2
|
474 |
+
- seaborn==0.13.2
|
475 |
+
- sentencepiece==0.2.0
|
476 |
+
- sentry-sdk==2.18.0
|
477 |
+
- setproctitle==1.3.3
|
478 |
+
- setuptools==72.1.0
|
479 |
+
- shap==0.46.0
|
480 |
+
- shtab==1.7.1
|
481 |
+
- slicer==0.0.8
|
482 |
+
- smclarify==0.5
|
483 |
+
- smdebug-rulesconfig==1.0.1
|
484 |
+
- smdistributed-dataparallel==2.1.0
|
485 |
+
- smmap==5.0.1
|
486 |
+
- smprof==0.3.334
|
487 |
+
- soupsieve==2.6
|
488 |
+
- tabulate==0.9.0
|
489 |
+
- tblib==3.0.0
|
490 |
+
- tenacity==9.0.0
|
491 |
+
- tensor-parallel==2.0.0
|
492 |
+
- termcolor==2.5.0
|
493 |
+
- tinycss2==1.4.0
|
494 |
+
- tokenizers==0.20.3
|
495 |
+
- torch==2.2.0
|
496 |
+
- torchaudio==2.2.0
|
497 |
+
- torchnet==0.0.4
|
498 |
+
- torchvision==0.17.0
|
499 |
+
- transformer-engine==0.12.0+170797
|
500 |
+
- transformers==4.46.0
|
501 |
+
- triton==2.2.0
|
502 |
+
- trl==0.12.0
|
503 |
+
- tyro==0.8.5
|
504 |
+
- visdom==0.2.4
|
505 |
+
- wandb==0.16.3
|
506 |
+
- webencodings==0.5.1
|
507 |
+
- websocket-client==1.8.0
|
508 |
+
- werkzeug==3.0.3
|
509 |
+
- xxhash==3.5.0
|
510 |
+
- xyzservices==2024.6.0
|
511 |
+
- yarg==0.1.9
|
512 |
+
- yarl==1.17.1
|
513 |
+
- zope-event==5.0
|
514 |
+
- zope-interface==7.0.1
|
515 |
+
prefix: /opt/conda
|
wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/config.yaml
ADDED
@@ -0,0 +1,948 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
wandb_version: 1
|
2 |
+
|
3 |
+
sagemaker_training_job_name:
|
4 |
+
desc: null
|
5 |
+
value: kushalarora-rvv-main-2024-11-13-17-55-42-664
|
6 |
+
batch_size:
|
7 |
+
desc: null
|
8 |
+
value: 64
|
9 |
+
cache_dir:
|
10 |
+
desc: null
|
11 |
+
value: /opt/ml/data/input/.cache
|
12 |
+
dataset:
|
13 |
+
desc: null
|
14 |
+
value: code-contests
|
15 |
+
gradient_accumulation_steps:
|
16 |
+
desc: null
|
17 |
+
value: 8
|
18 |
+
logging_steps:
|
19 |
+
desc: null
|
20 |
+
value: 5
|
21 |
+
lr:
|
22 |
+
desc: null
|
23 |
+
value: 1e-6
|
24 |
+
max_seq_length:
|
25 |
+
desc: null
|
26 |
+
value: 2048
|
27 |
+
model_name:
|
28 |
+
desc: null
|
29 |
+
value: Qwen/Qwen2.5-Coder-7B-Instruct
|
30 |
+
num_train_epochs:
|
31 |
+
desc: null
|
32 |
+
value: 2
|
33 |
+
output_dir:
|
34 |
+
desc: null
|
35 |
+
value: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048
|
36 |
+
push_to_hub:
|
37 |
+
desc: null
|
38 |
+
value: true
|
39 |
+
run_name:
|
40 |
+
desc: null
|
41 |
+
value: sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048
|
42 |
+
sagemaker_container_log_level:
|
43 |
+
desc: null
|
44 |
+
value: 20
|
45 |
+
sagemaker_instance_type:
|
46 |
+
desc: null
|
47 |
+
value: ml.p5.48xlarge
|
48 |
+
sagemaker_job_name:
|
49 |
+
desc: null
|
50 |
+
value: kushalarora-rvv-main-2024-11-13-17-55-42-664
|
51 |
+
sagemaker_program:
|
52 |
+
desc: null
|
53 |
+
value: sft.py
|
54 |
+
sagemaker_region:
|
55 |
+
desc: null
|
56 |
+
value: us-east-1
|
57 |
+
sagemaker_submit_directory:
|
58 |
+
desc: null
|
59 |
+
value: s3://tri-ml-datasets/kushalarora-rvv-main-2024-11-13-17-55-42-664/source/sourcedir.tar.gz
|
60 |
+
sagemaker_torch_distributed_enabled:
|
61 |
+
desc: null
|
62 |
+
value: 'true'
|
63 |
+
save_steps:
|
64 |
+
desc: null
|
65 |
+
value: 100
|
66 |
+
schedule:
|
67 |
+
desc: null
|
68 |
+
value: cosine
|
69 |
+
wandb_project:
|
70 |
+
desc: null
|
71 |
+
value: sft-codecontests-1112
|
72 |
+
weight_decay:
|
73 |
+
desc: null
|
74 |
+
value: 0.0
|
75 |
+
additional_framework_parameters:
|
76 |
+
desc: null
|
77 |
+
value:
|
78 |
+
sagemaker_instance_type: ml.p5.48xlarge
|
79 |
+
sagemaker_torch_distributed_enabled: true
|
80 |
+
channel_input_dirs:
|
81 |
+
desc: null
|
82 |
+
value: {}
|
83 |
+
current_host:
|
84 |
+
desc: null
|
85 |
+
value: algo-1
|
86 |
+
current_instance_group:
|
87 |
+
desc: null
|
88 |
+
value: homogeneousCluster
|
89 |
+
current_instance_group_hosts:
|
90 |
+
desc: null
|
91 |
+
value:
|
92 |
+
- algo-1
|
93 |
+
current_instance_type:
|
94 |
+
desc: null
|
95 |
+
value: ml.p5.48xlarge
|
96 |
+
distribution_hosts:
|
97 |
+
desc: null
|
98 |
+
value:
|
99 |
+
- algo-1
|
100 |
+
distribution_instance_groups:
|
101 |
+
desc: null
|
102 |
+
value:
|
103 |
+
- homogeneousCluster
|
104 |
+
framework_module:
|
105 |
+
desc: null
|
106 |
+
value: sagemaker_pytorch_container.training:main
|
107 |
+
hosts:
|
108 |
+
desc: null
|
109 |
+
value:
|
110 |
+
- algo-1
|
111 |
+
hyperparameters:
|
112 |
+
desc: null
|
113 |
+
value:
|
114 |
+
batch_size: 64
|
115 |
+
cache_dir: /opt/ml/data/input/.cache
|
116 |
+
dataset: code-contests
|
117 |
+
gradient_accumulation_steps: 8
|
118 |
+
logging_steps: 5
|
119 |
+
lr: 1e-6
|
120 |
+
max_seq_length: 2048
|
121 |
+
model_name: Qwen/Qwen2.5-Coder-7B-Instruct
|
122 |
+
num_train_epochs: 2
|
123 |
+
output_dir: /opt/ml/model/
|
124 |
+
push_to_hub: true
|
125 |
+
run_name: sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048
|
126 |
+
save_steps: 100
|
127 |
+
schedule: cosine
|
128 |
+
wandb_project: sft-codecontests-1112
|
129 |
+
weight_decay: 0.0
|
130 |
+
input_config_dir:
|
131 |
+
desc: null
|
132 |
+
value: /opt/ml/input/config
|
133 |
+
input_data_config:
|
134 |
+
desc: null
|
135 |
+
value: {}
|
136 |
+
input_dir:
|
137 |
+
desc: null
|
138 |
+
value: /opt/ml/input
|
139 |
+
instance_groups:
|
140 |
+
desc: null
|
141 |
+
value:
|
142 |
+
- homogeneousCluster
|
143 |
+
instance_groups_dict:
|
144 |
+
desc: null
|
145 |
+
value:
|
146 |
+
homogeneousCluster:
|
147 |
+
hosts:
|
148 |
+
- algo-1
|
149 |
+
instance_group_name: homogeneousCluster
|
150 |
+
instance_type: ml.p5.48xlarge
|
151 |
+
is_hetero:
|
152 |
+
desc: null
|
153 |
+
value: false
|
154 |
+
is_master:
|
155 |
+
desc: null
|
156 |
+
value: true
|
157 |
+
is_modelparallel_enabled:
|
158 |
+
desc: null
|
159 |
+
value: null
|
160 |
+
is_smddpmprun_installed:
|
161 |
+
desc: null
|
162 |
+
value: false
|
163 |
+
is_smddprun_installed:
|
164 |
+
desc: null
|
165 |
+
value: true
|
166 |
+
job_name:
|
167 |
+
desc: null
|
168 |
+
value: kushalarora-rvv-main-2024-11-13-17-55-42-664
|
169 |
+
log_level:
|
170 |
+
desc: null
|
171 |
+
value: passive
|
172 |
+
master_hostname:
|
173 |
+
desc: null
|
174 |
+
value: algo-1
|
175 |
+
model_dir:
|
176 |
+
desc: null
|
177 |
+
value: /opt/ml/model
|
178 |
+
module_dir:
|
179 |
+
desc: null
|
180 |
+
value: s3://tri-ml-datasets/kushalarora-rvv-main-2024-11-13-17-55-42-664/source/sourcedir.tar.gz
|
181 |
+
module_name:
|
182 |
+
desc: null
|
183 |
+
value: /opt/ml/code/qlearning_reasoning/training/sft
|
184 |
+
network_interface_name:
|
185 |
+
desc: null
|
186 |
+
value: eth0
|
187 |
+
num_cpus:
|
188 |
+
desc: null
|
189 |
+
value: 192
|
190 |
+
num_gpus:
|
191 |
+
desc: null
|
192 |
+
value: 8
|
193 |
+
num_neurons:
|
194 |
+
desc: null
|
195 |
+
value: 0
|
196 |
+
output_data_dir:
|
197 |
+
desc: null
|
198 |
+
value: /opt/ml/output/data
|
199 |
+
output_intermediate_dir:
|
200 |
+
desc: null
|
201 |
+
value: /opt/ml/output/intermediate
|
202 |
+
resource_config:
|
203 |
+
desc: null
|
204 |
+
value:
|
205 |
+
current_group_name: homogeneousCluster
|
206 |
+
current_host: algo-1
|
207 |
+
current_instance_type: ml.p5.48xlarge
|
208 |
+
hosts:
|
209 |
+
- algo-1
|
210 |
+
instance_groups:
|
211 |
+
- hosts:
|
212 |
+
- algo-1
|
213 |
+
instance_group_name: homogeneousCluster
|
214 |
+
instance_type: ml.p5.48xlarge
|
215 |
+
network_interface_name: eth0
|
216 |
+
user_entry_point:
|
217 |
+
desc: null
|
218 |
+
value: /opt/ml/code/qlearning_reasoning/training/sft.py
|
219 |
+
_wandb:
|
220 |
+
desc: null
|
221 |
+
value:
|
222 |
+
code_path: code/qlearning_reasoning/training/sft.py
|
223 |
+
python_version: 3.10.12
|
224 |
+
cli_version: 0.16.3
|
225 |
+
framework: huggingface
|
226 |
+
huggingface_version: 4.46.0
|
227 |
+
is_jupyter_run: false
|
228 |
+
is_kaggle_kernel: false
|
229 |
+
start_time: 1731524566.431221
|
230 |
+
t:
|
231 |
+
1:
|
232 |
+
- 1
|
233 |
+
- 5
|
234 |
+
- 11
|
235 |
+
- 41
|
236 |
+
- 49
|
237 |
+
- 51
|
238 |
+
- 53
|
239 |
+
- 55
|
240 |
+
- 71
|
241 |
+
- 84
|
242 |
+
2:
|
243 |
+
- 1
|
244 |
+
- 5
|
245 |
+
- 11
|
246 |
+
- 41
|
247 |
+
- 49
|
248 |
+
- 51
|
249 |
+
- 53
|
250 |
+
- 55
|
251 |
+
- 71
|
252 |
+
- 84
|
253 |
+
3:
|
254 |
+
- 7
|
255 |
+
- 9
|
256 |
+
- 13
|
257 |
+
- 19
|
258 |
+
- 23
|
259 |
+
4: 3.10.12
|
260 |
+
5: 0.16.3
|
261 |
+
6: 4.46.0
|
262 |
+
8:
|
263 |
+
- 5
|
264 |
+
9:
|
265 |
+
1: transformers_trainer
|
266 |
+
13: linux-x86_64
|
267 |
+
m:
|
268 |
+
- 1: train/global_step
|
269 |
+
6:
|
270 |
+
- 3
|
271 |
+
- 1: train/loss
|
272 |
+
5: 1
|
273 |
+
6:
|
274 |
+
- 1
|
275 |
+
- 1: train/grad_norm
|
276 |
+
5: 1
|
277 |
+
6:
|
278 |
+
- 1
|
279 |
+
- 1: train/learning_rate
|
280 |
+
5: 1
|
281 |
+
6:
|
282 |
+
- 1
|
283 |
+
- 1: train/epoch
|
284 |
+
5: 1
|
285 |
+
6:
|
286 |
+
- 1
|
287 |
+
- 1: eval/loss
|
288 |
+
5: 1
|
289 |
+
6:
|
290 |
+
- 1
|
291 |
+
- 1: eval/runtime
|
292 |
+
5: 1
|
293 |
+
6:
|
294 |
+
- 1
|
295 |
+
- 1: eval/samples_per_second
|
296 |
+
5: 1
|
297 |
+
6:
|
298 |
+
- 1
|
299 |
+
- 1: eval/steps_per_second
|
300 |
+
5: 1
|
301 |
+
6:
|
302 |
+
- 1
|
303 |
+
vocab_size:
|
304 |
+
desc: null
|
305 |
+
value: 151671
|
306 |
+
max_position_embeddings:
|
307 |
+
desc: null
|
308 |
+
value: 32768
|
309 |
+
hidden_size:
|
310 |
+
desc: null
|
311 |
+
value: 3584
|
312 |
+
intermediate_size:
|
313 |
+
desc: null
|
314 |
+
value: 18944
|
315 |
+
num_hidden_layers:
|
316 |
+
desc: null
|
317 |
+
value: 28
|
318 |
+
num_attention_heads:
|
319 |
+
desc: null
|
320 |
+
value: 28
|
321 |
+
use_sliding_window:
|
322 |
+
desc: null
|
323 |
+
value: false
|
324 |
+
sliding_window:
|
325 |
+
desc: null
|
326 |
+
value: null
|
327 |
+
max_window_layers:
|
328 |
+
desc: null
|
329 |
+
value: 28
|
330 |
+
num_key_value_heads:
|
331 |
+
desc: null
|
332 |
+
value: 4
|
333 |
+
hidden_act:
|
334 |
+
desc: null
|
335 |
+
value: silu
|
336 |
+
initializer_range:
|
337 |
+
desc: null
|
338 |
+
value: 0.02
|
339 |
+
rms_norm_eps:
|
340 |
+
desc: null
|
341 |
+
value: 1.0e-06
|
342 |
+
use_cache:
|
343 |
+
desc: null
|
344 |
+
value: false
|
345 |
+
rope_theta:
|
346 |
+
desc: null
|
347 |
+
value: 1000000.0
|
348 |
+
rope_scaling:
|
349 |
+
desc: null
|
350 |
+
value: null
|
351 |
+
attention_dropout:
|
352 |
+
desc: null
|
353 |
+
value: 0.0
|
354 |
+
return_dict:
|
355 |
+
desc: null
|
356 |
+
value: true
|
357 |
+
output_hidden_states:
|
358 |
+
desc: null
|
359 |
+
value: false
|
360 |
+
output_attentions:
|
361 |
+
desc: null
|
362 |
+
value: false
|
363 |
+
torchscript:
|
364 |
+
desc: null
|
365 |
+
value: false
|
366 |
+
torch_dtype:
|
367 |
+
desc: null
|
368 |
+
value: bfloat16
|
369 |
+
use_bfloat16:
|
370 |
+
desc: null
|
371 |
+
value: false
|
372 |
+
tf_legacy_loss:
|
373 |
+
desc: null
|
374 |
+
value: false
|
375 |
+
pruned_heads:
|
376 |
+
desc: null
|
377 |
+
value: {}
|
378 |
+
tie_word_embeddings:
|
379 |
+
desc: null
|
380 |
+
value: false
|
381 |
+
chunk_size_feed_forward:
|
382 |
+
desc: null
|
383 |
+
value: 0
|
384 |
+
is_encoder_decoder:
|
385 |
+
desc: null
|
386 |
+
value: false
|
387 |
+
is_decoder:
|
388 |
+
desc: null
|
389 |
+
value: false
|
390 |
+
cross_attention_hidden_size:
|
391 |
+
desc: null
|
392 |
+
value: null
|
393 |
+
add_cross_attention:
|
394 |
+
desc: null
|
395 |
+
value: false
|
396 |
+
tie_encoder_decoder:
|
397 |
+
desc: null
|
398 |
+
value: false
|
399 |
+
max_length:
|
400 |
+
desc: null
|
401 |
+
value: 20
|
402 |
+
min_length:
|
403 |
+
desc: null
|
404 |
+
value: 0
|
405 |
+
do_sample:
|
406 |
+
desc: null
|
407 |
+
value: false
|
408 |
+
early_stopping:
|
409 |
+
desc: null
|
410 |
+
value: false
|
411 |
+
num_beams:
|
412 |
+
desc: null
|
413 |
+
value: 1
|
414 |
+
num_beam_groups:
|
415 |
+
desc: null
|
416 |
+
value: 1
|
417 |
+
diversity_penalty:
|
418 |
+
desc: null
|
419 |
+
value: 0.0
|
420 |
+
temperature:
|
421 |
+
desc: null
|
422 |
+
value: 1.0
|
423 |
+
top_k:
|
424 |
+
desc: null
|
425 |
+
value: 50
|
426 |
+
top_p:
|
427 |
+
desc: null
|
428 |
+
value: 1.0
|
429 |
+
typical_p:
|
430 |
+
desc: null
|
431 |
+
value: 1.0
|
432 |
+
repetition_penalty:
|
433 |
+
desc: null
|
434 |
+
value: 1.0
|
435 |
+
length_penalty:
|
436 |
+
desc: null
|
437 |
+
value: 1.0
|
438 |
+
no_repeat_ngram_size:
|
439 |
+
desc: null
|
440 |
+
value: 0
|
441 |
+
encoder_no_repeat_ngram_size:
|
442 |
+
desc: null
|
443 |
+
value: 0
|
444 |
+
bad_words_ids:
|
445 |
+
desc: null
|
446 |
+
value: null
|
447 |
+
num_return_sequences:
|
448 |
+
desc: null
|
449 |
+
value: 1
|
450 |
+
output_scores:
|
451 |
+
desc: null
|
452 |
+
value: false
|
453 |
+
return_dict_in_generate:
|
454 |
+
desc: null
|
455 |
+
value: false
|
456 |
+
forced_bos_token_id:
|
457 |
+
desc: null
|
458 |
+
value: null
|
459 |
+
forced_eos_token_id:
|
460 |
+
desc: null
|
461 |
+
value: null
|
462 |
+
remove_invalid_values:
|
463 |
+
desc: null
|
464 |
+
value: false
|
465 |
+
exponential_decay_length_penalty:
|
466 |
+
desc: null
|
467 |
+
value: null
|
468 |
+
suppress_tokens:
|
469 |
+
desc: null
|
470 |
+
value: null
|
471 |
+
begin_suppress_tokens:
|
472 |
+
desc: null
|
473 |
+
value: null
|
474 |
+
architectures:
|
475 |
+
desc: null
|
476 |
+
value:
|
477 |
+
- Qwen2ForCausalLM
|
478 |
+
finetuning_task:
|
479 |
+
desc: null
|
480 |
+
value: null
|
481 |
+
id2label:
|
482 |
+
desc: null
|
483 |
+
value:
|
484 |
+
'0': LABEL_0
|
485 |
+
'1': LABEL_1
|
486 |
+
label2id:
|
487 |
+
desc: null
|
488 |
+
value:
|
489 |
+
LABEL_0: 0
|
490 |
+
LABEL_1: 1
|
491 |
+
tokenizer_class:
|
492 |
+
desc: null
|
493 |
+
value: null
|
494 |
+
prefix:
|
495 |
+
desc: null
|
496 |
+
value: null
|
497 |
+
bos_token_id:
|
498 |
+
desc: null
|
499 |
+
value: null
|
500 |
+
pad_token_id:
|
501 |
+
desc: null
|
502 |
+
value: 151643
|
503 |
+
eos_token_id:
|
504 |
+
desc: null
|
505 |
+
value: 151645
|
506 |
+
sep_token_id:
|
507 |
+
desc: null
|
508 |
+
value: null
|
509 |
+
decoder_start_token_id:
|
510 |
+
desc: null
|
511 |
+
value: null
|
512 |
+
task_specific_params:
|
513 |
+
desc: null
|
514 |
+
value: null
|
515 |
+
problem_type:
|
516 |
+
desc: null
|
517 |
+
value: null
|
518 |
+
_name_or_path:
|
519 |
+
desc: null
|
520 |
+
value: Qwen/Qwen2.5-Coder-7B-Instruct
|
521 |
+
_attn_implementation_autoset:
|
522 |
+
desc: null
|
523 |
+
value: true
|
524 |
+
transformers_version:
|
525 |
+
desc: null
|
526 |
+
value: 4.46.0
|
527 |
+
model_type:
|
528 |
+
desc: null
|
529 |
+
value: qwen2
|
530 |
+
overwrite_output_dir:
|
531 |
+
desc: null
|
532 |
+
value: false
|
533 |
+
do_train:
|
534 |
+
desc: null
|
535 |
+
value: false
|
536 |
+
do_eval:
|
537 |
+
desc: null
|
538 |
+
value: true
|
539 |
+
do_predict:
|
540 |
+
desc: null
|
541 |
+
value: false
|
542 |
+
eval_strategy:
|
543 |
+
desc: null
|
544 |
+
value: steps
|
545 |
+
prediction_loss_only:
|
546 |
+
desc: null
|
547 |
+
value: false
|
548 |
+
per_device_train_batch_size:
|
549 |
+
desc: null
|
550 |
+
value: 8
|
551 |
+
per_device_eval_batch_size:
|
552 |
+
desc: null
|
553 |
+
value: 8
|
554 |
+
per_gpu_train_batch_size:
|
555 |
+
desc: null
|
556 |
+
value: null
|
557 |
+
per_gpu_eval_batch_size:
|
558 |
+
desc: null
|
559 |
+
value: null
|
560 |
+
eval_accumulation_steps:
|
561 |
+
desc: null
|
562 |
+
value: null
|
563 |
+
eval_delay:
|
564 |
+
desc: null
|
565 |
+
value: 0
|
566 |
+
torch_empty_cache_steps:
|
567 |
+
desc: null
|
568 |
+
value: null
|
569 |
+
learning_rate:
|
570 |
+
desc: null
|
571 |
+
value: 1.0e-06
|
572 |
+
adam_beta1:
|
573 |
+
desc: null
|
574 |
+
value: 0.9
|
575 |
+
adam_beta2:
|
576 |
+
desc: null
|
577 |
+
value: 0.999
|
578 |
+
adam_epsilon:
|
579 |
+
desc: null
|
580 |
+
value: 1.0e-08
|
581 |
+
max_grad_norm:
|
582 |
+
desc: null
|
583 |
+
value: 1.0
|
584 |
+
max_steps:
|
585 |
+
desc: null
|
586 |
+
value: -1
|
587 |
+
lr_scheduler_type:
|
588 |
+
desc: null
|
589 |
+
value: cosine
|
590 |
+
lr_scheduler_kwargs:
|
591 |
+
desc: null
|
592 |
+
value: {}
|
593 |
+
warmup_ratio:
|
594 |
+
desc: null
|
595 |
+
value: 0.0
|
596 |
+
warmup_steps:
|
597 |
+
desc: null
|
598 |
+
value: 0
|
599 |
+
log_level_replica:
|
600 |
+
desc: null
|
601 |
+
value: warning
|
602 |
+
log_on_each_node:
|
603 |
+
desc: null
|
604 |
+
value: true
|
605 |
+
logging_dir:
|
606 |
+
desc: null
|
607 |
+
value: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/runs/Nov13_19-02-15_algo-1
|
608 |
+
logging_strategy:
|
609 |
+
desc: null
|
610 |
+
value: steps
|
611 |
+
logging_first_step:
|
612 |
+
desc: null
|
613 |
+
value: true
|
614 |
+
logging_nan_inf_filter:
|
615 |
+
desc: null
|
616 |
+
value: true
|
617 |
+
save_strategy:
|
618 |
+
desc: null
|
619 |
+
value: epoch
|
620 |
+
save_total_limit:
|
621 |
+
desc: null
|
622 |
+
value: null
|
623 |
+
save_safetensors:
|
624 |
+
desc: null
|
625 |
+
value: true
|
626 |
+
save_on_each_node:
|
627 |
+
desc: null
|
628 |
+
value: false
|
629 |
+
save_only_model:
|
630 |
+
desc: null
|
631 |
+
value: false
|
632 |
+
restore_callback_states_from_checkpoint:
|
633 |
+
desc: null
|
634 |
+
value: false
|
635 |
+
no_cuda:
|
636 |
+
desc: null
|
637 |
+
value: false
|
638 |
+
use_cpu:
|
639 |
+
desc: null
|
640 |
+
value: false
|
641 |
+
use_mps_device:
|
642 |
+
desc: null
|
643 |
+
value: false
|
644 |
+
seed:
|
645 |
+
desc: null
|
646 |
+
value: 42
|
647 |
+
data_seed:
|
648 |
+
desc: null
|
649 |
+
value: null
|
650 |
+
jit_mode_eval:
|
651 |
+
desc: null
|
652 |
+
value: false
|
653 |
+
use_ipex:
|
654 |
+
desc: null
|
655 |
+
value: false
|
656 |
+
bf16:
|
657 |
+
desc: null
|
658 |
+
value: false
|
659 |
+
fp16:
|
660 |
+
desc: null
|
661 |
+
value: false
|
662 |
+
fp16_opt_level:
|
663 |
+
desc: null
|
664 |
+
value: O1
|
665 |
+
half_precision_backend:
|
666 |
+
desc: null
|
667 |
+
value: auto
|
668 |
+
bf16_full_eval:
|
669 |
+
desc: null
|
670 |
+
value: false
|
671 |
+
fp16_full_eval:
|
672 |
+
desc: null
|
673 |
+
value: false
|
674 |
+
tf32:
|
675 |
+
desc: null
|
676 |
+
value: null
|
677 |
+
local_rank:
|
678 |
+
desc: null
|
679 |
+
value: 0
|
680 |
+
ddp_backend:
|
681 |
+
desc: null
|
682 |
+
value: null
|
683 |
+
tpu_num_cores:
|
684 |
+
desc: null
|
685 |
+
value: null
|
686 |
+
tpu_metrics_debug:
|
687 |
+
desc: null
|
688 |
+
value: false
|
689 |
+
debug:
|
690 |
+
desc: null
|
691 |
+
value: []
|
692 |
+
dataloader_drop_last:
|
693 |
+
desc: null
|
694 |
+
value: false
|
695 |
+
eval_steps:
|
696 |
+
desc: null
|
697 |
+
value: 5
|
698 |
+
dataloader_num_workers:
|
699 |
+
desc: null
|
700 |
+
value: 0
|
701 |
+
dataloader_prefetch_factor:
|
702 |
+
desc: null
|
703 |
+
value: null
|
704 |
+
past_index:
|
705 |
+
desc: null
|
706 |
+
value: -1
|
707 |
+
disable_tqdm:
|
708 |
+
desc: null
|
709 |
+
value: false
|
710 |
+
remove_unused_columns:
|
711 |
+
desc: null
|
712 |
+
value: true
|
713 |
+
label_names:
|
714 |
+
desc: null
|
715 |
+
value: null
|
716 |
+
load_best_model_at_end:
|
717 |
+
desc: null
|
718 |
+
value: false
|
719 |
+
metric_for_best_model:
|
720 |
+
desc: null
|
721 |
+
value: null
|
722 |
+
greater_is_better:
|
723 |
+
desc: null
|
724 |
+
value: null
|
725 |
+
ignore_data_skip:
|
726 |
+
desc: null
|
727 |
+
value: false
|
728 |
+
fsdp:
|
729 |
+
desc: null
|
730 |
+
value:
|
731 |
+
- full_shard
|
732 |
+
- auto_wrap
|
733 |
+
fsdp_min_num_params:
|
734 |
+
desc: null
|
735 |
+
value: 0
|
736 |
+
fsdp_config:
|
737 |
+
desc: null
|
738 |
+
value:
|
739 |
+
activation_checkpointing: false
|
740 |
+
auto_wrap_policy: TRANSFORMER_BASED_WRAP
|
741 |
+
backward_prefetch: BACKWARD_PRE
|
742 |
+
cpu_ram_efficient_loading: true
|
743 |
+
forward_prefetch: false
|
744 |
+
offload_params: false
|
745 |
+
sharding_strategy: FULL_SHARD
|
746 |
+
state_dict_type: FULL_STATE_DICT
|
747 |
+
sync_module_states: true
|
748 |
+
min_num_params: 0
|
749 |
+
xla: false
|
750 |
+
xla_fsdp_v2: false
|
751 |
+
xla_fsdp_grad_ckpt: false
|
752 |
+
fsdp_transformer_layer_cls_to_wrap:
|
753 |
+
desc: null
|
754 |
+
value: null
|
755 |
+
accelerator_config:
|
756 |
+
desc: null
|
757 |
+
value:
|
758 |
+
split_batches: false
|
759 |
+
dispatch_batches: null
|
760 |
+
even_batches: true
|
761 |
+
use_seedable_sampler: true
|
762 |
+
non_blocking: false
|
763 |
+
gradient_accumulation_kwargs: null
|
764 |
+
deepspeed:
|
765 |
+
desc: null
|
766 |
+
value: null
|
767 |
+
label_smoothing_factor:
|
768 |
+
desc: null
|
769 |
+
value: 0.0
|
770 |
+
optim:
|
771 |
+
desc: null
|
772 |
+
value: adamw_torch
|
773 |
+
optim_args:
|
774 |
+
desc: null
|
775 |
+
value: null
|
776 |
+
adafactor:
|
777 |
+
desc: null
|
778 |
+
value: false
|
779 |
+
group_by_length:
|
780 |
+
desc: null
|
781 |
+
value: false
|
782 |
+
length_column_name:
|
783 |
+
desc: null
|
784 |
+
value: length
|
785 |
+
report_to:
|
786 |
+
desc: null
|
787 |
+
value:
|
788 |
+
- wandb
|
789 |
+
ddp_find_unused_parameters:
|
790 |
+
desc: null
|
791 |
+
value: null
|
792 |
+
ddp_bucket_cap_mb:
|
793 |
+
desc: null
|
794 |
+
value: null
|
795 |
+
ddp_broadcast_buffers:
|
796 |
+
desc: null
|
797 |
+
value: null
|
798 |
+
dataloader_pin_memory:
|
799 |
+
desc: null
|
800 |
+
value: true
|
801 |
+
dataloader_persistent_workers:
|
802 |
+
desc: null
|
803 |
+
value: false
|
804 |
+
skip_memory_metrics:
|
805 |
+
desc: null
|
806 |
+
value: true
|
807 |
+
use_legacy_prediction_loop:
|
808 |
+
desc: null
|
809 |
+
value: false
|
810 |
+
resume_from_checkpoint:
|
811 |
+
desc: null
|
812 |
+
value: null
|
813 |
+
hub_model_id:
|
814 |
+
desc: null
|
815 |
+
value: sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_
|
816 |
+
hub_strategy:
|
817 |
+
desc: null
|
818 |
+
value: every_save
|
819 |
+
hub_token:
|
820 |
+
desc: null
|
821 |
+
value: <HUB_TOKEN>
|
822 |
+
hub_private_repo:
|
823 |
+
desc: null
|
824 |
+
value: false
|
825 |
+
hub_always_push:
|
826 |
+
desc: null
|
827 |
+
value: false
|
828 |
+
gradient_checkpointing:
|
829 |
+
desc: null
|
830 |
+
value: false
|
831 |
+
gradient_checkpointing_kwargs:
|
832 |
+
desc: null
|
833 |
+
value:
|
834 |
+
use_reentrant: false
|
835 |
+
include_inputs_for_metrics:
|
836 |
+
desc: null
|
837 |
+
value: false
|
838 |
+
include_for_metrics:
|
839 |
+
desc: null
|
840 |
+
value: []
|
841 |
+
eval_do_concat_batches:
|
842 |
+
desc: null
|
843 |
+
value: true
|
844 |
+
fp16_backend:
|
845 |
+
desc: null
|
846 |
+
value: auto
|
847 |
+
evaluation_strategy:
|
848 |
+
desc: null
|
849 |
+
value: null
|
850 |
+
push_to_hub_model_id:
|
851 |
+
desc: null
|
852 |
+
value: null
|
853 |
+
push_to_hub_organization:
|
854 |
+
desc: null
|
855 |
+
value: null
|
856 |
+
push_to_hub_token:
|
857 |
+
desc: null
|
858 |
+
value: <PUSH_TO_HUB_TOKEN>
|
859 |
+
mp_parameters:
|
860 |
+
desc: null
|
861 |
+
value: ''
|
862 |
+
auto_find_batch_size:
|
863 |
+
desc: null
|
864 |
+
value: false
|
865 |
+
full_determinism:
|
866 |
+
desc: null
|
867 |
+
value: false
|
868 |
+
torchdynamo:
|
869 |
+
desc: null
|
870 |
+
value: null
|
871 |
+
ray_scope:
|
872 |
+
desc: null
|
873 |
+
value: last
|
874 |
+
ddp_timeout:
|
875 |
+
desc: null
|
876 |
+
value: 1800
|
877 |
+
torch_compile:
|
878 |
+
desc: null
|
879 |
+
value: false
|
880 |
+
torch_compile_backend:
|
881 |
+
desc: null
|
882 |
+
value: null
|
883 |
+
torch_compile_mode:
|
884 |
+
desc: null
|
885 |
+
value: null
|
886 |
+
dispatch_batches:
|
887 |
+
desc: null
|
888 |
+
value: null
|
889 |
+
split_batches:
|
890 |
+
desc: null
|
891 |
+
value: null
|
892 |
+
include_tokens_per_second:
|
893 |
+
desc: null
|
894 |
+
value: false
|
895 |
+
include_num_input_tokens_seen:
|
896 |
+
desc: null
|
897 |
+
value: false
|
898 |
+
neftune_noise_alpha:
|
899 |
+
desc: null
|
900 |
+
value: null
|
901 |
+
optim_target_modules:
|
902 |
+
desc: null
|
903 |
+
value: null
|
904 |
+
batch_eval_metrics:
|
905 |
+
desc: null
|
906 |
+
value: false
|
907 |
+
eval_on_start:
|
908 |
+
desc: null
|
909 |
+
value: false
|
910 |
+
use_liger_kernel:
|
911 |
+
desc: null
|
912 |
+
value: false
|
913 |
+
eval_use_gather_object:
|
914 |
+
desc: null
|
915 |
+
value: false
|
916 |
+
dataset_text_field:
|
917 |
+
desc: null
|
918 |
+
value: text
|
919 |
+
packing:
|
920 |
+
desc: null
|
921 |
+
value: false
|
922 |
+
dataset_num_proc:
|
923 |
+
desc: null
|
924 |
+
value: null
|
925 |
+
dataset_batch_size:
|
926 |
+
desc: null
|
927 |
+
value: 64
|
928 |
+
model_init_kwargs:
|
929 |
+
desc: null
|
930 |
+
value: null
|
931 |
+
dataset_kwargs:
|
932 |
+
desc: null
|
933 |
+
value: {}
|
934 |
+
eval_packing:
|
935 |
+
desc: null
|
936 |
+
value: null
|
937 |
+
num_of_sequences:
|
938 |
+
desc: null
|
939 |
+
value: 1024
|
940 |
+
chars_per_token:
|
941 |
+
desc: null
|
942 |
+
value: <CHARS_PER_TOKEN>
|
943 |
+
use_liger:
|
944 |
+
desc: null
|
945 |
+
value: false
|
946 |
+
model/num_parameters:
|
947 |
+
desc: null
|
948 |
+
value: 951599936
|
wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/diff.patch
ADDED
@@ -0,0 +1,49 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
diff --git a/qlearning_reasoning/training/sft.py b/qlearning_reasoning/training/sft.py
|
2 |
+
index 583fba2..1c2b359 100644
|
3 |
+
--- a/qlearning_reasoning/training/sft.py
|
4 |
+
+++ b/qlearning_reasoning/training/sft.py
|
5 |
+
@@ -51,6 +51,7 @@ flags.DEFINE_string('fsdp', "full_shard auto_wrap", 'If to use FSDP')
|
6 |
+
flags.DEFINE_string('fsdp_config', "qlearning_reasoning/accelerate_configs/fsdp_config.json", 'FSDP config to use.')
|
7 |
+
flags.DEFINE_boolean("push_to_hub", False, "Should the model be uploaded to Huggingface hub?")
|
8 |
+
flags.DEFINE_boolean('add_step_tokens', True, 'Add step tokens')
|
9 |
+
+flags.DEFINE_string("hub_model_id", None, "Hub model id. If not defined it will be truncated run name.")
|
10 |
+
|
11 |
+
START_OF_STEP_TOKEN = '[STEP]'
|
12 |
+
END_OF_STEP_TOKEN = '[/STEP]'
|
13 |
+
@@ -60,6 +61,10 @@ START_OF_REVISION_TOKEN = '[REVISION]'
|
14 |
+
END_OF_REVISION_TOKEN = '[/REVISION]'
|
15 |
+
|
16 |
+
def main(_):
|
17 |
+
+
|
18 |
+
+ if FLAGS.hub_model_id is None:
|
19 |
+
+ FLAGS.hub_model_id = FLAGS.run_name[:96]
|
20 |
+
+
|
21 |
+
if not os.path.exists(FLAGS.output_dir):
|
22 |
+
os.makedirs(FLAGS.output_dir, exist_ok=True)
|
23 |
+
os.makedirs(f"{FLAGS.output_dir}/{FLAGS.run_name}", exist_ok=True)
|
24 |
+
@@ -219,11 +224,11 @@ def main(_):
|
25 |
+
optim=FLAGS.optim,
|
26 |
+
weight_decay=FLAGS.weight_decay,
|
27 |
+
warmup_steps=FLAGS.warmup_steps,
|
28 |
+
- max_steps=num_steps,
|
29 |
+
fsdp=FLAGS.fsdp,
|
30 |
+
fsdp_config=FLAGS.fsdp_config if FLAGS.fsdp != '' else None,
|
31 |
+
push_to_hub=FLAGS.push_to_hub,
|
32 |
+
num_train_epochs=num_train_epochs,
|
33 |
+
+ hub_model_id=FLAGS.hub_model_id
|
34 |
+
)
|
35 |
+
|
36 |
+
flags_dict = FLAGS.flag_values_dict()
|
37 |
+
diff --git a/sagemaker/launch_sagemaker.py b/sagemaker/launch_sagemaker.py
|
38 |
+
index 3316ede..98ed205 100644
|
39 |
+
--- a/sagemaker/launch_sagemaker.py
|
40 |
+
+++ b/sagemaker/launch_sagemaker.py
|
41 |
+
@@ -224,7 +224,7 @@ def main(argv):
|
42 |
+
# TODO: verify if this breaks the code
|
43 |
+
environment = {
|
44 |
+
"PYTHONPATH": "/opt/ml/code/qlearning_reasoning/",
|
45 |
+
- "HF_HOME": "/opt/ml/input/data/training/hf-cache",
|
46 |
+
+ "HF_HOME": "/opt/ml/data/input/.cache",
|
47 |
+
"HF_TOKEN": get_hf_token(),
|
48 |
+
"WANDB_API_KEY": get_wandb_api_key(),
|
49 |
+
"SM_USE_RESERVED_CAPACITY": "1",
|
wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
ADDED
The diff for this file is too large to render.
See raw diff
|
|
wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/requirements.txt
ADDED
@@ -0,0 +1,289 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
absl-py==2.1.0
|
2 |
+
accelerate==1.1.1
|
3 |
+
aiohappyeyeballs==2.4.3
|
4 |
+
aiohttp==3.10.10
|
5 |
+
aiosignal==1.3.1
|
6 |
+
annotated-types==0.7.0
|
7 |
+
antlr4-python3-runtime==4.9.3
|
8 |
+
apex==0.1
|
9 |
+
appdirs==1.4.4
|
10 |
+
argparse==1.4.0
|
11 |
+
asttokens==2.4.1
|
12 |
+
async-timeout==4.0.3
|
13 |
+
attrdict==2.0.1
|
14 |
+
attrs==23.2.0
|
15 |
+
autocommand==2.2.2
|
16 |
+
awscli==1.33.40
|
17 |
+
backcall==0.2.0
|
18 |
+
backports.tarfile==1.2.0
|
19 |
+
bcrypt==4.2.0
|
20 |
+
beautifulsoup4==4.12.3
|
21 |
+
bleach==6.2.0
|
22 |
+
blis==0.7.10
|
23 |
+
bokeh==3.5.1
|
24 |
+
boltons==23.0.0
|
25 |
+
boto3==1.34.158
|
26 |
+
botocore==1.34.158
|
27 |
+
brotli==1.0.9
|
28 |
+
cached-property==1.5.2
|
29 |
+
cachetools==5.5.0
|
30 |
+
catalogue==2.0.10
|
31 |
+
certifi==2024.7.4
|
32 |
+
cffi==1.15.1
|
33 |
+
chardet==5.2.0
|
34 |
+
charset-normalizer==3.2.0
|
35 |
+
click==8.1.7
|
36 |
+
cloudpathlib==0.18.1
|
37 |
+
cloudpickle==2.2.1
|
38 |
+
cmake==3.30.2
|
39 |
+
colorama==0.4.6
|
40 |
+
comm==0.2.2
|
41 |
+
conda-libmamba-solver==23.3.0
|
42 |
+
conda-package-handling==2.2.0
|
43 |
+
conda-package-streaming==0.9.0
|
44 |
+
conda==23.3.1
|
45 |
+
confection==0.1.4
|
46 |
+
contourpy==1.2.1
|
47 |
+
cryptography==41.0.7
|
48 |
+
cycler==0.12.1
|
49 |
+
cymem==2.0.8
|
50 |
+
cython==3.0.11
|
51 |
+
datasets==3.1.0
|
52 |
+
debugpy==1.8.5
|
53 |
+
decorator==5.1.1
|
54 |
+
defusedxml==0.7.1
|
55 |
+
dill==0.3.8
|
56 |
+
docker-pycreds==0.4.0
|
57 |
+
docker==7.1.0
|
58 |
+
docopt==0.6.2
|
59 |
+
docstring-parser==0.16
|
60 |
+
docutils==0.16
|
61 |
+
einops==0.8.0
|
62 |
+
exceptiongroup==1.2.2
|
63 |
+
executing==2.0.1
|
64 |
+
fastai==2.7.16
|
65 |
+
fastcore==1.5.55
|
66 |
+
fastdownload==0.0.7
|
67 |
+
fastjsonschema==2.20.0
|
68 |
+
fastprogress==1.0.3
|
69 |
+
filelock==3.15.4
|
70 |
+
fire==0.6.0
|
71 |
+
flash-attn==2.5.5
|
72 |
+
fonttools==4.53.1
|
73 |
+
frozenlist==1.5.0
|
74 |
+
fsspec==2024.6.1
|
75 |
+
gevent==24.2.1
|
76 |
+
gitdb==4.0.11
|
77 |
+
gitpython==3.1.43
|
78 |
+
gmpy2==2.1.5
|
79 |
+
google-pasta==0.2.0
|
80 |
+
greenlet==3.0.3
|
81 |
+
h5py==3.11.0
|
82 |
+
huggingface-hub==0.26.2
|
83 |
+
hydra-core==1.3.2
|
84 |
+
idna==3.7
|
85 |
+
imageio==2.34.2
|
86 |
+
importlib-metadata==6.11.0
|
87 |
+
importlib-resources==6.1.1
|
88 |
+
inflect==7.3.1
|
89 |
+
inotify-simple==1.2.1
|
90 |
+
ipykernel==6.29.1
|
91 |
+
ipython==8.12.3
|
92 |
+
jaraco.context==5.3.0
|
93 |
+
jaraco.functools==4.0.1
|
94 |
+
jaraco.text==3.12.1
|
95 |
+
jedi==0.19.1
|
96 |
+
jinja2==3.1.4
|
97 |
+
jmespath==1.0.1
|
98 |
+
joblib==1.4.2
|
99 |
+
jsonpatch==1.32
|
100 |
+
jsonpointer==2.0
|
101 |
+
jsonschema-specifications==2023.12.1
|
102 |
+
jsonschema==4.23.0
|
103 |
+
jupyter-client==8.6.2
|
104 |
+
jupyter-core==5.7.2
|
105 |
+
jupyterlab-pygments==0.3.0
|
106 |
+
kiwisolver==1.4.5
|
107 |
+
langcodes==3.4.0
|
108 |
+
language-data==1.2.0
|
109 |
+
libmambapy==1.4.2
|
110 |
+
llvmlite==0.43.0
|
111 |
+
mamba==1.4.2
|
112 |
+
marisa-trie==1.1.0
|
113 |
+
markdown-it-py==3.0.0
|
114 |
+
markupsafe==2.1.5
|
115 |
+
matplotlib-inline==0.1.7
|
116 |
+
matplotlib==3.9.1
|
117 |
+
mdurl==0.1.2
|
118 |
+
mistune==3.0.2
|
119 |
+
more-itertools==10.3.0
|
120 |
+
mpi4py==3.1.6
|
121 |
+
mpmath==1.3.0
|
122 |
+
multidict==6.1.0
|
123 |
+
multiprocess==0.70.16
|
124 |
+
munkres==1.1.4
|
125 |
+
murmurhash==1.0.10
|
126 |
+
nbclient==0.10.0
|
127 |
+
nbconvert==7.16.4
|
128 |
+
nbformat==5.10.4
|
129 |
+
nest-asyncio==1.6.0
|
130 |
+
networkx==3.3
|
131 |
+
ninja==1.11.1.1
|
132 |
+
numba==0.60.0
|
133 |
+
numpy==1.24.3
|
134 |
+
nvidia-cublas-cu12==12.1.3.1
|
135 |
+
nvidia-cuda-cupti-cu12==12.1.105
|
136 |
+
nvidia-cuda-nvrtc-cu12==12.1.105
|
137 |
+
nvidia-cuda-runtime-cu12==12.1.105
|
138 |
+
nvidia-cudnn-cu12==8.9.2.26
|
139 |
+
nvidia-cufft-cu12==11.0.2.54
|
140 |
+
nvidia-curand-cu12==10.3.2.106
|
141 |
+
nvidia-cusolver-cu12==11.4.5.107
|
142 |
+
nvidia-cusparse-cu12==12.1.0.106
|
143 |
+
nvidia-ml-py==12.535.161
|
144 |
+
nvidia-nccl-cu12==2.19.3
|
145 |
+
nvidia-nvjitlink-cu12==12.6.77
|
146 |
+
nvidia-nvtx-cu12==12.1.105
|
147 |
+
nvitop==1.3.2
|
148 |
+
omegaconf==2.3.0
|
149 |
+
opencv-python==4.10.0.84
|
150 |
+
ordered-set==4.1.0
|
151 |
+
packaging==23.1
|
152 |
+
pandas==2.2.2
|
153 |
+
pandocfilters==1.5.1
|
154 |
+
paramiko==3.4.0
|
155 |
+
parso==0.8.4
|
156 |
+
pathos==0.3.2
|
157 |
+
pexpect==4.9.0
|
158 |
+
pickleshare==0.7.5
|
159 |
+
pillow==10.3.0
|
160 |
+
pip-chill==1.0.3
|
161 |
+
pip==24.1.2
|
162 |
+
pipreqs==0.5.0
|
163 |
+
platformdirs==4.2.2
|
164 |
+
plotly==5.23.0
|
165 |
+
pluggy==1.3.0
|
166 |
+
ply==3.11
|
167 |
+
pox==0.3.4
|
168 |
+
ppft==1.7.6.8
|
169 |
+
preshed==3.0.9
|
170 |
+
prompt-toolkit==3.0.47
|
171 |
+
propcache==0.2.0
|
172 |
+
protobuf==3.20.3
|
173 |
+
psutil==6.0.0
|
174 |
+
ptyprocess==0.7.0
|
175 |
+
pure-eval==0.2.3
|
176 |
+
pyarrow==17.0.0
|
177 |
+
pyasn1==0.6.0
|
178 |
+
pybind11-global==2.13.1
|
179 |
+
pybind11==2.13.1
|
180 |
+
pycosat==0.6.4
|
181 |
+
pycparser==2.21
|
182 |
+
pydantic-core==2.20.1
|
183 |
+
pydantic==2.8.2
|
184 |
+
pyfunctional==1.5.0
|
185 |
+
pygments==2.18.0
|
186 |
+
pynacl==1.5.0
|
187 |
+
pyopenssl==23.2.0
|
188 |
+
pyparsing==3.1.2
|
189 |
+
pyqt5-sip==12.12.2
|
190 |
+
pyqt5==5.15.9
|
191 |
+
pysocks==1.7.1
|
192 |
+
python-dateutil==2.9.0
|
193 |
+
pytz==2024.1
|
194 |
+
pyyaml==6.0.2
|
195 |
+
pyzmq==26.1.0
|
196 |
+
referencing==0.35.1
|
197 |
+
regex==2024.11.6
|
198 |
+
requests==2.32.3
|
199 |
+
retrying==1.3.4
|
200 |
+
rich==13.7.1
|
201 |
+
rpds-py==0.20.0
|
202 |
+
rsa==4.7.2
|
203 |
+
ruamel.yaml.clib==0.2.7
|
204 |
+
ruamel.yaml==0.17.32
|
205 |
+
s3fs==0.4.2
|
206 |
+
s3transfer==0.10.2
|
207 |
+
safetensors==0.4.5
|
208 |
+
sagemaker-experiments==0.1.45
|
209 |
+
sagemaker-pytorch-training==2.8.1
|
210 |
+
sagemaker-training==4.8.1
|
211 |
+
sagemaker==2.224.2
|
212 |
+
schema==0.7.7
|
213 |
+
scikit-learn==1.3.2
|
214 |
+
scipy==1.14.0
|
215 |
+
seaborn==0.13.2
|
216 |
+
sentencepiece==0.2.0
|
217 |
+
sentry-sdk==2.18.0
|
218 |
+
setproctitle==1.3.3
|
219 |
+
setuptools==72.1.0
|
220 |
+
shap==0.46.0
|
221 |
+
shellingham==1.5.4
|
222 |
+
shtab==1.7.1
|
223 |
+
sip==6.7.12
|
224 |
+
six==1.16.0
|
225 |
+
slicer==0.0.8
|
226 |
+
smart-open==7.0.4
|
227 |
+
smclarify==0.5
|
228 |
+
smdebug-rulesconfig==1.0.1
|
229 |
+
smdistributed-dataparallel==2.1.0
|
230 |
+
smmap==5.0.1
|
231 |
+
smprof==0.3.334
|
232 |
+
soupsieve==2.6
|
233 |
+
spacy-legacy==3.0.12
|
234 |
+
spacy-loggers==1.0.5
|
235 |
+
spacy==3.7.5
|
236 |
+
srsly==2.4.8
|
237 |
+
stack-data==0.6.2
|
238 |
+
sympy==1.13.0
|
239 |
+
tabulate==0.9.0
|
240 |
+
tblib==3.0.0
|
241 |
+
tenacity==9.0.0
|
242 |
+
tensor-parallel==2.0.0
|
243 |
+
termcolor==2.5.0
|
244 |
+
thinc==8.2.5
|
245 |
+
threadpoolctl==3.5.0
|
246 |
+
tinycss2==1.4.0
|
247 |
+
tokenizers==0.20.3
|
248 |
+
toml==0.10.2
|
249 |
+
tomli==2.0.1
|
250 |
+
toolz==0.12.0
|
251 |
+
torch==2.2.0
|
252 |
+
torchaudio==2.2.0
|
253 |
+
torchdata==0.7.0
|
254 |
+
torchnet==0.0.4
|
255 |
+
torchtext==0.16.0
|
256 |
+
torchvision==0.17.0
|
257 |
+
tornado==6.4.1
|
258 |
+
tqdm==4.66.5
|
259 |
+
traitlets==5.14.3
|
260 |
+
transformer-engine==0.12.0+170797
|
261 |
+
transformers==4.46.0
|
262 |
+
triton==2.2.0
|
263 |
+
trl==0.12.0
|
264 |
+
typeguard==4.3.0
|
265 |
+
typer-slim==0.12.3
|
266 |
+
typer==0.12.3
|
267 |
+
typing-extensions==4.12.2
|
268 |
+
tyro==0.8.5
|
269 |
+
tzdata==2024.1
|
270 |
+
unicodedata2==15.1.0
|
271 |
+
urllib3==1.26.19
|
272 |
+
visdom==0.2.4
|
273 |
+
wandb==0.16.3
|
274 |
+
wasabi==1.1.2
|
275 |
+
wcwidth==0.2.13
|
276 |
+
weasel==0.4.1
|
277 |
+
webencodings==0.5.1
|
278 |
+
websocket-client==1.8.0
|
279 |
+
werkzeug==3.0.3
|
280 |
+
wheel==0.41.2
|
281 |
+
wrapt==1.16.0
|
282 |
+
xxhash==3.5.0
|
283 |
+
xyzservices==2024.6.0
|
284 |
+
yarg==0.1.9
|
285 |
+
yarl==1.17.1
|
286 |
+
zipp==3.19.2
|
287 |
+
zope.event==5.0
|
288 |
+
zope.interface==7.0.1
|
289 |
+
zstandard==0.19.0
|
wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/wandb-metadata.json
ADDED
@@ -0,0 +1,1069 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"os": "Linux-5.10.226-214.879.amzn2.x86_64-x86_64-with-glibc2.31",
|
3 |
+
"python": "3.10.12",
|
4 |
+
"heartbeatAt": "2024-11-13T19:02:47.010671",
|
5 |
+
"startedAt": "2024-11-13T19:02:46.426230",
|
6 |
+
"docker": null,
|
7 |
+
"cuda": null,
|
8 |
+
"args": [
|
9 |
+
"--batch_size",
|
10 |
+
"64",
|
11 |
+
"--cache_dir",
|
12 |
+
"/opt/ml/data/input/.cache",
|
13 |
+
"--dataset",
|
14 |
+
"code-contests",
|
15 |
+
"--gradient_accumulation_steps",
|
16 |
+
"8",
|
17 |
+
"--logging_steps",
|
18 |
+
"5",
|
19 |
+
"--lr",
|
20 |
+
"1e-6",
|
21 |
+
"--max_seq_length",
|
22 |
+
"2048",
|
23 |
+
"--model_name",
|
24 |
+
"Qwen/Qwen2.5-Coder-7B-Instruct",
|
25 |
+
"--num_train_epochs",
|
26 |
+
"2",
|
27 |
+
"--output_dir",
|
28 |
+
"/opt/ml/model/",
|
29 |
+
"--push_to_hub",
|
30 |
+
"True",
|
31 |
+
"--run_name",
|
32 |
+
"sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048",
|
33 |
+
"--save_steps",
|
34 |
+
"100",
|
35 |
+
"--schedule",
|
36 |
+
"cosine",
|
37 |
+
"--wandb_project",
|
38 |
+
"sft-codecontests-1112",
|
39 |
+
"--weight_decay",
|
40 |
+
"0.0"
|
41 |
+
],
|
42 |
+
"state": "running",
|
43 |
+
"program": "/opt/ml/code/qlearning_reasoning/training/sft.py",
|
44 |
+
"codePathLocal": "qlearning_reasoning/training/sft.py",
|
45 |
+
"codePath": "qlearning_reasoning/training/sft.py",
|
46 |
+
"git": {
|
47 |
+
"remote": "git@github.com:TRI-ML/reasoning-value-verifiers.git",
|
48 |
+
"commit": "cbfd05147f15c55be83f4996eebdb305c17f119d"
|
49 |
+
},
|
50 |
+
"email": null,
|
51 |
+
"root": "/opt/ml/code",
|
52 |
+
"host": "algo-1",
|
53 |
+
"username": "root",
|
54 |
+
"executable": "/opt/conda/bin/python3.10",
|
55 |
+
"cpu_count": 96,
|
56 |
+
"cpu_count_logical": 192,
|
57 |
+
"cpu_freq": {
|
58 |
+
"current": 2829.910838541667,
|
59 |
+
"min": 0.0,
|
60 |
+
"max": 0.0
|
61 |
+
},
|
62 |
+
"cpu_freq_per_core": [
|
63 |
+
{
|
64 |
+
"current": 2850.788,
|
65 |
+
"min": 0.0,
|
66 |
+
"max": 0.0
|
67 |
+
},
|
68 |
+
{
|
69 |
+
"current": 3406.159,
|
70 |
+
"min": 0.0,
|
71 |
+
"max": 0.0
|
72 |
+
},
|
73 |
+
{
|
74 |
+
"current": 2605.944,
|
75 |
+
"min": 0.0,
|
76 |
+
"max": 0.0
|
77 |
+
},
|
78 |
+
{
|
79 |
+
"current": 2608.162,
|
80 |
+
"min": 0.0,
|
81 |
+
"max": 0.0
|
82 |
+
},
|
83 |
+
{
|
84 |
+
"current": 2818.169,
|
85 |
+
"min": 0.0,
|
86 |
+
"max": 0.0
|
87 |
+
},
|
88 |
+
{
|
89 |
+
"current": 2591.551,
|
90 |
+
"min": 0.0,
|
91 |
+
"max": 0.0
|
92 |
+
},
|
93 |
+
{
|
94 |
+
"current": 2621.486,
|
95 |
+
"min": 0.0,
|
96 |
+
"max": 0.0
|
97 |
+
},
|
98 |
+
{
|
99 |
+
"current": 2607.504,
|
100 |
+
"min": 0.0,
|
101 |
+
"max": 0.0
|
102 |
+
},
|
103 |
+
{
|
104 |
+
"current": 2999.734,
|
105 |
+
"min": 0.0,
|
106 |
+
"max": 0.0
|
107 |
+
},
|
108 |
+
{
|
109 |
+
"current": 2991.449,
|
110 |
+
"min": 0.0,
|
111 |
+
"max": 0.0
|
112 |
+
},
|
113 |
+
{
|
114 |
+
"current": 3123.177,
|
115 |
+
"min": 0.0,
|
116 |
+
"max": 0.0
|
117 |
+
},
|
118 |
+
{
|
119 |
+
"current": 3053.641,
|
120 |
+
"min": 0.0,
|
121 |
+
"max": 0.0
|
122 |
+
},
|
123 |
+
{
|
124 |
+
"current": 3003.014,
|
125 |
+
"min": 0.0,
|
126 |
+
"max": 0.0
|
127 |
+
},
|
128 |
+
{
|
129 |
+
"current": 3000.178,
|
130 |
+
"min": 0.0,
|
131 |
+
"max": 0.0
|
132 |
+
},
|
133 |
+
{
|
134 |
+
"current": 3432.256,
|
135 |
+
"min": 0.0,
|
136 |
+
"max": 0.0
|
137 |
+
},
|
138 |
+
{
|
139 |
+
"current": 2999.802,
|
140 |
+
"min": 0.0,
|
141 |
+
"max": 0.0
|
142 |
+
},
|
143 |
+
{
|
144 |
+
"current": 3005.725,
|
145 |
+
"min": 0.0,
|
146 |
+
"max": 0.0
|
147 |
+
},
|
148 |
+
{
|
149 |
+
"current": 3565.885,
|
150 |
+
"min": 0.0,
|
151 |
+
"max": 0.0
|
152 |
+
},
|
153 |
+
{
|
154 |
+
"current": 2992.339,
|
155 |
+
"min": 0.0,
|
156 |
+
"max": 0.0
|
157 |
+
},
|
158 |
+
{
|
159 |
+
"current": 3104.468,
|
160 |
+
"min": 0.0,
|
161 |
+
"max": 0.0
|
162 |
+
},
|
163 |
+
{
|
164 |
+
"current": 3182.411,
|
165 |
+
"min": 0.0,
|
166 |
+
"max": 0.0
|
167 |
+
},
|
168 |
+
{
|
169 |
+
"current": 3013.217,
|
170 |
+
"min": 0.0,
|
171 |
+
"max": 0.0
|
172 |
+
},
|
173 |
+
{
|
174 |
+
"current": 3131.904,
|
175 |
+
"min": 0.0,
|
176 |
+
"max": 0.0
|
177 |
+
},
|
178 |
+
{
|
179 |
+
"current": 3007.314,
|
180 |
+
"min": 0.0,
|
181 |
+
"max": 0.0
|
182 |
+
},
|
183 |
+
{
|
184 |
+
"current": 2625.175,
|
185 |
+
"min": 0.0,
|
186 |
+
"max": 0.0
|
187 |
+
},
|
188 |
+
{
|
189 |
+
"current": 2653.407,
|
190 |
+
"min": 0.0,
|
191 |
+
"max": 0.0
|
192 |
+
},
|
193 |
+
{
|
194 |
+
"current": 2610.871,
|
195 |
+
"min": 0.0,
|
196 |
+
"max": 0.0
|
197 |
+
},
|
198 |
+
{
|
199 |
+
"current": 2793.855,
|
200 |
+
"min": 0.0,
|
201 |
+
"max": 0.0
|
202 |
+
},
|
203 |
+
{
|
204 |
+
"current": 2619.232,
|
205 |
+
"min": 0.0,
|
206 |
+
"max": 0.0
|
207 |
+
},
|
208 |
+
{
|
209 |
+
"current": 2627.886,
|
210 |
+
"min": 0.0,
|
211 |
+
"max": 0.0
|
212 |
+
},
|
213 |
+
{
|
214 |
+
"current": 2613.748,
|
215 |
+
"min": 0.0,
|
216 |
+
"max": 0.0
|
217 |
+
},
|
218 |
+
{
|
219 |
+
"current": 2318.039,
|
220 |
+
"min": 0.0,
|
221 |
+
"max": 0.0
|
222 |
+
},
|
223 |
+
{
|
224 |
+
"current": 3546.661,
|
225 |
+
"min": 0.0,
|
226 |
+
"max": 0.0
|
227 |
+
},
|
228 |
+
{
|
229 |
+
"current": 3022.56,
|
230 |
+
"min": 0.0,
|
231 |
+
"max": 0.0
|
232 |
+
},
|
233 |
+
{
|
234 |
+
"current": 3035.28,
|
235 |
+
"min": 0.0,
|
236 |
+
"max": 0.0
|
237 |
+
},
|
238 |
+
{
|
239 |
+
"current": 3541.443,
|
240 |
+
"min": 0.0,
|
241 |
+
"max": 0.0
|
242 |
+
},
|
243 |
+
{
|
244 |
+
"current": 3163.55,
|
245 |
+
"min": 0.0,
|
246 |
+
"max": 0.0
|
247 |
+
},
|
248 |
+
{
|
249 |
+
"current": 3023.79,
|
250 |
+
"min": 0.0,
|
251 |
+
"max": 0.0
|
252 |
+
},
|
253 |
+
{
|
254 |
+
"current": 3033.204,
|
255 |
+
"min": 0.0,
|
256 |
+
"max": 0.0
|
257 |
+
},
|
258 |
+
{
|
259 |
+
"current": 3031.695,
|
260 |
+
"min": 0.0,
|
261 |
+
"max": 0.0
|
262 |
+
},
|
263 |
+
{
|
264 |
+
"current": 3034.587,
|
265 |
+
"min": 0.0,
|
266 |
+
"max": 0.0
|
267 |
+
},
|
268 |
+
{
|
269 |
+
"current": 3031.71,
|
270 |
+
"min": 0.0,
|
271 |
+
"max": 0.0
|
272 |
+
},
|
273 |
+
{
|
274 |
+
"current": 3029.116,
|
275 |
+
"min": 0.0,
|
276 |
+
"max": 0.0
|
277 |
+
},
|
278 |
+
{
|
279 |
+
"current": 3293.701,
|
280 |
+
"min": 0.0,
|
281 |
+
"max": 0.0
|
282 |
+
},
|
283 |
+
{
|
284 |
+
"current": 3033.78,
|
285 |
+
"min": 0.0,
|
286 |
+
"max": 0.0
|
287 |
+
},
|
288 |
+
{
|
289 |
+
"current": 3032.034,
|
290 |
+
"min": 0.0,
|
291 |
+
"max": 0.0
|
292 |
+
},
|
293 |
+
{
|
294 |
+
"current": 3057.365,
|
295 |
+
"min": 0.0,
|
296 |
+
"max": 0.0
|
297 |
+
},
|
298 |
+
{
|
299 |
+
"current": 3028.183,
|
300 |
+
"min": 0.0,
|
301 |
+
"max": 0.0
|
302 |
+
},
|
303 |
+
{
|
304 |
+
"current": 3596.439,
|
305 |
+
"min": 0.0,
|
306 |
+
"max": 0.0
|
307 |
+
},
|
308 |
+
{
|
309 |
+
"current": 3327.853,
|
310 |
+
"min": 0.0,
|
311 |
+
"max": 0.0
|
312 |
+
},
|
313 |
+
{
|
314 |
+
"current": 3226.463,
|
315 |
+
"min": 0.0,
|
316 |
+
"max": 0.0
|
317 |
+
},
|
318 |
+
{
|
319 |
+
"current": 3259.361,
|
320 |
+
"min": 0.0,
|
321 |
+
"max": 0.0
|
322 |
+
},
|
323 |
+
{
|
324 |
+
"current": 3284.076,
|
325 |
+
"min": 0.0,
|
326 |
+
"max": 0.0
|
327 |
+
},
|
328 |
+
{
|
329 |
+
"current": 3257.698,
|
330 |
+
"min": 0.0,
|
331 |
+
"max": 0.0
|
332 |
+
},
|
333 |
+
{
|
334 |
+
"current": 3280.289,
|
335 |
+
"min": 0.0,
|
336 |
+
"max": 0.0
|
337 |
+
},
|
338 |
+
{
|
339 |
+
"current": 3258.834,
|
340 |
+
"min": 0.0,
|
341 |
+
"max": 0.0
|
342 |
+
},
|
343 |
+
{
|
344 |
+
"current": 2517.346,
|
345 |
+
"min": 0.0,
|
346 |
+
"max": 0.0
|
347 |
+
},
|
348 |
+
{
|
349 |
+
"current": 2553.091,
|
350 |
+
"min": 0.0,
|
351 |
+
"max": 0.0
|
352 |
+
},
|
353 |
+
{
|
354 |
+
"current": 2552.693,
|
355 |
+
"min": 0.0,
|
356 |
+
"max": 0.0
|
357 |
+
},
|
358 |
+
{
|
359 |
+
"current": 2689.091,
|
360 |
+
"min": 0.0,
|
361 |
+
"max": 0.0
|
362 |
+
},
|
363 |
+
{
|
364 |
+
"current": 2503.301,
|
365 |
+
"min": 0.0,
|
366 |
+
"max": 0.0
|
367 |
+
},
|
368 |
+
{
|
369 |
+
"current": 2524.4,
|
370 |
+
"min": 0.0,
|
371 |
+
"max": 0.0
|
372 |
+
},
|
373 |
+
{
|
374 |
+
"current": 2521.768,
|
375 |
+
"min": 0.0,
|
376 |
+
"max": 0.0
|
377 |
+
},
|
378 |
+
{
|
379 |
+
"current": 2530.364,
|
380 |
+
"min": 0.0,
|
381 |
+
"max": 0.0
|
382 |
+
},
|
383 |
+
{
|
384 |
+
"current": 3258.225,
|
385 |
+
"min": 0.0,
|
386 |
+
"max": 0.0
|
387 |
+
},
|
388 |
+
{
|
389 |
+
"current": 3597.406,
|
390 |
+
"min": 0.0,
|
391 |
+
"max": 0.0
|
392 |
+
},
|
393 |
+
{
|
394 |
+
"current": 3279.658,
|
395 |
+
"min": 0.0,
|
396 |
+
"max": 0.0
|
397 |
+
},
|
398 |
+
{
|
399 |
+
"current": 3292.257,
|
400 |
+
"min": 0.0,
|
401 |
+
"max": 0.0
|
402 |
+
},
|
403 |
+
{
|
404 |
+
"current": 3265.351,
|
405 |
+
"min": 0.0,
|
406 |
+
"max": 0.0
|
407 |
+
},
|
408 |
+
{
|
409 |
+
"current": 3286.715,
|
410 |
+
"min": 0.0,
|
411 |
+
"max": 0.0
|
412 |
+
},
|
413 |
+
{
|
414 |
+
"current": 3268.777,
|
415 |
+
"min": 0.0,
|
416 |
+
"max": 0.0
|
417 |
+
},
|
418 |
+
{
|
419 |
+
"current": 3242.068,
|
420 |
+
"min": 0.0,
|
421 |
+
"max": 0.0
|
422 |
+
},
|
423 |
+
{
|
424 |
+
"current": 3288.908,
|
425 |
+
"min": 0.0,
|
426 |
+
"max": 0.0
|
427 |
+
},
|
428 |
+
{
|
429 |
+
"current": 3278.645,
|
430 |
+
"min": 0.0,
|
431 |
+
"max": 0.0
|
432 |
+
},
|
433 |
+
{
|
434 |
+
"current": 3215.578,
|
435 |
+
"min": 0.0,
|
436 |
+
"max": 0.0
|
437 |
+
},
|
438 |
+
{
|
439 |
+
"current": 3327.101,
|
440 |
+
"min": 0.0,
|
441 |
+
"max": 0.0
|
442 |
+
},
|
443 |
+
{
|
444 |
+
"current": 3357.075,
|
445 |
+
"min": 0.0,
|
446 |
+
"max": 0.0
|
447 |
+
},
|
448 |
+
{
|
449 |
+
"current": 3144.838,
|
450 |
+
"min": 0.0,
|
451 |
+
"max": 0.0
|
452 |
+
},
|
453 |
+
{
|
454 |
+
"current": 3005.517,
|
455 |
+
"min": 0.0,
|
456 |
+
"max": 0.0
|
457 |
+
},
|
458 |
+
{
|
459 |
+
"current": 3597.381,
|
460 |
+
"min": 0.0,
|
461 |
+
"max": 0.0
|
462 |
+
},
|
463 |
+
{
|
464 |
+
"current": 3596.175,
|
465 |
+
"min": 0.0,
|
466 |
+
"max": 0.0
|
467 |
+
},
|
468 |
+
{
|
469 |
+
"current": 3317.258,
|
470 |
+
"min": 0.0,
|
471 |
+
"max": 0.0
|
472 |
+
},
|
473 |
+
{
|
474 |
+
"current": 3310.249,
|
475 |
+
"min": 0.0,
|
476 |
+
"max": 0.0
|
477 |
+
},
|
478 |
+
{
|
479 |
+
"current": 3140.687,
|
480 |
+
"min": 0.0,
|
481 |
+
"max": 0.0
|
482 |
+
},
|
483 |
+
{
|
484 |
+
"current": 3398.954,
|
485 |
+
"min": 0.0,
|
486 |
+
"max": 0.0
|
487 |
+
},
|
488 |
+
{
|
489 |
+
"current": 3318.433,
|
490 |
+
"min": 0.0,
|
491 |
+
"max": 0.0
|
492 |
+
},
|
493 |
+
{
|
494 |
+
"current": 3342.35,
|
495 |
+
"min": 0.0,
|
496 |
+
"max": 0.0
|
497 |
+
},
|
498 |
+
{
|
499 |
+
"current": 3328.575,
|
500 |
+
"min": 0.0,
|
501 |
+
"max": 0.0
|
502 |
+
},
|
503 |
+
{
|
504 |
+
"current": 3499.892,
|
505 |
+
"min": 0.0,
|
506 |
+
"max": 0.0
|
507 |
+
},
|
508 |
+
{
|
509 |
+
"current": 3344.864,
|
510 |
+
"min": 0.0,
|
511 |
+
"max": 0.0
|
512 |
+
},
|
513 |
+
{
|
514 |
+
"current": 3355.802,
|
515 |
+
"min": 0.0,
|
516 |
+
"max": 0.0
|
517 |
+
},
|
518 |
+
{
|
519 |
+
"current": 3305.032,
|
520 |
+
"min": 0.0,
|
521 |
+
"max": 0.0
|
522 |
+
},
|
523 |
+
{
|
524 |
+
"current": 3318.07,
|
525 |
+
"min": 0.0,
|
526 |
+
"max": 0.0
|
527 |
+
},
|
528 |
+
{
|
529 |
+
"current": 3336.546,
|
530 |
+
"min": 0.0,
|
531 |
+
"max": 0.0
|
532 |
+
},
|
533 |
+
{
|
534 |
+
"current": 3335.458,
|
535 |
+
"min": 0.0,
|
536 |
+
"max": 0.0
|
537 |
+
},
|
538 |
+
{
|
539 |
+
"current": 3596.561,
|
540 |
+
"min": 0.0,
|
541 |
+
"max": 0.0
|
542 |
+
},
|
543 |
+
{
|
544 |
+
"current": 2884.473,
|
545 |
+
"min": 0.0,
|
546 |
+
"max": 0.0
|
547 |
+
},
|
548 |
+
{
|
549 |
+
"current": 2934.098,
|
550 |
+
"min": 0.0,
|
551 |
+
"max": 0.0
|
552 |
+
},
|
553 |
+
{
|
554 |
+
"current": 2621.598,
|
555 |
+
"min": 0.0,
|
556 |
+
"max": 0.0
|
557 |
+
},
|
558 |
+
{
|
559 |
+
"current": 2645.184,
|
560 |
+
"min": 0.0,
|
561 |
+
"max": 0.0
|
562 |
+
},
|
563 |
+
{
|
564 |
+
"current": 2480.726,
|
565 |
+
"min": 0.0,
|
566 |
+
"max": 0.0
|
567 |
+
},
|
568 |
+
{
|
569 |
+
"current": 2515.912,
|
570 |
+
"min": 0.0,
|
571 |
+
"max": 0.0
|
572 |
+
},
|
573 |
+
{
|
574 |
+
"current": 2702.57,
|
575 |
+
"min": 0.0,
|
576 |
+
"max": 0.0
|
577 |
+
},
|
578 |
+
{
|
579 |
+
"current": 2671.473,
|
580 |
+
"min": 0.0,
|
581 |
+
"max": 0.0
|
582 |
+
},
|
583 |
+
{
|
584 |
+
"current": 3092.321,
|
585 |
+
"min": 0.0,
|
586 |
+
"max": 0.0
|
587 |
+
},
|
588 |
+
{
|
589 |
+
"current": 3092.204,
|
590 |
+
"min": 0.0,
|
591 |
+
"max": 0.0
|
592 |
+
},
|
593 |
+
{
|
594 |
+
"current": 3217.287,
|
595 |
+
"min": 0.0,
|
596 |
+
"max": 0.0
|
597 |
+
},
|
598 |
+
{
|
599 |
+
"current": 3161.646,
|
600 |
+
"min": 0.0,
|
601 |
+
"max": 0.0
|
602 |
+
},
|
603 |
+
{
|
604 |
+
"current": 3089.273,
|
605 |
+
"min": 0.0,
|
606 |
+
"max": 0.0
|
607 |
+
},
|
608 |
+
{
|
609 |
+
"current": 3088.447,
|
610 |
+
"min": 0.0,
|
611 |
+
"max": 0.0
|
612 |
+
},
|
613 |
+
{
|
614 |
+
"current": 3559.214,
|
615 |
+
"min": 0.0,
|
616 |
+
"max": 0.0
|
617 |
+
},
|
618 |
+
{
|
619 |
+
"current": 3218.611,
|
620 |
+
"min": 0.0,
|
621 |
+
"max": 0.0
|
622 |
+
},
|
623 |
+
{
|
624 |
+
"current": 3021.622,
|
625 |
+
"min": 0.0,
|
626 |
+
"max": 0.0
|
627 |
+
},
|
628 |
+
{
|
629 |
+
"current": 3332.084,
|
630 |
+
"min": 0.0,
|
631 |
+
"max": 0.0
|
632 |
+
},
|
633 |
+
{
|
634 |
+
"current": 3019.854,
|
635 |
+
"min": 0.0,
|
636 |
+
"max": 0.0
|
637 |
+
},
|
638 |
+
{
|
639 |
+
"current": 3213.831,
|
640 |
+
"min": 0.0,
|
641 |
+
"max": 0.0
|
642 |
+
},
|
643 |
+
{
|
644 |
+
"current": 3092.774,
|
645 |
+
"min": 0.0,
|
646 |
+
"max": 0.0
|
647 |
+
},
|
648 |
+
{
|
649 |
+
"current": 3015.868,
|
650 |
+
"min": 0.0,
|
651 |
+
"max": 0.0
|
652 |
+
},
|
653 |
+
{
|
654 |
+
"current": 3082.752,
|
655 |
+
"min": 0.0,
|
656 |
+
"max": 0.0
|
657 |
+
},
|
658 |
+
{
|
659 |
+
"current": 3029.559,
|
660 |
+
"min": 0.0,
|
661 |
+
"max": 0.0
|
662 |
+
},
|
663 |
+
{
|
664 |
+
"current": 2873.142,
|
665 |
+
"min": 0.0,
|
666 |
+
"max": 0.0
|
667 |
+
},
|
668 |
+
{
|
669 |
+
"current": 2215.288,
|
670 |
+
"min": 0.0,
|
671 |
+
"max": 0.0
|
672 |
+
},
|
673 |
+
{
|
674 |
+
"current": 3193.443,
|
675 |
+
"min": 0.0,
|
676 |
+
"max": 0.0
|
677 |
+
},
|
678 |
+
{
|
679 |
+
"current": 2928.717,
|
680 |
+
"min": 0.0,
|
681 |
+
"max": 0.0
|
682 |
+
},
|
683 |
+
{
|
684 |
+
"current": 2990.366,
|
685 |
+
"min": 0.0,
|
686 |
+
"max": 0.0
|
687 |
+
},
|
688 |
+
{
|
689 |
+
"current": 2923.847,
|
690 |
+
"min": 0.0,
|
691 |
+
"max": 0.0
|
692 |
+
},
|
693 |
+
{
|
694 |
+
"current": 2869.364,
|
695 |
+
"min": 0.0,
|
696 |
+
"max": 0.0
|
697 |
+
},
|
698 |
+
{
|
699 |
+
"current": 2859.189,
|
700 |
+
"min": 0.0,
|
701 |
+
"max": 0.0
|
702 |
+
},
|
703 |
+
{
|
704 |
+
"current": 3049.001,
|
705 |
+
"min": 0.0,
|
706 |
+
"max": 0.0
|
707 |
+
},
|
708 |
+
{
|
709 |
+
"current": 2805.206,
|
710 |
+
"min": 0.0,
|
711 |
+
"max": 0.0
|
712 |
+
},
|
713 |
+
{
|
714 |
+
"current": 3094.081,
|
715 |
+
"min": 0.0,
|
716 |
+
"max": 0.0
|
717 |
+
},
|
718 |
+
{
|
719 |
+
"current": 3187.222,
|
720 |
+
"min": 0.0,
|
721 |
+
"max": 0.0
|
722 |
+
},
|
723 |
+
{
|
724 |
+
"current": 2853.945,
|
725 |
+
"min": 0.0,
|
726 |
+
"max": 0.0
|
727 |
+
},
|
728 |
+
{
|
729 |
+
"current": 2944.73,
|
730 |
+
"min": 0.0,
|
731 |
+
"max": 0.0
|
732 |
+
},
|
733 |
+
{
|
734 |
+
"current": 2809.161,
|
735 |
+
"min": 0.0,
|
736 |
+
"max": 0.0
|
737 |
+
},
|
738 |
+
{
|
739 |
+
"current": 2806.787,
|
740 |
+
"min": 0.0,
|
741 |
+
"max": 0.0
|
742 |
+
},
|
743 |
+
{
|
744 |
+
"current": 3024.225,
|
745 |
+
"min": 0.0,
|
746 |
+
"max": 0.0
|
747 |
+
},
|
748 |
+
{
|
749 |
+
"current": 3032.173,
|
750 |
+
"min": 0.0,
|
751 |
+
"max": 0.0
|
752 |
+
},
|
753 |
+
{
|
754 |
+
"current": 3033.195,
|
755 |
+
"min": 0.0,
|
756 |
+
"max": 0.0
|
757 |
+
},
|
758 |
+
{
|
759 |
+
"current": 3561.287,
|
760 |
+
"min": 0.0,
|
761 |
+
"max": 0.0
|
762 |
+
},
|
763 |
+
{
|
764 |
+
"current": 2987.872,
|
765 |
+
"min": 0.0,
|
766 |
+
"max": 0.0
|
767 |
+
},
|
768 |
+
{
|
769 |
+
"current": 3051.252,
|
770 |
+
"min": 0.0,
|
771 |
+
"max": 0.0
|
772 |
+
},
|
773 |
+
{
|
774 |
+
"current": 3379.017,
|
775 |
+
"min": 0.0,
|
776 |
+
"max": 0.0
|
777 |
+
},
|
778 |
+
{
|
779 |
+
"current": 3044.319,
|
780 |
+
"min": 0.0,
|
781 |
+
"max": 0.0
|
782 |
+
},
|
783 |
+
{
|
784 |
+
"current": 3595.62,
|
785 |
+
"min": 0.0,
|
786 |
+
"max": 0.0
|
787 |
+
},
|
788 |
+
{
|
789 |
+
"current": 3261.737,
|
790 |
+
"min": 0.0,
|
791 |
+
"max": 0.0
|
792 |
+
},
|
793 |
+
{
|
794 |
+
"current": 3257.013,
|
795 |
+
"min": 0.0,
|
796 |
+
"max": 0.0
|
797 |
+
},
|
798 |
+
{
|
799 |
+
"current": 3288.317,
|
800 |
+
"min": 0.0,
|
801 |
+
"max": 0.0
|
802 |
+
},
|
803 |
+
{
|
804 |
+
"current": 3268.675,
|
805 |
+
"min": 0.0,
|
806 |
+
"max": 0.0
|
807 |
+
},
|
808 |
+
{
|
809 |
+
"current": 3269.579,
|
810 |
+
"min": 0.0,
|
811 |
+
"max": 0.0
|
812 |
+
},
|
813 |
+
{
|
814 |
+
"current": 3266.213,
|
815 |
+
"min": 0.0,
|
816 |
+
"max": 0.0
|
817 |
+
},
|
818 |
+
{
|
819 |
+
"current": 3265.148,
|
820 |
+
"min": 0.0,
|
821 |
+
"max": 0.0
|
822 |
+
},
|
823 |
+
{
|
824 |
+
"current": 2554.152,
|
825 |
+
"min": 0.0,
|
826 |
+
"max": 0.0
|
827 |
+
},
|
828 |
+
{
|
829 |
+
"current": 2556.63,
|
830 |
+
"min": 0.0,
|
831 |
+
"max": 0.0
|
832 |
+
},
|
833 |
+
{
|
834 |
+
"current": 2540.571,
|
835 |
+
"min": 0.0,
|
836 |
+
"max": 0.0
|
837 |
+
},
|
838 |
+
{
|
839 |
+
"current": 2512.005,
|
840 |
+
"min": 0.0,
|
841 |
+
"max": 0.0
|
842 |
+
},
|
843 |
+
{
|
844 |
+
"current": 2560.992,
|
845 |
+
"min": 0.0,
|
846 |
+
"max": 0.0
|
847 |
+
},
|
848 |
+
{
|
849 |
+
"current": 2526.641,
|
850 |
+
"min": 0.0,
|
851 |
+
"max": 0.0
|
852 |
+
},
|
853 |
+
{
|
854 |
+
"current": 2528.267,
|
855 |
+
"min": 0.0,
|
856 |
+
"max": 0.0
|
857 |
+
},
|
858 |
+
{
|
859 |
+
"current": 2583.061,
|
860 |
+
"min": 0.0,
|
861 |
+
"max": 0.0
|
862 |
+
},
|
863 |
+
{
|
864 |
+
"current": 3257.652,
|
865 |
+
"min": 0.0,
|
866 |
+
"max": 0.0
|
867 |
+
},
|
868 |
+
{
|
869 |
+
"current": 3598.217,
|
870 |
+
"min": 0.0,
|
871 |
+
"max": 0.0
|
872 |
+
},
|
873 |
+
{
|
874 |
+
"current": 3282.578,
|
875 |
+
"min": 0.0,
|
876 |
+
"max": 0.0
|
877 |
+
},
|
878 |
+
{
|
879 |
+
"current": 3261.664,
|
880 |
+
"min": 0.0,
|
881 |
+
"max": 0.0
|
882 |
+
},
|
883 |
+
{
|
884 |
+
"current": 3268.246,
|
885 |
+
"min": 0.0,
|
886 |
+
"max": 0.0
|
887 |
+
},
|
888 |
+
{
|
889 |
+
"current": 3331.937,
|
890 |
+
"min": 0.0,
|
891 |
+
"max": 0.0
|
892 |
+
},
|
893 |
+
{
|
894 |
+
"current": 3265.603,
|
895 |
+
"min": 0.0,
|
896 |
+
"max": 0.0
|
897 |
+
},
|
898 |
+
{
|
899 |
+
"current": 3258.501,
|
900 |
+
"min": 0.0,
|
901 |
+
"max": 0.0
|
902 |
+
},
|
903 |
+
{
|
904 |
+
"current": 3337.045,
|
905 |
+
"min": 0.0,
|
906 |
+
"max": 0.0
|
907 |
+
},
|
908 |
+
{
|
909 |
+
"current": 3259.949,
|
910 |
+
"min": 0.0,
|
911 |
+
"max": 0.0
|
912 |
+
},
|
913 |
+
{
|
914 |
+
"current": 3226.058,
|
915 |
+
"min": 0.0,
|
916 |
+
"max": 0.0
|
917 |
+
},
|
918 |
+
{
|
919 |
+
"current": 3250.308,
|
920 |
+
"min": 0.0,
|
921 |
+
"max": 0.0
|
922 |
+
},
|
923 |
+
{
|
924 |
+
"current": 3208.026,
|
925 |
+
"min": 0.0,
|
926 |
+
"max": 0.0
|
927 |
+
},
|
928 |
+
{
|
929 |
+
"current": 2974.043,
|
930 |
+
"min": 0.0,
|
931 |
+
"max": 0.0
|
932 |
+
},
|
933 |
+
{
|
934 |
+
"current": 3216.706,
|
935 |
+
"min": 0.0,
|
936 |
+
"max": 0.0
|
937 |
+
},
|
938 |
+
{
|
939 |
+
"current": 3598.55,
|
940 |
+
"min": 0.0,
|
941 |
+
"max": 0.0
|
942 |
+
},
|
943 |
+
{
|
944 |
+
"current": 3598.261,
|
945 |
+
"min": 0.0,
|
946 |
+
"max": 0.0
|
947 |
+
},
|
948 |
+
{
|
949 |
+
"current": 3222.637,
|
950 |
+
"min": 0.0,
|
951 |
+
"max": 0.0
|
952 |
+
},
|
953 |
+
{
|
954 |
+
"current": 3227.662,
|
955 |
+
"min": 0.0,
|
956 |
+
"max": 0.0
|
957 |
+
},
|
958 |
+
{
|
959 |
+
"current": 3143.781,
|
960 |
+
"min": 0.0,
|
961 |
+
"max": 0.0
|
962 |
+
},
|
963 |
+
{
|
964 |
+
"current": 3215.567,
|
965 |
+
"min": 0.0,
|
966 |
+
"max": 0.0
|
967 |
+
},
|
968 |
+
{
|
969 |
+
"current": 3250.612,
|
970 |
+
"min": 0.0,
|
971 |
+
"max": 0.0
|
972 |
+
},
|
973 |
+
{
|
974 |
+
"current": 3195.784,
|
975 |
+
"min": 0.0,
|
976 |
+
"max": 0.0
|
977 |
+
},
|
978 |
+
{
|
979 |
+
"current": 3219.289,
|
980 |
+
"min": 0.0,
|
981 |
+
"max": 0.0
|
982 |
+
},
|
983 |
+
{
|
984 |
+
"current": 3289.163,
|
985 |
+
"min": 0.0,
|
986 |
+
"max": 0.0
|
987 |
+
},
|
988 |
+
{
|
989 |
+
"current": 3247.875,
|
990 |
+
"min": 0.0,
|
991 |
+
"max": 0.0
|
992 |
+
},
|
993 |
+
{
|
994 |
+
"current": 3249.112,
|
995 |
+
"min": 0.0,
|
996 |
+
"max": 0.0
|
997 |
+
},
|
998 |
+
{
|
999 |
+
"current": 3209.3,
|
1000 |
+
"min": 0.0,
|
1001 |
+
"max": 0.0
|
1002 |
+
},
|
1003 |
+
{
|
1004 |
+
"current": 3187.212,
|
1005 |
+
"min": 0.0,
|
1006 |
+
"max": 0.0
|
1007 |
+
},
|
1008 |
+
{
|
1009 |
+
"current": 3288.003,
|
1010 |
+
"min": 0.0,
|
1011 |
+
"max": 0.0
|
1012 |
+
},
|
1013 |
+
{
|
1014 |
+
"current": 3169.417,
|
1015 |
+
"min": 0.0,
|
1016 |
+
"max": 0.0
|
1017 |
+
},
|
1018 |
+
{
|
1019 |
+
"current": 3596.071,
|
1020 |
+
"min": 0.0,
|
1021 |
+
"max": 0.0
|
1022 |
+
}
|
1023 |
+
],
|
1024 |
+
"disk": {
|
1025 |
+
"/": {
|
1026 |
+
"total": 119.94140625,
|
1027 |
+
"used": 56.4015998840332
|
1028 |
+
}
|
1029 |
+
},
|
1030 |
+
"gpu": "NVIDIA H100 80GB HBM3",
|
1031 |
+
"gpu_count": 8,
|
1032 |
+
"gpu_devices": [
|
1033 |
+
{
|
1034 |
+
"name": "NVIDIA H100 80GB HBM3",
|
1035 |
+
"memory_total": 85520809984
|
1036 |
+
},
|
1037 |
+
{
|
1038 |
+
"name": "NVIDIA H100 80GB HBM3",
|
1039 |
+
"memory_total": 85520809984
|
1040 |
+
},
|
1041 |
+
{
|
1042 |
+
"name": "NVIDIA H100 80GB HBM3",
|
1043 |
+
"memory_total": 85520809984
|
1044 |
+
},
|
1045 |
+
{
|
1046 |
+
"name": "NVIDIA H100 80GB HBM3",
|
1047 |
+
"memory_total": 85520809984
|
1048 |
+
},
|
1049 |
+
{
|
1050 |
+
"name": "NVIDIA H100 80GB HBM3",
|
1051 |
+
"memory_total": 85520809984
|
1052 |
+
},
|
1053 |
+
{
|
1054 |
+
"name": "NVIDIA H100 80GB HBM3",
|
1055 |
+
"memory_total": 85520809984
|
1056 |
+
},
|
1057 |
+
{
|
1058 |
+
"name": "NVIDIA H100 80GB HBM3",
|
1059 |
+
"memory_total": 85520809984
|
1060 |
+
},
|
1061 |
+
{
|
1062 |
+
"name": "NVIDIA H100 80GB HBM3",
|
1063 |
+
"memory_total": 85520809984
|
1064 |
+
}
|
1065 |
+
],
|
1066 |
+
"memory": {
|
1067 |
+
"total": 1999.9661331176758
|
1068 |
+
}
|
1069 |
+
}
|
wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/wandb-summary.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"train/loss": 0.856, "train/grad_norm": 2.515625, "train/learning_rate": 3.086582838174551e-07, "train/epoch": 1.22, "train/global_step": 15, "_timestamp": 1731525055.2001407, "_runtime": 488.7689197063446, "_step": 6, "eval/loss": 0.7827465534210205, "eval/runtime": 1.6563, "eval/samples_per_second": 61.584, "eval/steps_per_second": 1.208}
|
wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/logs/debug-internal.log
ADDED
@@ -0,0 +1,373 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
2024-11-13 19:02:46,431 INFO StreamThr :1939 [internal.py:wandb_internal():86] W&B internal server running at pid: 1939, started at: 2024-11-13 19:02:46.431351
|
2 |
+
2024-11-13 19:02:46,433 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status
|
3 |
+
2024-11-13 19:02:46,434 INFO WriterThread:1939 [datastore.py:open_for_write():87] open: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/run-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1.wandb
|
4 |
+
2024-11-13 19:02:46,435 DEBUG SenderThread:1939 [sender.py:send():382] send: header
|
5 |
+
2024-11-13 19:02:46,442 DEBUG SenderThread:1939 [sender.py:send():382] send: run
|
6 |
+
2024-11-13 19:02:46,772 INFO SenderThread:1939 [dir_watcher.py:__init__():211] watching files in: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files
|
7 |
+
2024-11-13 19:02:46,772 INFO SenderThread:1939 [sender.py:_start_run_threads():1136] run started: kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1 with start time 1731524566.431221
|
8 |
+
2024-11-13 19:02:46,782 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: check_version
|
9 |
+
2024-11-13 19:02:46,782 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: check_version
|
10 |
+
2024-11-13 19:02:46,859 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: run_start
|
11 |
+
2024-11-13 19:02:46,947 DEBUG HandlerThread:1939 [system_info.py:__init__():27] System info init
|
12 |
+
2024-11-13 19:02:46,947 DEBUG HandlerThread:1939 [system_info.py:__init__():42] System info init done
|
13 |
+
2024-11-13 19:02:46,947 INFO HandlerThread:1939 [system_monitor.py:start():194] Starting system monitor
|
14 |
+
2024-11-13 19:02:46,947 INFO SystemMonitor:1939 [system_monitor.py:_start():158] Starting system asset monitoring threads
|
15 |
+
2024-11-13 19:02:46,948 INFO HandlerThread:1939 [system_monitor.py:probe():214] Collecting system info
|
16 |
+
2024-11-13 19:02:46,948 INFO SystemMonitor:1939 [interfaces.py:start():190] Started cpu monitoring
|
17 |
+
2024-11-13 19:02:46,949 INFO SystemMonitor:1939 [interfaces.py:start():190] Started disk monitoring
|
18 |
+
2024-11-13 19:02:46,949 INFO SystemMonitor:1939 [interfaces.py:start():190] Started gpu monitoring
|
19 |
+
2024-11-13 19:02:46,951 INFO SystemMonitor:1939 [interfaces.py:start():190] Started memory monitoring
|
20 |
+
2024-11-13 19:02:46,952 INFO SystemMonitor:1939 [interfaces.py:start():190] Started network monitoring
|
21 |
+
2024-11-13 19:02:47,010 DEBUG HandlerThread:1939 [system_info.py:probe():151] Probing system
|
22 |
+
2024-11-13 19:02:47,013 DEBUG HandlerThread:1939 [system_info.py:_probe_git():136] Probing git
|
23 |
+
2024-11-13 19:02:47,021 DEBUG HandlerThread:1939 [system_info.py:_probe_git():144] Probing git done
|
24 |
+
2024-11-13 19:02:47,021 DEBUG HandlerThread:1939 [system_info.py:probe():199] Probing system done
|
25 |
+
2024-11-13 19:02:47,022 DEBUG HandlerThread:1939 [system_monitor.py:probe():223] {'os': 'Linux-5.10.226-214.879.amzn2.x86_64-x86_64-with-glibc2.31', 'python': '3.10.12', 'heartbeatAt': '2024-11-13T19:02:47.010671', 'startedAt': '2024-11-13T19:02:46.426230', 'docker': None, 'cuda': None, 'args': ('--batch_size', '64', '--cache_dir', '/opt/ml/data/input/.cache', '--dataset', 'code-contests', '--gradient_accumulation_steps', '8', '--logging_steps', '5', '--lr', '1e-6', '--max_seq_length', '2048', '--model_name', 'Qwen/Qwen2.5-Coder-7B-Instruct', '--num_train_epochs', '2', '--output_dir', '/opt/ml/model/', '--push_to_hub', 'True', '--run_name', 'sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048', '--save_steps', '100', '--schedule', 'cosine', '--wandb_project', 'sft-codecontests-1112', '--weight_decay', '0.0'), 'state': 'running', 'program': '/opt/ml/code/qlearning_reasoning/training/sft.py', 'codePathLocal': 'qlearning_reasoning/training/sft.py', 'codePath': 'qlearning_reasoning/training/sft.py', 'git': {'remote': 'git@github.com:TRI-ML/reasoning-value-verifiers.git', 'commit': 'cbfd05147f15c55be83f4996eebdb305c17f119d'}, 'email': None, 'root': '/opt/ml/code', 'host': 'algo-1', 'username': 'root', 'executable': '/opt/conda/bin/python3.10', 'cpu_count': 96, 'cpu_count_logical': 192, 'cpu_freq': {'current': 2829.910838541667, 'min': 0.0, 'max': 0.0}, 'cpu_freq_per_core': [{'current': 2850.788, 'min': 0.0, 'max': 0.0}, {'current': 3406.159, 'min': 0.0, 'max': 0.0}, {'current': 2605.944, 'min': 0.0, 'max': 0.0}, {'current': 2608.162, 'min': 0.0, 'max': 0.0}, {'current': 2818.169, 'min': 0.0, 'max': 0.0}, {'current': 2591.551, 'min': 0.0, 'max': 0.0}, {'current': 2621.486, 'min': 0.0, 'max': 0.0}, {'current': 2607.504, 'min': 0.0, 'max': 0.0}, {'current': 2999.734, 'min': 0.0, 'max': 0.0}, {'current': 2991.449, 'min': 0.0, 'max': 0.0}, {'current': 3123.177, 'min': 0.0, 'max': 0.0}, {'current': 3053.641, 'min': 0.0, 'max': 0.0}, {'current': 3003.014, 'min': 0.0, 'max': 0.0}, {'current': 3000.178, 'min': 0.0, 'max': 0.0}, {'current': 3432.256, 'min': 0.0, 'max': 0.0}, {'current': 2999.802, 'min': 0.0, 'max': 0.0}, {'current': 3005.725, 'min': 0.0, 'max': 0.0}, {'current': 3565.885, 'min': 0.0, 'max': 0.0}, {'current': 2992.339, 'min': 0.0, 'max': 0.0}, {'current': 3104.468, 'min': 0.0, 'max': 0.0}, {'current': 3182.411, 'min': 0.0, 'max': 0.0}, {'current': 3013.217, 'min': 0.0, 'max': 0.0}, {'current': 3131.904, 'min': 0.0, 'max': 0.0}, {'current': 3007.314, 'min': 0.0, 'max': 0.0}, {'current': 2625.175, 'min': 0.0, 'max': 0.0}, {'current': 2653.407, 'min': 0.0, 'max': 0.0}, {'current': 2610.871, 'min': 0.0, 'max': 0.0}, {'current': 2793.855, 'min': 0.0, 'max': 0.0}, {'current': 2619.232, 'min': 0.0, 'max': 0.0}, {'current': 2627.886, 'min': 0.0, 'max': 0.0}, {'current': 2613.748, 'min': 0.0, 'max': 0.0}, {'current': 2318.039, 'min': 0.0, 'max': 0.0}, {'current': 3546.661, 'min': 0.0, 'max': 0.0}, {'current': 3022.56, 'min': 0.0, 'max': 0.0}, {'current': 3035.28, 'min': 0.0, 'max': 0.0}, {'current': 3541.443, 'min': 0.0, 'max': 0.0}, {'current': 3163.55, 'min': 0.0, 'max': 0.0}, {'current': 3023.79, 'min': 0.0, 'max': 0.0}, {'current': 3033.204, 'min': 0.0, 'max': 0.0}, {'current': 3031.695, 'min': 0.0, 'max': 0.0}, {'current': 3034.587, 'min': 0.0, 'max': 0.0}, {'current': 3031.71, 'min': 0.0, 'max': 0.0}, {'current': 3029.116, 'min': 0.0, 'max': 0.0}, {'current': 3293.701, 'min': 0.0, 'max': 0.0}, {'current': 3033.78, 'min': 0.0, 'max': 0.0}, {'current': 3032.034, 'min': 0.0, 'max': 0.0}, {'current': 3057.365, 'min': 0.0, 'max': 0.0}, {'current': 3028.183, 'min': 0.0, 'max': 0.0}, {'current': 3596.439, 'min': 0.0, 'max': 0.0}, {'current': 3327.853, 'min': 0.0, 'max': 0.0}, {'current': 3226.463, 'min': 0.0, 'max': 0.0}, {'current': 3259.361, 'min': 0.0, 'max': 0.0}, {'current': 3284.076, 'min': 0.0, 'max': 0.0}, {'current': 3257.698, 'min': 0.0, 'max': 0.0}, {'current': 3280.289, 'min': 0.0, 'max': 0.0}, {'current': 3258.834, 'min': 0.0, 'max': 0.0}, {'current': 2517.346, 'min': 0.0, 'max': 0.0}, {'current': 2553.091, 'min': 0.0, 'max': 0.0}, {'current': 2552.693, 'min': 0.0, 'max': 0.0}, {'current': 2689.091, 'min': 0.0, 'max': 0.0}, {'current': 2503.301, 'min': 0.0, 'max': 0.0}, {'current': 2524.4, 'min': 0.0, 'max': 0.0}, {'current': 2521.768, 'min': 0.0, 'max': 0.0}, {'current': 2530.364, 'min': 0.0, 'max': 0.0}, {'current': 3258.225, 'min': 0.0, 'max': 0.0}, {'current': 3597.406, 'min': 0.0, 'max': 0.0}, {'current': 3279.658, 'min': 0.0, 'max': 0.0}, {'current': 3292.257, 'min': 0.0, 'max': 0.0}, {'current': 3265.351, 'min': 0.0, 'max': 0.0}, {'current': 3286.715, 'min': 0.0, 'max': 0.0}, {'current': 3268.777, 'min': 0.0, 'max': 0.0}, {'current': 3242.068, 'min': 0.0, 'max': 0.0}, {'current': 3288.908, 'min': 0.0, 'max': 0.0}, {'current': 3278.645, 'min': 0.0, 'max': 0.0}, {'current': 3215.578, 'min': 0.0, 'max': 0.0}, {'current': 3327.101, 'min': 0.0, 'max': 0.0}, {'current': 3357.075, 'min': 0.0, 'max': 0.0}, {'current': 3144.838, 'min': 0.0, 'max': 0.0}, {'current': 3005.517, 'min': 0.0, 'max': 0.0}, {'current': 3597.381, 'min': 0.0, 'max': 0.0}, {'current': 3596.175, 'min': 0.0, 'max': 0.0}, {'current': 3317.258, 'min': 0.0, 'max': 0.0}, {'current': 3310.249, 'min': 0.0, 'max': 0.0}, {'current': 3140.687, 'min': 0.0, 'max': 0.0}, {'current': 3398.954, 'min': 0.0, 'max': 0.0}, {'current': 3318.433, 'min': 0.0, 'max': 0.0}, {'current': 3342.35, 'min': 0.0, 'max': 0.0}, {'current': 3328.575, 'min': 0.0, 'max': 0.0}, {'current': 3499.892, 'min': 0.0, 'max': 0.0}, {'current': 3344.864, 'min': 0.0, 'max': 0.0}, {'current': 3355.802, 'min': 0.0, 'max': 0.0}, {'current': 3305.032, 'min': 0.0, 'max': 0.0}, {'current': 3318.07, 'min': 0.0, 'max': 0.0}, {'current': 3336.546, 'min': 0.0, 'max': 0.0}, {'current': 3335.458, 'min': 0.0, 'max': 0.0}, {'current': 3596.561, 'min': 0.0, 'max': 0.0}, {'current': 2884.473, 'min': 0.0, 'max': 0.0}, {'current': 2934.098, 'min': 0.0, 'max': 0.0}, {'current': 2621.598, 'min': 0.0, 'max': 0.0}, {'current': 2645.184, 'min': 0.0, 'max': 0.0}, {'current': 2480.726, 'min': 0.0, 'max': 0.0}, {'current': 2515.912, 'min': 0.0, 'max': 0.0}, {'current': 2702.57, 'min': 0.0, 'max': 0.0}, {'current': 2671.473, 'min': 0.0, 'max': 0.0}, {'current': 3092.321, 'min': 0.0, 'max': 0.0}, {'current': 3092.204, 'min': 0.0, 'max': 0.0}, {'current': 3217.287, 'min': 0.0, 'max': 0.0}, {'current': 3161.646, 'min': 0.0, 'max': 0.0}, {'current': 3089.273, 'min': 0.0, 'max': 0.0}, {'current': 3088.447, 'min': 0.0, 'max': 0.0}, {'current': 3559.214, 'min': 0.0, 'max': 0.0}, {'current': 3218.611, 'min': 0.0, 'max': 0.0}, {'current': 3021.622, 'min': 0.0, 'max': 0.0}, {'current': 3332.084, 'min': 0.0, 'max': 0.0}, {'current': 3019.854, 'min': 0.0, 'max': 0.0}, {'current': 3213.831, 'min': 0.0, 'max': 0.0}, {'current': 3092.774, 'min': 0.0, 'max': 0.0}, {'current': 3015.868, 'min': 0.0, 'max': 0.0}, {'current': 3082.752, 'min': 0.0, 'max': 0.0}, {'current': 3029.559, 'min': 0.0, 'max': 0.0}, {'current': 2873.142, 'min': 0.0, 'max': 0.0}, {'current': 2215.288, 'min': 0.0, 'max': 0.0}, {'current': 3193.443, 'min': 0.0, 'max': 0.0}, {'current': 2928.717, 'min': 0.0, 'max': 0.0}, {'current': 2990.366, 'min': 0.0, 'max': 0.0}, {'current': 2923.847, 'min': 0.0, 'max': 0.0}, {'current': 2869.364, 'min': 0.0, 'max': 0.0}, {'current': 2859.189, 'min': 0.0, 'max': 0.0}, {'current': 3049.001, 'min': 0.0, 'max': 0.0}, {'current': 2805.206, 'min': 0.0, 'max': 0.0}, {'current': 3094.081, 'min': 0.0, 'max': 0.0}, {'current': 3187.222, 'min': 0.0, 'max': 0.0}, {'current': 2853.945, 'min': 0.0, 'max': 0.0}, {'current': 2944.73, 'min': 0.0, 'max': 0.0}, {'current': 2809.161, 'min': 0.0, 'max': 0.0}, {'current': 2806.787, 'min': 0.0, 'max': 0.0}, {'current': 3024.225, 'min': 0.0, 'max': 0.0}, {'current': 3032.173, 'min': 0.0, 'max': 0.0}, {'current': 3033.195, 'min': 0.0, 'max': 0.0}, {'current': 3561.287, 'min': 0.0, 'max': 0.0}, {'current': 2987.872, 'min': 0.0, 'max': 0.0}, {'current': 3051.252, 'min': 0.0, 'max': 0.0}, {'current': 3379.017, 'min': 0.0, 'max': 0.0}, {'current': 3044.319, 'min': 0.0, 'max': 0.0}, {'current': 3595.62, 'min': 0.0, 'max': 0.0}, {'current': 3261.737, 'min': 0.0, 'max': 0.0}, {'current': 3257.013, 'min': 0.0, 'max': 0.0}, {'current': 3288.317, 'min': 0.0, 'max': 0.0}, {'current': 3268.675, 'min': 0.0, 'max': 0.0}, {'current': 3269.579, 'min': 0.0, 'max': 0.0}, {'current': 3266.213, 'min': 0.0, 'max': 0.0}, {'current': 3265.148, 'min': 0.0, 'max': 0.0}, {'current': 2554.152, 'min': 0.0, 'max': 0.0}, {'current': 2556.63, 'min': 0.0, 'max': 0.0}, {'current': 2540.571, 'min': 0.0, 'max': 0.0}, {'current': 2512.005, 'min': 0.0, 'max': 0.0}, {'current': 2560.992, 'min': 0.0, 'max': 0.0}, {'current': 2526.641, 'min': 0.0, 'max': 0.0}, {'current': 2528.267, 'min': 0.0, 'max': 0.0}, {'current': 2583.061, 'min': 0.0, 'max': 0.0}, {'current': 3257.652, 'min': 0.0, 'max': 0.0}, {'current': 3598.217, 'min': 0.0, 'max': 0.0}, {'current': 3282.578, 'min': 0.0, 'max': 0.0}, {'current': 3261.664, 'min': 0.0, 'max': 0.0}, {'current': 3268.246, 'min': 0.0, 'max': 0.0}, {'current': 3331.937, 'min': 0.0, 'max': 0.0}, {'current': 3265.603, 'min': 0.0, 'max': 0.0}, {'current': 3258.501, 'min': 0.0, 'max': 0.0}, {'current': 3337.045, 'min': 0.0, 'max': 0.0}, {'current': 3259.949, 'min': 0.0, 'max': 0.0}, {'current': 3226.058, 'min': 0.0, 'max': 0.0}, {'current': 3250.308, 'min': 0.0, 'max': 0.0}, {'current': 3208.026, 'min': 0.0, 'max': 0.0}, {'current': 2974.043, 'min': 0.0, 'max': 0.0}, {'current': 3216.706, 'min': 0.0, 'max': 0.0}, {'current': 3598.55, 'min': 0.0, 'max': 0.0}, {'current': 3598.261, 'min': 0.0, 'max': 0.0}, {'current': 3222.637, 'min': 0.0, 'max': 0.0}, {'current': 3227.662, 'min': 0.0, 'max': 0.0}, {'current': 3143.781, 'min': 0.0, 'max': 0.0}, {'current': 3215.567, 'min': 0.0, 'max': 0.0}, {'current': 3250.612, 'min': 0.0, 'max': 0.0}, {'current': 3195.784, 'min': 0.0, 'max': 0.0}, {'current': 3219.289, 'min': 0.0, 'max': 0.0}, {'current': 3289.163, 'min': 0.0, 'max': 0.0}, {'current': 3247.875, 'min': 0.0, 'max': 0.0}, {'current': 3249.112, 'min': 0.0, 'max': 0.0}, {'current': 3209.3, 'min': 0.0, 'max': 0.0}, {'current': 3187.212, 'min': 0.0, 'max': 0.0}, {'current': 3288.003, 'min': 0.0, 'max': 0.0}, {'current': 3169.417, 'min': 0.0, 'max': 0.0}, {'current': 3596.071, 'min': 0.0, 'max': 0.0}], 'disk': {'/': {'total': 119.94140625, 'used': 56.4015998840332}}, 'gpu': 'NVIDIA H100 80GB HBM3', 'gpu_count': 8, 'gpu_devices': [{'name': 'NVIDIA H100 80GB HBM3', 'memory_total': 85520809984}, {'name': 'NVIDIA H100 80GB HBM3', 'memory_total': 85520809984}, {'name': 'NVIDIA H100 80GB HBM3', 'memory_total': 85520809984}, {'name': 'NVIDIA H100 80GB HBM3', 'memory_total': 85520809984}, {'name': 'NVIDIA H100 80GB HBM3', 'memory_total': 85520809984}, {'name': 'NVIDIA H100 80GB HBM3', 'memory_total': 85520809984}, {'name': 'NVIDIA H100 80GB HBM3', 'memory_total': 85520809984}, {'name': 'NVIDIA H100 80GB HBM3', 'memory_total': 85520809984}], 'memory': {'total': 1999.9661331176758}}
|
26 |
+
2024-11-13 19:02:47,022 INFO HandlerThread:1939 [system_monitor.py:probe():224] Finished collecting system info
|
27 |
+
2024-11-13 19:02:47,022 INFO HandlerThread:1939 [system_monitor.py:probe():227] Publishing system info
|
28 |
+
2024-11-13 19:02:47,022 DEBUG HandlerThread:1939 [system_info.py:_save_conda():208] Saving list of conda packages installed into the current environment
|
29 |
+
2024-11-13 19:02:47,775 INFO Thread-12 :1939 [dir_watcher.py:_on_file_created():271] file/dir created: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/conda-environment.yaml
|
30 |
+
2024-11-13 19:03:01,345 DEBUG HandlerThread:1939 [system_info.py:_save_conda():220] Saving conda packages done
|
31 |
+
2024-11-13 19:03:01,345 DEBUG HandlerThread:1939 [system_info.py:_save_code():45] Saving code
|
32 |
+
2024-11-13 19:03:01,349 DEBUG HandlerThread:1939 [system_info.py:_save_code():66] Saving code done
|
33 |
+
2024-11-13 19:03:01,349 DEBUG HandlerThread:1939 [system_info.py:_save_patches():83] Saving git patches
|
34 |
+
2024-11-13 19:03:01,776 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/conda-environment.yaml
|
35 |
+
2024-11-13 19:03:01,776 INFO Thread-12 :1939 [dir_watcher.py:_on_file_created():271] file/dir created: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/code/qlearning_reasoning/training/sft.py
|
36 |
+
2024-11-13 19:03:01,776 INFO Thread-12 :1939 [dir_watcher.py:_on_file_created():271] file/dir created: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/code/qlearning_reasoning
|
37 |
+
2024-11-13 19:03:01,776 INFO Thread-12 :1939 [dir_watcher.py:_on_file_created():271] file/dir created: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/code/qlearning_reasoning/training
|
38 |
+
2024-11-13 19:03:01,776 INFO Thread-12 :1939 [dir_watcher.py:_on_file_created():271] file/dir created: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/code
|
39 |
+
2024-11-13 19:03:01,989 DEBUG HandlerThread:1939 [system_info.py:_save_patches():125] Saving git patches done
|
40 |
+
2024-11-13 19:03:01,991 INFO HandlerThread:1939 [system_monitor.py:probe():229] Finished publishing system info
|
41 |
+
2024-11-13 19:03:01,993 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
42 |
+
2024-11-13 19:03:01,993 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: keepalive
|
43 |
+
2024-11-13 19:03:01,994 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
44 |
+
2024-11-13 19:03:01,994 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: keepalive
|
45 |
+
2024-11-13 19:03:01,994 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
46 |
+
2024-11-13 19:03:01,994 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: keepalive
|
47 |
+
2024-11-13 19:03:01,994 DEBUG SenderThread:1939 [sender.py:send():382] send: files
|
48 |
+
2024-11-13 19:03:01,995 INFO SenderThread:1939 [sender.py:_save_file():1403] saving file wandb-metadata.json with policy now
|
49 |
+
2024-11-13 19:03:01,995 INFO SenderThread:1939 [sender.py:_save_file():1403] saving file code/qlearning_reasoning/training/sft.py with policy now
|
50 |
+
2024-11-13 19:03:01,995 INFO SenderThread:1939 [sender.py:_save_file():1403] saving file diff.patch with policy now
|
51 |
+
2024-11-13 19:03:01,999 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: python_packages
|
52 |
+
2024-11-13 19:03:01,999 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: python_packages
|
53 |
+
2024-11-13 19:03:02,001 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
54 |
+
2024-11-13 19:03:02,003 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
55 |
+
2024-11-13 19:03:02,003 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
56 |
+
2024-11-13 19:03:02,111 DEBUG SenderThread:1939 [sender.py:send():382] send: telemetry
|
57 |
+
2024-11-13 19:03:02,111 DEBUG SenderThread:1939 [sender.py:send():382] send: config
|
58 |
+
2024-11-13 19:03:02,112 DEBUG SenderThread:1939 [sender.py:send():382] send: metric
|
59 |
+
2024-11-13 19:03:02,112 DEBUG SenderThread:1939 [sender.py:send():382] send: telemetry
|
60 |
+
2024-11-13 19:03:02,112 DEBUG SenderThread:1939 [sender.py:send():382] send: metric
|
61 |
+
2024-11-13 19:03:02,113 WARNING SenderThread:1939 [sender.py:send_metric():1354] Seen metric with glob (shouldn't happen)
|
62 |
+
2024-11-13 19:03:02,113 DEBUG SenderThread:1939 [sender.py:send():382] send: telemetry
|
63 |
+
2024-11-13 19:03:02,113 DEBUG SenderThread:1939 [sender.py:send():382] send: telemetry
|
64 |
+
2024-11-13 19:03:02,113 DEBUG SenderThread:1939 [sender.py:send():382] send: config
|
65 |
+
2024-11-13 19:03:02,551 INFO wandb-upload_1:1939 [upload_job.py:push():131] Uploaded file /tmp/tmpm_h7h9v0wandb/pjhigh2e-code/qlearning_reasoning/training/sft.py
|
66 |
+
2024-11-13 19:03:02,554 INFO wandb-upload_0:1939 [upload_job.py:push():131] Uploaded file /tmp/tmpm_h7h9v0wandb/f98paeu8-wandb-metadata.json
|
67 |
+
2024-11-13 19:03:02,556 INFO wandb-upload_2:1939 [upload_job.py:push():131] Uploaded file /tmp/tmpm_h7h9v0wandb/cc8zuviy-diff.patch
|
68 |
+
2024-11-13 19:03:02,776 INFO Thread-12 :1939 [dir_watcher.py:_on_file_created():271] file/dir created: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/diff.patch
|
69 |
+
2024-11-13 19:03:02,776 INFO Thread-12 :1939 [dir_watcher.py:_on_file_created():271] file/dir created: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
70 |
+
2024-11-13 19:03:02,777 INFO Thread-12 :1939 [dir_watcher.py:_on_file_created():271] file/dir created: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/requirements.txt
|
71 |
+
2024-11-13 19:03:02,777 INFO Thread-12 :1939 [dir_watcher.py:_on_file_created():271] file/dir created: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/wandb-metadata.json
|
72 |
+
2024-11-13 19:03:04,777 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
73 |
+
2024-11-13 19:03:07,114 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
74 |
+
2024-11-13 19:03:12,115 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
75 |
+
2024-11-13 19:03:16,999 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
76 |
+
2024-11-13 19:03:17,000 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
77 |
+
2024-11-13 19:03:17,002 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
78 |
+
2024-11-13 19:03:18,090 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
79 |
+
2024-11-13 19:03:18,782 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/config.yaml
|
80 |
+
2024-11-13 19:03:23,250 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
81 |
+
2024-11-13 19:03:26,095 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: partial_history
|
82 |
+
2024-11-13 19:03:26,097 DEBUG SenderThread:1939 [sender.py:send():382] send: metric
|
83 |
+
2024-11-13 19:03:26,097 DEBUG SenderThread:1939 [sender.py:send():382] send: metric
|
84 |
+
2024-11-13 19:03:26,097 DEBUG SenderThread:1939 [sender.py:send():382] send: metric
|
85 |
+
2024-11-13 19:03:26,097 DEBUG SenderThread:1939 [sender.py:send():382] send: metric
|
86 |
+
2024-11-13 19:03:26,097 DEBUG SenderThread:1939 [sender.py:send():382] send: history
|
87 |
+
2024-11-13 19:03:26,098 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: summary_record
|
88 |
+
2024-11-13 19:03:26,099 INFO SenderThread:1939 [sender.py:_save_file():1403] saving file wandb-summary.json with policy end
|
89 |
+
2024-11-13 19:03:26,785 INFO Thread-12 :1939 [dir_watcher.py:_on_file_created():271] file/dir created: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/wandb-summary.json
|
90 |
+
2024-11-13 19:03:28,785 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
91 |
+
2024-11-13 19:03:29,100 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
92 |
+
2024-11-13 19:03:31,999 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
93 |
+
2024-11-13 19:03:32,000 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
94 |
+
2024-11-13 19:03:32,002 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
95 |
+
2024-11-13 19:03:34,147 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
96 |
+
2024-11-13 19:03:39,147 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
97 |
+
2024-11-13 19:03:44,148 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
98 |
+
2024-11-13 19:03:46,955 DEBUG SystemMonitor:1939 [system_monitor.py:_start():172] Starting system metrics aggregation loop
|
99 |
+
2024-11-13 19:03:46,959 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
|
100 |
+
2024-11-13 19:03:47,000 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
101 |
+
2024-11-13 19:03:47,000 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
102 |
+
2024-11-13 19:03:47,043 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
103 |
+
2024-11-13 19:03:49,636 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
104 |
+
2024-11-13 19:03:50,813 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
105 |
+
2024-11-13 19:03:50,814 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/config.yaml
|
106 |
+
2024-11-13 19:03:54,814 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
107 |
+
2024-11-13 19:03:59,814 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
108 |
+
2024-11-13 19:04:02,000 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
109 |
+
2024-11-13 19:04:02,000 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
110 |
+
2024-11-13 19:04:02,002 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
111 |
+
2024-11-13 19:04:05,061 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
112 |
+
2024-11-13 19:04:10,062 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
113 |
+
2024-11-13 19:04:14,822 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
114 |
+
2024-11-13 19:04:15,419 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
115 |
+
2024-11-13 19:04:16,962 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
|
116 |
+
2024-11-13 19:04:17,000 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
117 |
+
2024-11-13 19:04:17,000 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
118 |
+
2024-11-13 19:04:17,002 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
119 |
+
2024-11-13 19:04:21,087 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
120 |
+
2024-11-13 19:04:26,088 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
121 |
+
2024-11-13 19:04:31,089 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
122 |
+
2024-11-13 19:04:32,000 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
123 |
+
2024-11-13 19:04:32,000 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
124 |
+
2024-11-13 19:04:32,043 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
125 |
+
2024-11-13 19:04:36,587 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
126 |
+
2024-11-13 19:04:36,829 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
127 |
+
2024-11-13 19:04:41,588 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
128 |
+
2024-11-13 19:04:46,588 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
129 |
+
2024-11-13 19:04:46,965 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
|
130 |
+
2024-11-13 19:04:47,000 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
131 |
+
2024-11-13 19:04:47,000 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
132 |
+
2024-11-13 19:04:47,002 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
133 |
+
2024-11-13 19:04:52,111 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
134 |
+
2024-11-13 19:04:57,111 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
135 |
+
2024-11-13 19:04:57,383 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: partial_history
|
136 |
+
2024-11-13 19:04:57,384 DEBUG SenderThread:1939 [sender.py:send():382] send: history
|
137 |
+
2024-11-13 19:04:57,385 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: summary_record
|
138 |
+
2024-11-13 19:04:57,385 INFO SenderThread:1939 [sender.py:_save_file():1403] saving file wandb-summary.json with policy end
|
139 |
+
2024-11-13 19:04:57,837 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/wandb-summary.json
|
140 |
+
2024-11-13 19:04:58,837 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
141 |
+
2024-11-13 19:04:59,002 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: partial_history
|
142 |
+
2024-11-13 19:04:59,004 DEBUG SenderThread:1939 [sender.py:send():382] send: metric
|
143 |
+
2024-11-13 19:04:59,005 DEBUG SenderThread:1939 [sender.py:send():382] send: metric
|
144 |
+
2024-11-13 19:04:59,005 DEBUG SenderThread:1939 [sender.py:send():382] send: metric
|
145 |
+
2024-11-13 19:04:59,005 DEBUG SenderThread:1939 [sender.py:send():382] send: metric
|
146 |
+
2024-11-13 19:04:59,006 DEBUG SenderThread:1939 [sender.py:send():382] send: history
|
147 |
+
2024-11-13 19:04:59,006 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: summary_record
|
148 |
+
2024-11-13 19:04:59,006 INFO SenderThread:1939 [sender.py:_save_file():1403] saving file wandb-summary.json with policy end
|
149 |
+
2024-11-13 19:04:59,838 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/wandb-summary.json
|
150 |
+
2024-11-13 19:05:00,838 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
151 |
+
2024-11-13 19:05:02,000 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
152 |
+
2024-11-13 19:05:02,001 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
153 |
+
2024-11-13 19:05:02,002 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
154 |
+
2024-11-13 19:05:03,111 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
155 |
+
2024-11-13 19:05:08,112 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
156 |
+
2024-11-13 19:05:13,113 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
157 |
+
2024-11-13 19:05:16,967 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
|
158 |
+
2024-11-13 19:05:17,000 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
159 |
+
2024-11-13 19:05:17,001 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
160 |
+
2024-11-13 19:05:17,043 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
161 |
+
2024-11-13 19:05:18,123 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
162 |
+
2024-11-13 19:05:22,846 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
163 |
+
2024-11-13 19:05:23,982 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
164 |
+
2024-11-13 19:05:25,847 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/config.yaml
|
165 |
+
2024-11-13 19:05:30,486 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
166 |
+
2024-11-13 19:05:32,000 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
167 |
+
2024-11-13 19:05:32,001 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
168 |
+
2024-11-13 19:05:32,002 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
169 |
+
2024-11-13 19:05:36,117 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
170 |
+
2024-11-13 19:05:41,118 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
171 |
+
2024-11-13 19:05:46,513 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
172 |
+
2024-11-13 19:05:46,854 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
173 |
+
2024-11-13 19:05:46,970 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
|
174 |
+
2024-11-13 19:05:47,000 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
175 |
+
2024-11-13 19:05:47,001 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
176 |
+
2024-11-13 19:05:47,003 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
177 |
+
2024-11-13 19:05:52,097 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
178 |
+
2024-11-13 19:05:57,098 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
179 |
+
2024-11-13 19:06:02,001 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
180 |
+
2024-11-13 19:06:02,001 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
181 |
+
2024-11-13 19:06:02,043 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
182 |
+
2024-11-13 19:06:02,152 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
183 |
+
2024-11-13 19:06:07,153 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
184 |
+
2024-11-13 19:06:08,862 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
185 |
+
2024-11-13 19:06:12,639 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
186 |
+
2024-11-13 19:06:16,973 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
|
187 |
+
2024-11-13 19:06:17,001 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
188 |
+
2024-11-13 19:06:17,002 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
189 |
+
2024-11-13 19:06:17,004 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
190 |
+
2024-11-13 19:06:18,127 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
191 |
+
2024-11-13 19:06:23,127 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
192 |
+
2024-11-13 19:06:28,128 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
193 |
+
2024-11-13 19:06:30,870 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
194 |
+
2024-11-13 19:06:32,001 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
195 |
+
2024-11-13 19:06:32,001 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
196 |
+
2024-11-13 19:06:32,003 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
197 |
+
2024-11-13 19:06:33,143 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
198 |
+
2024-11-13 19:06:38,144 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
199 |
+
2024-11-13 19:06:43,144 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
200 |
+
2024-11-13 19:06:46,975 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
|
201 |
+
2024-11-13 19:06:47,001 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
202 |
+
2024-11-13 19:06:47,001 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
203 |
+
2024-11-13 19:06:47,043 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
204 |
+
2024-11-13 19:06:49,137 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
205 |
+
2024-11-13 19:06:51,931 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: partial_history
|
206 |
+
2024-11-13 19:06:51,932 DEBUG SenderThread:1939 [sender.py:send():382] send: history
|
207 |
+
2024-11-13 19:06:51,933 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: summary_record
|
208 |
+
2024-11-13 19:06:51,934 INFO SenderThread:1939 [sender.py:_save_file():1403] saving file wandb-summary.json with policy end
|
209 |
+
2024-11-13 19:06:52,878 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
210 |
+
2024-11-13 19:06:52,878 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/wandb-summary.json
|
211 |
+
2024-11-13 19:06:53,541 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: partial_history
|
212 |
+
2024-11-13 19:06:53,543 DEBUG SenderThread:1939 [sender.py:send():382] send: history
|
213 |
+
2024-11-13 19:06:53,543 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: summary_record
|
214 |
+
2024-11-13 19:06:53,544 INFO SenderThread:1939 [sender.py:_save_file():1403] saving file wandb-summary.json with policy end
|
215 |
+
2024-11-13 19:06:53,878 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/wandb-summary.json
|
216 |
+
2024-11-13 19:06:54,545 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
217 |
+
2024-11-13 19:06:54,878 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
218 |
+
2024-11-13 19:06:59,546 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
219 |
+
2024-11-13 19:07:02,001 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
220 |
+
2024-11-13 19:07:02,002 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
221 |
+
2024-11-13 19:07:02,004 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
222 |
+
2024-11-13 19:07:05,164 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
223 |
+
2024-11-13 19:07:10,165 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
224 |
+
2024-11-13 19:07:15,165 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
225 |
+
2024-11-13 19:07:16,886 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
226 |
+
2024-11-13 19:07:16,978 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
|
227 |
+
2024-11-13 19:07:17,001 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
228 |
+
2024-11-13 19:07:17,002 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
229 |
+
2024-11-13 19:07:17,003 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
230 |
+
2024-11-13 19:07:21,108 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
231 |
+
2024-11-13 19:07:26,108 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
232 |
+
2024-11-13 19:07:31,109 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
233 |
+
2024-11-13 19:07:32,001 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
234 |
+
2024-11-13 19:07:32,002 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
235 |
+
2024-11-13 19:07:32,043 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
236 |
+
2024-11-13 19:07:36,934 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
237 |
+
2024-11-13 19:07:38,894 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
238 |
+
2024-11-13 19:07:41,935 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
239 |
+
2024-11-13 19:07:46,935 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
240 |
+
2024-11-13 19:07:46,981 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
|
241 |
+
2024-11-13 19:07:47,002 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
242 |
+
2024-11-13 19:07:47,002 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
243 |
+
2024-11-13 19:07:47,004 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
244 |
+
2024-11-13 19:07:52,135 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
245 |
+
2024-11-13 19:07:57,136 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
246 |
+
2024-11-13 19:08:02,136 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
247 |
+
2024-11-13 19:08:02,856 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
248 |
+
2024-11-13 19:08:02,923 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
249 |
+
2024-11-13 19:08:02,924 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
250 |
+
2024-11-13 19:08:08,076 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
251 |
+
2024-11-13 19:08:13,077 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
252 |
+
2024-11-13 19:08:16,985 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
|
253 |
+
2024-11-13 19:08:17,407 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
254 |
+
2024-11-13 19:08:17,407 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
255 |
+
2024-11-13 19:08:17,447 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
256 |
+
2024-11-13 19:08:18,545 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
257 |
+
2024-11-13 19:08:23,546 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
258 |
+
2024-11-13 19:08:28,546 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
259 |
+
2024-11-13 19:08:32,914 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
260 |
+
2024-11-13 19:08:33,843 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
261 |
+
2024-11-13 19:08:34,729 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
262 |
+
2024-11-13 19:08:34,729 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
263 |
+
2024-11-13 19:08:34,771 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
264 |
+
2024-11-13 19:08:39,802 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
265 |
+
2024-11-13 19:08:44,803 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
266 |
+
2024-11-13 19:08:46,987 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
|
267 |
+
2024-11-13 19:08:48,007 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
268 |
+
2024-11-13 19:08:48,007 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
269 |
+
2024-11-13 19:08:48,008 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
270 |
+
2024-11-13 19:08:48,921 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
271 |
+
2024-11-13 19:08:49,945 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
272 |
+
2024-11-13 19:08:50,922 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
273 |
+
2024-11-13 19:08:52,923 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
274 |
+
2024-11-13 19:08:54,924 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
275 |
+
2024-11-13 19:08:55,154 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
276 |
+
2024-11-13 19:08:56,925 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
277 |
+
2024-11-13 19:08:58,926 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
278 |
+
2024-11-13 19:09:00,371 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
279 |
+
2024-11-13 19:09:00,948 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
280 |
+
2024-11-13 19:09:03,950 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
281 |
+
2024-11-13 19:09:05,541 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
282 |
+
2024-11-13 19:09:05,782 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
283 |
+
2024-11-13 19:09:06,102 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
284 |
+
2024-11-13 19:09:06,102 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
285 |
+
2024-11-13 19:09:11,263 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
286 |
+
2024-11-13 19:09:16,264 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
287 |
+
2024-11-13 19:09:16,990 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
|
288 |
+
2024-11-13 19:09:20,961 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
289 |
+
2024-11-13 19:09:22,073 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
290 |
+
2024-11-13 19:09:22,272 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
291 |
+
2024-11-13 19:09:22,272 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
292 |
+
2024-11-13 19:09:27,401 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
293 |
+
2024-11-13 19:09:32,401 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
294 |
+
2024-11-13 19:09:36,126 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
295 |
+
2024-11-13 19:09:36,666 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
296 |
+
2024-11-13 19:09:36,666 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
297 |
+
2024-11-13 19:09:38,052 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
298 |
+
2024-11-13 19:09:39,964 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
299 |
+
2024-11-13 19:09:43,769 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
300 |
+
2024-11-13 19:09:46,992 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
|
301 |
+
2024-11-13 19:09:47,846 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
302 |
+
2024-11-13 19:09:47,847 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
303 |
+
2024-11-13 19:09:47,847 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
304 |
+
2024-11-13 19:09:49,230 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
305 |
+
2024-11-13 19:09:54,231 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
306 |
+
2024-11-13 19:09:59,231 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
307 |
+
2024-11-13 19:10:02,847 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
308 |
+
2024-11-13 19:10:02,848 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
309 |
+
2024-11-13 19:10:02,848 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
310 |
+
2024-11-13 19:10:05,230 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
311 |
+
2024-11-13 19:10:10,231 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
312 |
+
2024-11-13 19:10:11,977 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
313 |
+
2024-11-13 19:10:15,248 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
314 |
+
2024-11-13 19:10:16,996 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
|
315 |
+
2024-11-13 19:10:17,847 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
316 |
+
2024-11-13 19:10:17,848 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
317 |
+
2024-11-13 19:10:17,848 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
318 |
+
2024-11-13 19:10:21,230 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
319 |
+
2024-11-13 19:10:26,231 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
320 |
+
2024-11-13 19:10:31,231 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
321 |
+
2024-11-13 19:10:32,847 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
322 |
+
2024-11-13 19:10:32,847 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
323 |
+
2024-11-13 19:10:32,847 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
324 |
+
2024-11-13 19:10:33,985 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
325 |
+
2024-11-13 19:10:37,230 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
326 |
+
2024-11-13 19:10:42,231 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
327 |
+
2024-11-13 19:10:46,998 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
|
328 |
+
2024-11-13 19:10:47,847 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
329 |
+
2024-11-13 19:10:47,847 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
330 |
+
2024-11-13 19:10:47,887 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
331 |
+
2024-11-13 19:10:47,921 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
332 |
+
2024-11-13 19:10:53,230 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
333 |
+
2024-11-13 19:10:53,575 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: partial_history
|
334 |
+
2024-11-13 19:10:53,576 DEBUG SenderThread:1939 [sender.py:send():382] send: history
|
335 |
+
2024-11-13 19:10:53,576 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: summary_record
|
336 |
+
2024-11-13 19:10:53,577 INFO SenderThread:1939 [sender.py:_save_file():1403] saving file wandb-summary.json with policy end
|
337 |
+
2024-11-13 19:10:53,992 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/wandb-summary.json
|
338 |
+
2024-11-13 19:10:54,993 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
339 |
+
2024-11-13 19:10:55,200 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: partial_history
|
340 |
+
2024-11-13 19:10:55,201 DEBUG SenderThread:1939 [sender.py:send():382] send: history
|
341 |
+
2024-11-13 19:10:55,201 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: summary_record
|
342 |
+
2024-11-13 19:10:55,203 INFO SenderThread:1939 [sender.py:_save_file():1403] saving file wandb-summary.json with policy end
|
343 |
+
2024-11-13 19:10:55,993 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
344 |
+
2024-11-13 19:10:55,994 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/wandb-summary.json
|
345 |
+
2024-11-13 19:10:56,994 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
346 |
+
2024-11-13 19:10:57,994 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
347 |
+
2024-11-13 19:10:59,204 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
348 |
+
2024-11-13 19:11:02,848 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
349 |
+
2024-11-13 19:11:02,848 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
350 |
+
2024-11-13 19:11:02,849 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
351 |
+
2024-11-13 19:11:04,230 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
352 |
+
2024-11-13 19:11:09,231 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
353 |
+
2024-11-13 19:11:14,231 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
354 |
+
2024-11-13 19:11:17,000 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
|
355 |
+
2024-11-13 19:11:17,847 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
356 |
+
2024-11-13 19:11:17,848 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
357 |
+
2024-11-13 19:11:17,848 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
358 |
+
2024-11-13 19:11:19,311 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
359 |
+
2024-11-13 19:11:20,004 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
360 |
+
2024-11-13 19:11:24,312 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
361 |
+
2024-11-13 19:11:29,312 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
362 |
+
2024-11-13 19:11:32,848 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
363 |
+
2024-11-13 19:11:32,848 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
364 |
+
2024-11-13 19:11:32,849 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
365 |
+
2024-11-13 19:11:35,231 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
366 |
+
2024-11-13 19:11:40,232 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
367 |
+
2024-11-13 19:11:44,014 INFO Thread-12 :1939 [dir_watcher.py:_on_file_modified():288] file/dir modified: /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/files/output.log
|
368 |
+
2024-11-13 19:11:46,173 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
369 |
+
2024-11-13 19:11:47,003 DEBUG SenderThread:1939 [sender.py:send():382] send: stats
|
370 |
+
2024-11-13 19:11:47,847 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: internal_messages
|
371 |
+
2024-11-13 19:11:47,847 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: stop_status
|
372 |
+
2024-11-13 19:11:47,848 DEBUG SenderThread:1939 [sender.py:send_request():409] send_request: stop_status
|
373 |
+
2024-11-13 19:11:51,230 DEBUG HandlerThread:1939 [handler.py:handle_request():146] handle_request: status_report
|
wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/logs/debug.log
ADDED
@@ -0,0 +1,30 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
2024-11-13 19:02:46,428 INFO MainThread:173 [wandb_setup.py:_flush():76] Current SDK version is 0.16.3
|
2 |
+
2024-11-13 19:02:46,428 INFO MainThread:173 [wandb_setup.py:_flush():76] Configure stats pid to 173
|
3 |
+
2024-11-13 19:02:46,428 INFO MainThread:173 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
|
4 |
+
2024-11-13 19:02:46,428 INFO MainThread:173 [wandb_setup.py:_flush():76] Loading settings from /opt/ml/code/wandb/settings
|
5 |
+
2024-11-13 19:02:46,428 INFO MainThread:173 [wandb_setup.py:_flush():76] Loading settings from environment variables: {'root_dir': '/opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048', 'api_key': '***REDACTED***', 'project': 'sft-codecontests-1112', 'run_name': 'sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048'}
|
6 |
+
2024-11-13 19:02:46,428 INFO MainThread:173 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
|
7 |
+
2024-11-13 19:02:46,428 INFO MainThread:173 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': 'qlearning_reasoning/training/sft.py', 'program_abspath': '/opt/ml/code/qlearning_reasoning/training/sft.py', 'program': '/opt/ml/code/qlearning_reasoning/training/sft.py'}
|
8 |
+
2024-11-13 19:02:46,428 INFO MainThread:173 [wandb_init.py:_log_setup():526] Logging user logs to /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/logs/debug.log
|
9 |
+
2024-11-13 19:02:46,428 INFO MainThread:173 [wandb_init.py:_log_setup():527] Logging internal logs to /opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/logs/debug-internal.log
|
10 |
+
2024-11-13 19:02:46,428 INFO MainThread:173 [wandb_init.py:init():566] calling init triggers
|
11 |
+
2024-11-13 19:02:46,428 INFO MainThread:173 [wandb_init.py:init():573] wandb.init called with sweep_config: {}
|
12 |
+
config: {'sagemaker_training_job_name': 'kushalarora-rvv-main-2024-11-13-17-55-42-664', 'batch_size': 64, 'cache_dir': '/opt/ml/data/input/.cache', 'dataset': 'code-contests', 'gradient_accumulation_steps': 8, 'logging_steps': 5, 'lr': '1e-6', 'max_seq_length': 2048, 'model_name': 'Qwen/Qwen2.5-Coder-7B-Instruct', 'num_train_epochs': 2, 'output_dir': '/opt/ml/output', 'push_to_hub': 'true', 'run_name': 'sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048', 'sagemaker_container_log_level': 20, 'sagemaker_instance_type': 'ml.p5.48xlarge', 'sagemaker_job_name': 'kushalarora-rvv-main-2024-11-13-17-55-42-664', 'sagemaker_program': 'sft.py', 'sagemaker_region': 'us-east-1', 'sagemaker_submit_directory': 's3://tri-ml-datasets/kushalarora-rvv-main-2024-11-13-17-55-42-664/source/sourcedir.tar.gz', 'sagemaker_torch_distributed_enabled': 'true', 'save_steps': 100, 'schedule': 'cosine', 'wandb_project': 'sft-codecontests-1112', 'weight_decay': 0.0, 'additional_framework_parameters': {'sagemaker_instance_type': 'ml.p5.48xlarge', 'sagemaker_torch_distributed_enabled': True}, 'channel_input_dirs': {}, 'current_host': 'algo-1', 'current_instance_group': 'homogeneousCluster', 'current_instance_group_hosts': ['algo-1'], 'current_instance_type': 'ml.p5.48xlarge', 'distribution_hosts': ['algo-1'], 'distribution_instance_groups': ['homogeneousCluster'], 'framework_module': 'sagemaker_pytorch_container.training:main', 'hosts': ['algo-1'], 'hyperparameters': {'batch_size': 64, 'cache_dir': '/opt/ml/data/input/.cache', 'dataset': 'code-contests', 'gradient_accumulation_steps': 8, 'logging_steps': 5, 'lr': '1e-6', 'max_seq_length': 2048, 'model_name': 'Qwen/Qwen2.5-Coder-7B-Instruct', 'num_train_epochs': 2, 'output_dir': '/opt/ml/model/', 'push_to_hub': True, 'run_name': 'sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048', 'save_steps': 100, 'schedule': 'cosine', 'wandb_project': 'sft-codecontests-1112', 'weight_decay': 0.0}, 'input_config_dir': '/opt/ml/input/config', 'input_data_config': {}, 'input_dir': '/opt/ml/input', 'instance_groups': ['homogeneousCluster'], 'instance_groups_dict': {'homogeneousCluster': {'hosts': ['algo-1'], 'instance_group_name': 'homogeneousCluster', 'instance_type': 'ml.p5.48xlarge'}}, 'is_hetero': False, 'is_master': True, 'is_modelparallel_enabled': None, 'is_smddpmprun_installed': False, 'is_smddprun_installed': True, 'job_name': 'kushalarora-rvv-main-2024-11-13-17-55-42-664', 'log_level': 20, 'master_hostname': 'algo-1', 'model_dir': '/opt/ml/model', 'module_dir': 's3://tri-ml-datasets/kushalarora-rvv-main-2024-11-13-17-55-42-664/source/sourcedir.tar.gz', 'module_name': '/opt/ml/code/qlearning_reasoning/training/sft', 'network_interface_name': 'eth0', 'num_cpus': 192, 'num_gpus': 8, 'num_neurons': 0, 'output_data_dir': '/opt/ml/output/data', 'output_intermediate_dir': '/opt/ml/output/intermediate', 'resource_config': {'current_group_name': 'homogeneousCluster', 'current_host': 'algo-1', 'current_instance_type': 'ml.p5.48xlarge', 'hosts': ['algo-1'], 'instance_groups': [{'hosts': ['algo-1'], 'instance_group_name': 'homogeneousCluster', 'instance_type': 'ml.p5.48xlarge'}], 'network_interface_name': 'eth0'}, 'user_entry_point': '/opt/ml/code/qlearning_reasoning/training/sft.py'}
|
13 |
+
2024-11-13 19:02:46,428 INFO MainThread:173 [wandb_init.py:init():616] starting backend
|
14 |
+
2024-11-13 19:02:46,428 INFO MainThread:173 [wandb_init.py:init():620] setting up manager
|
15 |
+
2024-11-13 19:02:46,429 INFO MainThread:173 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
|
16 |
+
2024-11-13 19:02:46,431 INFO MainThread:173 [wandb_init.py:init():628] backend started and connected
|
17 |
+
2024-11-13 19:02:46,434 INFO MainThread:173 [wandb_init.py:init():720] updated telemetry
|
18 |
+
2024-11-13 19:02:46,440 INFO MainThread:173 [wandb_init.py:init():753] communicating run to backend with 90.0 second timeout
|
19 |
+
2024-11-13 19:02:46,781 INFO MainThread:173 [wandb_run.py:_on_init():2262] communicating current version
|
20 |
+
2024-11-13 19:02:46,812 INFO MainThread:173 [wandb_run.py:_on_init():2271] got version response upgrade_message: "wandb version 0.18.6 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
|
21 |
+
|
22 |
+
2024-11-13 19:02:46,813 INFO MainThread:173 [wandb_init.py:init():804] starting run threads in backend
|
23 |
+
2024-11-13 19:03:02,000 INFO MainThread:173 [wandb_run.py:_console_start():2241] atexit reg
|
24 |
+
2024-11-13 19:03:02,000 INFO MainThread:173 [wandb_run.py:_redirect():2096] redirect: wrap_raw
|
25 |
+
2024-11-13 19:03:02,000 INFO MainThread:173 [wandb_run.py:_redirect():2161] Wrapping output streams.
|
26 |
+
2024-11-13 19:03:02,001 INFO MainThread:173 [wandb_run.py:_redirect():2186] Redirects installed.
|
27 |
+
2024-11-13 19:03:02,001 INFO MainThread:173 [wandb_init.py:init():847] run started, returning control to user process
|
28 |
+
2024-11-13 19:03:02,004 INFO MainThread:173 [wandb_run.py:_config_callback():1343] config_cb None None {'vocab_size': 151671, 'max_position_embeddings': 32768, 'hidden_size': 3584, 'intermediate_size': 18944, 'num_hidden_layers': 28, 'num_attention_heads': 28, 'use_sliding_window': False, 'sliding_window': None, 'max_window_layers': 28, 'num_key_value_heads': 4, 'hidden_act': 'silu', 'initializer_range': 0.02, 'rms_norm_eps': 1e-06, 'use_cache': False, 'rope_theta': 1000000.0, 'rope_scaling': None, 'attention_dropout': 0.0, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'bfloat16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['Qwen2ForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': 151643, 'eos_token_id': 151645, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'Qwen/Qwen2.5-Coder-7B-Instruct', '_attn_implementation_autoset': True, 'transformers_version': '4.46.0', 'model_type': 'qwen2', 'output_dir': '/opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'eval_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 8, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 8, 'eval_accumulation_steps': None, 'eval_delay': 0, 'torch_empty_cache_steps': None, 'learning_rate': 1e-06, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 2, 'max_steps': -1, 'lr_scheduler_type': 'cosine', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/opt/ml/model//sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048/runs/Nov13_19-02-15_algo-1', 'logging_strategy': 'steps', 'logging_first_step': True, 'logging_steps': 5, 'logging_nan_inf_filter': True, 'save_strategy': 'epoch', 'save_steps': 100, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'restore_callback_states_from_checkpoint': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': 5, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': 'sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_acc-8_len-2048', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': ['full_shard', 'auto_wrap'], 'fsdp_min_num_params': 0, 'fsdp_config': {'activation_checkpointing': False, 'auto_wrap_policy': 'TRANSFORMER_BASED_WRAP', 'backward_prefetch': 'BACKWARD_PRE', 'cpu_ram_efficient_loading': True, 'forward_prefetch': False, 'offload_params': False, 'sharding_strategy': 'FULL_SHARD', 'state_dict_type': 'FULL_STATE_DICT', 'sync_module_states': True, 'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': 'sft-codecontests-qwen_ds-code-contests_model-Qwen2.5-Coder-7B-Instruct_sch-cosine_lr-1e-6_bs-64_', 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': {'use_reentrant': False}, 'include_inputs_for_metrics': False, 'include_for_metrics': [], 'eval_do_concat_batches': True, 'fp16_backend': 'auto', 'evaluation_strategy': None, 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None, 'batch_eval_metrics': False, 'eval_on_start': False, 'use_liger_kernel': False, 'eval_use_gather_object': False, 'dataset_text_field': 'text', 'packing': False, 'max_seq_length': 2048, 'dataset_num_proc': None, 'dataset_batch_size': 64, 'model_init_kwargs': None, 'dataset_kwargs': {}, 'eval_packing': None, 'num_of_sequences': 1024, 'chars_per_token': '<CHARS_PER_TOKEN>', 'use_liger': False}
|
29 |
+
2024-11-13 19:03:02,008 INFO MainThread:173 [wandb_config.py:__setitem__():151] config set model/num_parameters = 951599936 - <bound method Run._config_callback of <wandb.sdk.wandb_run.Run object at 0x7fb464957310>>
|
30 |
+
2024-11-13 19:03:02,009 INFO MainThread:173 [wandb_run.py:_config_callback():1343] config_cb model/num_parameters 951599936 None
|
wandb/run-20241113_190246-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1/run-kushalarora-rvv-main-2024-11-13-17-55-42-664-g0ro8r-algo-1.wandb
ADDED
Binary file (392 kB). View file
|
|