Training in progress, step 400
Browse files- .gitattributes +1 -0
- adapter_model.safetensors +1 -1
- runs/May22_05-46-00_92dcc4555414/events.out.tfevents.1716356830.92dcc4555414.24.0 +2 -2
- wandb/debug-internal.log +0 -0
- wandb/run-20240522_054348-vgrzs6jq/files/output.log +38 -0
- wandb/run-20240522_054348-vgrzs6jq/files/wandb-summary.json +1 -1
- wandb/run-20240522_054348-vgrzs6jq/logs/debug-internal.log +0 -0
- wandb/run-20240522_054348-vgrzs6jq/run-vgrzs6jq.wandb +0 -0
.gitattributes
CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
36 |
+
wandb/run-20240522_054348-vgrzs6jq/run-vgrzs6jq.wandb filter=lfs diff=lfs merge=lfs -text
|
adapter_model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1279323952
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:adf2b8d6fdde761c05450f36be5841531cf2f9b29c363ab64000f9d096968d68
|
3 |
size 1279323952
|
runs/May22_05-46-00_92dcc4555414/events.out.tfevents.1716356830.92dcc4555414.24.0
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:98faef80cb9585607e93d29d7a512c5167bb550499e5fa895d9dd7d486f27b88
|
3 |
+
size 35517
|
wandb/debug-internal.log
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
wandb/run-20240522_054348-vgrzs6jq/files/output.log
CHANGED
@@ -698,3 +698,41 @@ tokenizer config file saved in /kaggle/working/checkpoint-300/tokenizer_config.j
|
|
698 |
Special tokens file saved in /kaggle/working/checkpoint-300/special_tokens_map.json
|
699 |
tokenizer config file saved in /kaggle/working/tokenizer_config.json
|
700 |
Special tokens file saved in /kaggle/working/special_tokens_map.json
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
698 |
Special tokens file saved in /kaggle/working/checkpoint-300/special_tokens_map.json
|
699 |
tokenizer config file saved in /kaggle/working/tokenizer_config.json
|
700 |
Special tokens file saved in /kaggle/working/special_tokens_map.json
|
701 |
+
Deleting older checkpoint [/kaggle/working/checkpoint-200] due to args.save_total_limit
|
702 |
+
***** Running Evaluation *****
|
703 |
+
Num examples = 1801
|
704 |
+
Batch size = 1
|
705 |
+
Saving model checkpoint to /kaggle/working/checkpoint-400
|
706 |
+
loading configuration file config.json from cache at /root/.cache/huggingface/hub/models--meta-llama--Llama-2-7b-chat-hf/snapshots/f5db02db724555f92da89c216ac04704f23d4590/config.json
|
707 |
+
Model config LlamaConfig {
|
708 |
+
"_name_or_path": "meta-llama/Llama-2-7b-chat-hf",
|
709 |
+
"architectures": [
|
710 |
+
"LlamaForCausalLM"
|
711 |
+
],
|
712 |
+
"attention_bias": false,
|
713 |
+
"attention_dropout": 0.0,
|
714 |
+
"bos_token_id": 1,
|
715 |
+
"eos_token_id": 2,
|
716 |
+
"hidden_act": "silu",
|
717 |
+
"hidden_size": 4096,
|
718 |
+
"initializer_range": 0.02,
|
719 |
+
"intermediate_size": 11008,
|
720 |
+
"max_position_embeddings": 4096,
|
721 |
+
"model_type": "llama",
|
722 |
+
"num_attention_heads": 32,
|
723 |
+
"num_hidden_layers": 32,
|
724 |
+
"num_key_value_heads": 32,
|
725 |
+
"pretraining_tp": 1,
|
726 |
+
"rms_norm_eps": 1e-05,
|
727 |
+
"rope_scaling": null,
|
728 |
+
"rope_theta": 10000.0,
|
729 |
+
"tie_word_embeddings": false,
|
730 |
+
"torch_dtype": "float16",
|
731 |
+
"transformers_version": "4.39.3",
|
732 |
+
"use_cache": true,
|
733 |
+
"vocab_size": 32000
|
734 |
+
}
|
735 |
+
tokenizer config file saved in /kaggle/working/checkpoint-400/tokenizer_config.json
|
736 |
+
Special tokens file saved in /kaggle/working/checkpoint-400/special_tokens_map.json
|
737 |
+
tokenizer config file saved in /kaggle/working/tokenizer_config.json
|
738 |
+
Special tokens file saved in /kaggle/working/special_tokens_map.json
|
wandb/run-20240522_054348-vgrzs6jq/files/wandb-summary.json
CHANGED
@@ -1 +1 @@
|
|
1 |
-
{"train/loss": 0.
|
|
|
1 |
+
{"train/loss": 0.0197, "train/grad_norm": 0.09612330794334412, "train/learning_rate": 1.9158962821045113e-05, "train/rewards/chosen": -8.07567024230957, "train/rewards/rejected": -30.504592895507812, "train/rewards/accuracies": 1.0, "train/rewards/margins": 22.428918838500977, "train/logps/rejected": -5489.53271484375, "train/logps/chosen": -2377.00732421875, "train/logits/rejected": -3.1316120624542236, "train/logits/chosen": -2.682544231414795, "train/epoch": 0.22, "train/global_step": 400, "_timestamp": 1716386464.2571802, "_runtime": 29835.621697187424, "_step": 43, "eval/loss": 0.17712955176830292, "eval/runtime": 5069.1824, "eval/samples_per_second": 0.355, "eval/steps_per_second": 0.355, "eval/rewards/chosen": -8.881494522094727, "eval/rewards/rejected": -29.159704208374023, "eval/rewards/accuracies": 0.9600222110748291, "eval/rewards/margins": 20.278209686279297, "eval/logps/rejected": -5199.47314453125, "eval/logps/chosen": -2491.822265625, "eval/logits/rejected": -3.1585352420806885, "eval/logits/chosen": -2.733020782470703}
|
wandb/run-20240522_054348-vgrzs6jq/logs/debug-internal.log
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
wandb/run-20240522_054348-vgrzs6jq/run-vgrzs6jq.wandb
CHANGED
Binary files a/wandb/run-20240522_054348-vgrzs6jq/run-vgrzs6jq.wandb and b/wandb/run-20240522_054348-vgrzs6jq/run-vgrzs6jq.wandb differ
|
|