ShengdingHu commited on
Commit
44daab9
1 Parent(s): 379cf8b

Training in progress, step 200

Browse files
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2db779b1c68f7ae8ffeb4dd2de8e2563b3f7743c0a57928cc788a2b2c4a539c7
3
- size 2631685
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5b060b206e6fb1504b07b307d826d38db5ba2fb1381a666b3c94f6bab9833e50
3
+ size 335851
runs/Feb04_09-56-16_node1/1643940000.8292065/events.out.tfevents.1643940000.node1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:05cd5c0696e9f666decec38653bdccc7c5681fd08656f011a6855df8c9dc1b88
3
+ size 5026
runs/Feb04_09-56-16_node1/events.out.tfevents.1643940000.node1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:34391dcf4362fc2380529845e033b890409a170b0014c87a3b6fc985d0777f04
3
+ size 4314
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d084e8b52f5c7cc4a6a29f9f14decebc1ce43459b020a9c05c1c22bd9831401a
3
  size 3183
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5f396571923132052a008693cef3a6acfe656e9be0d56e2e10ea73893ffe1b46
3
  size 3183
training_config.json CHANGED
@@ -1 +1 @@
1
- {"dataset_config_name": ["en"], "delta_type": "lora", "do_eval": true, "do_test": true, "do_train": true, "eval_dataset_config_name": ["en"], "eval_dataset_name": "sst2", "eval_steps": 200, "evaluation_strategy": "steps", "greater_is_better": true, "learning_rate": 0.0003, "load_best_model_at_end": true, "lora_r": 8, "max_source_length": 128, "metric_for_best_model": "average_metrics", "model_name_or_path": "../../../../plm_cache/t5-base", "num_train_epochs": 3, "output_dir": "outputs/bitfit/t5-base/sst2", "overwrite_output_dir": true, "per_device_eval_batch_size": 32, "per_device_train_batch_size": 32, "predict_with_generate": true, "push_to_hub": true, "save_steps": 200, "save_strategy": "steps", "save_total_limit": 1, "seed": 42, "split_validation_test": true, "task_name": "sst2", "test_dataset_config_name": ["en"], "test_dataset_name": "sst2", "tokenizer_name": "../../../../plm_cache/t5-base", "unfrozen_modules": ["deltas", "layer_norm", "final_layer_norm"], "warmup_steps": 0}
 
1
+ {"dataset_config_name": ["en"], "delta_type": "soft_prompt", "do_eval": true, "do_test": true, "do_train": true, "eval_dataset_config_name": ["en"], "eval_dataset_name": "sst2", "eval_steps": 200, "evaluation_strategy": "steps", "greater_is_better": true, "learning_rate": 0.0003, "load_best_model_at_end": true, "max_source_length": 128, "metric_for_best_model": "average_metrics", "model_name_or_path": "../../../../plm_cache/t5-base", "num_train_epochs": 3, "output_dir": "outputs/soft_prompt/t5-base/sst2", "overwrite_output_dir": true, "per_device_eval_batch_size": 32, "per_device_train_batch_size": 32, "predict_with_generate": true, "push_to_hub": true, "save_steps": 200, "save_strategy": "steps", "save_total_limit": 1, "seed": 42, "soft_token_num": 100, "split_validation_test": true, "task_name": "sst2", "test_dataset_config_name": ["en"], "test_dataset_name": "sst2", "tokenizer_name": "../../../../plm_cache/t5-base", "unfrozen_modules": ["deltas"], "warmup_steps": 0}