mattchurgin commited on
Commit
0629305
1 Parent(s): 7812a20

End of training

Browse files
all_results.json CHANGED
@@ -1,14 +1,14 @@
1
  {
2
- "epoch": 3.0,
3
  "eval_loss": Infinity,
4
- "eval_runtime": 10.6149,
5
  "eval_samples": 301,
6
- "eval_samples_per_second": 28.356,
7
- "eval_steps_per_second": 3.58,
8
  "eval_wer": 1.0,
9
- "train_loss": 3.7429571440725615,
10
- "train_runtime": 101.8953,
11
  "train_samples": 704,
12
- "train_samples_per_second": 20.727,
13
- "train_steps_per_second": 2.591
14
  }
1
  {
2
+ "epoch": 1.0,
3
  "eval_loss": Infinity,
4
+ "eval_runtime": 10.6101,
5
  "eval_samples": 301,
6
+ "eval_samples_per_second": 28.369,
7
+ "eval_steps_per_second": 3.582,
8
  "eval_wer": 1.0,
9
+ "train_loss": 4.538612365722656,
10
+ "train_runtime": 31.4322,
11
  "train_samples": 704,
12
+ "train_samples_per_second": 22.397,
13
+ "train_steps_per_second": 2.8
14
  }
eval_results.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
- "epoch": 3.0,
3
  "eval_loss": Infinity,
4
- "eval_runtime": 10.6149,
5
  "eval_samples": 301,
6
- "eval_samples_per_second": 28.356,
7
- "eval_steps_per_second": 3.58,
8
  "eval_wer": 1.0
9
  }
1
  {
2
+ "epoch": 1.0,
3
  "eval_loss": Infinity,
4
+ "eval_runtime": 10.6101,
5
  "eval_samples": 301,
6
+ "eval_samples_per_second": 28.369,
7
+ "eval_steps_per_second": 3.582,
8
  "eval_wer": 1.0
9
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:941203e6a17701b67c1cafc3d6cb41d4cf74fc16cad3faa57ca84153d5da0cd2
3
  size 1247823
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:612dd692031680aae02deafa36c7fa8dadce27d4b31ce9acb75461c1e48b9ddb
3
  size 1247823
run.sh CHANGED
@@ -4,7 +4,7 @@ python run_speech_recognition_ctc.py \
4
  --dataset_config_name="ab" \
5
  --output_dir="./" \
6
  --overwrite_output_dir \
7
- --num_train_epochs="3" \
8
  --save_steps="100" \
9
  --eval_steps="500" \
10
  --per_device_train_batch_size="8" \
@@ -21,3 +21,4 @@ python run_speech_recognition_ctc.py \
21
  --push_to_hub \
22
  --use_auth_token \
23
  --do_train --do_eval
 
4
  --dataset_config_name="ab" \
5
  --output_dir="./" \
6
  --overwrite_output_dir \
7
+ --num_train_epochs="1" \
8
  --save_steps="100" \
9
  --eval_steps="500" \
10
  --per_device_train_batch_size="8" \
21
  --push_to_hub \
22
  --use_auth_token \
23
  --do_train --do_eval
24
+
special_tokens_map.json CHANGED
@@ -1 +1 @@
1
- {"bos_token": "<s>", "eos_token": "</s>", "unk_token": "[UNK]", "pad_token": "[PAD]", "additional_special_tokens": [{"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}]}
1
+ {"bos_token": "<s>", "eos_token": "</s>", "unk_token": "[UNK]", "pad_token": "[PAD]", "additional_special_tokens": [{"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}]}
test_load_cv_en.py ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
1
+
2
+ from transformers import AutoModelForCTC, AutoProcessor
3
+ from datasets import load_dataset
4
+ train_ds = load_dataset('mozilla-foundation/common_voice_7_0', 'en', split='train[10:20]')
5
+
6
+ print(len(train_ds))
train_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
- "epoch": 3.0,
3
- "train_loss": 3.7429571440725615,
4
- "train_runtime": 101.8953,
5
  "train_samples": 704,
6
- "train_samples_per_second": 20.727,
7
- "train_steps_per_second": 2.591
8
  }
1
  {
2
+ "epoch": 1.0,
3
+ "train_loss": 4.538612365722656,
4
+ "train_runtime": 31.4322,
5
  "train_samples": 704,
6
+ "train_samples_per_second": 22.397,
7
+ "train_steps_per_second": 2.8
8
  }
trainer_state.json CHANGED
@@ -1,25 +1,25 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 3.0,
5
- "global_step": 264,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
9
  "log_history": [
10
  {
11
- "epoch": 3.0,
12
- "step": 264,
13
- "total_flos": 391664972335104.0,
14
- "train_loss": 3.7429571440725615,
15
- "train_runtime": 101.8953,
16
- "train_samples_per_second": 20.727,
17
- "train_steps_per_second": 2.591
18
  }
19
  ],
20
- "max_steps": 264,
21
- "num_train_epochs": 3,
22
- "total_flos": 391664972335104.0,
23
  "trial_name": null,
24
  "trial_params": null
25
  }
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 1.0,
5
+ "global_step": 88,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
9
  "log_history": [
10
  {
11
+ "epoch": 1.0,
12
+ "step": 88,
13
+ "total_flos": 130613499224064.0,
14
+ "train_loss": 4.538612365722656,
15
+ "train_runtime": 31.4322,
16
+ "train_samples_per_second": 22.397,
17
+ "train_steps_per_second": 2.8
18
  }
19
  ],
20
+ "max_steps": 88,
21
+ "num_train_epochs": 1,
22
+ "total_flos": 130613499224064.0,
23
  "trial_name": null,
24
  "trial_params": null
25
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:59923f372893625f0de9829d44a4b74c369997fe40d4b7f08759e5a30b318f24
3
  size 2991
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d118e42535c934c70d07a3714d6d1920cb49fd8cfe6d64fc5e126f0905ace316
3
  size 2991