mattchurgin
commited on
Commit
•
0629305
1
Parent(s):
7812a20
End of training
Browse files- all_results.json +8 -8
- eval_results.json +4 -4
- pytorch_model.bin +1 -1
- run.sh +2 -1
- special_tokens_map.json +1 -1
- test_load_cv_en.py +6 -0
- train_results.json +5 -5
- trainer_state.json +12 -12
- training_args.bin +1 -1
all_results.json
CHANGED
@@ -1,14 +1,14 @@
|
|
1 |
{
|
2 |
-
"epoch":
|
3 |
"eval_loss": Infinity,
|
4 |
-
"eval_runtime": 10.
|
5 |
"eval_samples": 301,
|
6 |
-
"eval_samples_per_second": 28.
|
7 |
-
"eval_steps_per_second": 3.
|
8 |
"eval_wer": 1.0,
|
9 |
-
"train_loss":
|
10 |
-
"train_runtime":
|
11 |
"train_samples": 704,
|
12 |
-
"train_samples_per_second":
|
13 |
-
"train_steps_per_second": 2.
|
14 |
}
|
1 |
{
|
2 |
+
"epoch": 1.0,
|
3 |
"eval_loss": Infinity,
|
4 |
+
"eval_runtime": 10.6101,
|
5 |
"eval_samples": 301,
|
6 |
+
"eval_samples_per_second": 28.369,
|
7 |
+
"eval_steps_per_second": 3.582,
|
8 |
"eval_wer": 1.0,
|
9 |
+
"train_loss": 4.538612365722656,
|
10 |
+
"train_runtime": 31.4322,
|
11 |
"train_samples": 704,
|
12 |
+
"train_samples_per_second": 22.397,
|
13 |
+
"train_steps_per_second": 2.8
|
14 |
}
|
eval_results.json
CHANGED
@@ -1,9 +1,9 @@
|
|
1 |
{
|
2 |
-
"epoch":
|
3 |
"eval_loss": Infinity,
|
4 |
-
"eval_runtime": 10.
|
5 |
"eval_samples": 301,
|
6 |
-
"eval_samples_per_second": 28.
|
7 |
-
"eval_steps_per_second": 3.
|
8 |
"eval_wer": 1.0
|
9 |
}
|
1 |
{
|
2 |
+
"epoch": 1.0,
|
3 |
"eval_loss": Infinity,
|
4 |
+
"eval_runtime": 10.6101,
|
5 |
"eval_samples": 301,
|
6 |
+
"eval_samples_per_second": 28.369,
|
7 |
+
"eval_steps_per_second": 3.582,
|
8 |
"eval_wer": 1.0
|
9 |
}
|
pytorch_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1247823
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:612dd692031680aae02deafa36c7fa8dadce27d4b31ce9acb75461c1e48b9ddb
|
3 |
size 1247823
|
run.sh
CHANGED
@@ -4,7 +4,7 @@ python run_speech_recognition_ctc.py \
|
|
4 |
--dataset_config_name="ab" \
|
5 |
--output_dir="./" \
|
6 |
--overwrite_output_dir \
|
7 |
-
--num_train_epochs="
|
8 |
--save_steps="100" \
|
9 |
--eval_steps="500" \
|
10 |
--per_device_train_batch_size="8" \
|
@@ -21,3 +21,4 @@ python run_speech_recognition_ctc.py \
|
|
21 |
--push_to_hub \
|
22 |
--use_auth_token \
|
23 |
--do_train --do_eval
|
|
4 |
--dataset_config_name="ab" \
|
5 |
--output_dir="./" \
|
6 |
--overwrite_output_dir \
|
7 |
+
--num_train_epochs="1" \
|
8 |
--save_steps="100" \
|
9 |
--eval_steps="500" \
|
10 |
--per_device_train_batch_size="8" \
|
21 |
--push_to_hub \
|
22 |
--use_auth_token \
|
23 |
--do_train --do_eval
|
24 |
+
|
special_tokens_map.json
CHANGED
@@ -1 +1 @@
|
|
1 |
-
{"bos_token": "<s>", "eos_token": "</s>", "unk_token": "[UNK]", "pad_token": "[PAD]", "additional_special_tokens": [{"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}]}
|
1 |
+
{"bos_token": "<s>", "eos_token": "</s>", "unk_token": "[UNK]", "pad_token": "[PAD]", "additional_special_tokens": [{"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}]}
|
test_load_cv_en.py
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
|
2 |
+
from transformers import AutoModelForCTC, AutoProcessor
|
3 |
+
from datasets import load_dataset
|
4 |
+
train_ds = load_dataset('mozilla-foundation/common_voice_7_0', 'en', split='train[10:20]')
|
5 |
+
|
6 |
+
print(len(train_ds))
|
train_results.json
CHANGED
@@ -1,8 +1,8 @@
|
|
1 |
{
|
2 |
-
"epoch":
|
3 |
-
"train_loss":
|
4 |
-
"train_runtime":
|
5 |
"train_samples": 704,
|
6 |
-
"train_samples_per_second":
|
7 |
-
"train_steps_per_second": 2.
|
8 |
}
|
1 |
{
|
2 |
+
"epoch": 1.0,
|
3 |
+
"train_loss": 4.538612365722656,
|
4 |
+
"train_runtime": 31.4322,
|
5 |
"train_samples": 704,
|
6 |
+
"train_samples_per_second": 22.397,
|
7 |
+
"train_steps_per_second": 2.8
|
8 |
}
|
trainer_state.json
CHANGED
@@ -1,25 +1,25 @@
|
|
1 |
{
|
2 |
"best_metric": null,
|
3 |
"best_model_checkpoint": null,
|
4 |
-
"epoch":
|
5 |
-
"global_step":
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
9 |
"log_history": [
|
10 |
{
|
11 |
-
"epoch":
|
12 |
-
"step":
|
13 |
-
"total_flos":
|
14 |
-
"train_loss":
|
15 |
-
"train_runtime":
|
16 |
-
"train_samples_per_second":
|
17 |
-
"train_steps_per_second": 2.
|
18 |
}
|
19 |
],
|
20 |
-
"max_steps":
|
21 |
-
"num_train_epochs":
|
22 |
-
"total_flos":
|
23 |
"trial_name": null,
|
24 |
"trial_params": null
|
25 |
}
|
1 |
{
|
2 |
"best_metric": null,
|
3 |
"best_model_checkpoint": null,
|
4 |
+
"epoch": 1.0,
|
5 |
+
"global_step": 88,
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
9 |
"log_history": [
|
10 |
{
|
11 |
+
"epoch": 1.0,
|
12 |
+
"step": 88,
|
13 |
+
"total_flos": 130613499224064.0,
|
14 |
+
"train_loss": 4.538612365722656,
|
15 |
+
"train_runtime": 31.4322,
|
16 |
+
"train_samples_per_second": 22.397,
|
17 |
+
"train_steps_per_second": 2.8
|
18 |
}
|
19 |
],
|
20 |
+
"max_steps": 88,
|
21 |
+
"num_train_epochs": 1,
|
22 |
+
"total_flos": 130613499224064.0,
|
23 |
"trial_name": null,
|
24 |
"trial_params": null
|
25 |
}
|
training_args.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 2991
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d118e42535c934c70d07a3714d6d1920cb49fd8cfe6d64fc5e126f0905ace316
|
3 |
size 2991
|