13048909972 commited on
Commit
540cc1d
1 Parent(s): f771621

Training in progress, step 400

Browse files
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bf74c072c9a82fa79b630c2a15e3f6d6bc8b0561c591b468ca1ded58f42e2ea8
3
  size 1262091761
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a364cb1fa5839ec73cef456a2b868c16845a58e01cb3e6b0257b65d652456d89
3
  size 1262091761
runs/Dec09_09-01-28_cuda11-renwei-docker/1639012200.6069648/events.out.tfevents.1639012200.cuda11-renwei-docker.296.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bff574ade91fcc7f5390b57bc7082a456a666fea8a3aa525d4f1dd511e39f91f
3
+ size 4920
runs/Dec09_09-01-28_cuda11-renwei-docker/events.out.tfevents.1639012200.cuda11-renwei-docker.296.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f095fadee554b40e27ba1bf9c3df16892f87f6c1f50b14bafa0ad3cc60ef0294
3
+ size 6148
special_tokens_map.json CHANGED
@@ -1 +1 @@
1
- {"bos_token": "<s>", "eos_token": "</s>", "unk_token": "[UNK]", "pad_token": "[PAD]", "additional_special_tokens": [{"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}]}
 
1
+ {"bos_token": "<s>", "eos_token": "</s>", "unk_token": "[UNK]", "pad_token": "[PAD]", "additional_special_tokens": [{"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}]}
tokenizer_config.json CHANGED
@@ -1 +1 @@
1
- {"unk_token": "[UNK]", "bos_token": "<s>", "eos_token": "</s>", "pad_token": "[PAD]", "do_lower_case": false, "word_delimiter_token": "|", "special_tokens_map_file": null, "tokenizer_file": null, "name_or_path": "./wav2vec2-common_voice-tr-demo", "tokenizer_class": "Wav2Vec2CTCTokenizer"}
 
1
+ {"unk_token": "[UNK]", "bos_token": "<s>", "eos_token": "</s>", "pad_token": "[PAD]", "do_lower_case": false, "word_delimiter_token": "|", "special_tokens_map_file": null, "tokenizer_file": null, "name_or_path": "/project/dave/neural_network/model/xlsr/xlsr_speech-recognition/wav2vec2-common_voice-tr-demo", "tokenizer_class": "Wav2Vec2CTCTokenizer"}
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ed402f5ac7e7b90ccfff497218eca7058afe46ab9a7dbc59230a0a24c224dfd6
3
- size 2927
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:062fa2866af9d9264789ececb76883edf1ef7d144638e061c9382a6e56e435c0
3
+ size 3055