Jeska commited on
Commit
a464cbb
1 Parent(s): 9b865d5

Training in progress, epoch 1

Browse files
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "Jeska/BertjeWDialDataALLQonly09",
3
  "architectures": [
4
  "BertForSequenceClassification"
5
  ],
1
  {
2
+ "_name_or_path": "outputDAQonly09/",
3
  "architectures": [
4
  "BertForSequenceClassification"
5
  ],
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3ee8f2e53286fd05918afbc27226d76402f1a4bbcd70e3db77f8f56e0cfea337
3
  size 437190509
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:137617742b8a518f91e2d2e683119e5639e3fb5bced849cba0b0504543ad61c1
3
  size 437190509
runs/Dec15_13-02-52_jbuhmann/1639573387.7159743/events.out.tfevents.1639573387.jbuhmann.27644.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:93b719e28d5650e253c70b54dc7257e3fcff8bc3513ac1f408546e2374f1511e
3
+ size 4611
runs/Dec15_13-02-52_jbuhmann/events.out.tfevents.1639573387.jbuhmann.27644.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:309397975710ce0993ce96fe48de7e40bfbde9208b222e4b38c93a71c3ad1146
3
+ size 17071
tokenizer_config.json CHANGED
@@ -1 +1 @@
1
- {"do_lower_case": false, "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "tokenize_chinese_chars": true, "strip_accents": null, "model_max_length": 512, "special_tokens_map_file": "/home/ubuntu/.cache/huggingface/transformers/adb82a117c09b0f8768357de8e836a9e0610730782f82edc49dd0020c48f1d03.dd8bd9bfd3664b530ea4e645105f557769387b3da9f79bdb55ed556bdd80611d", "name_or_path": "Jeska/BertjeWDialDataALLQonly09", "do_basic_tokenize": true, "never_split": null, "tokenizer_class": "BertTokenizer"}
1
+ {"do_lower_case": false, "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "tokenize_chinese_chars": true, "strip_accents": null, "model_max_length": 512, "special_tokens_map_file": "/home/ubuntu/.cache/huggingface/transformers/adb82a117c09b0f8768357de8e836a9e0610730782f82edc49dd0020c48f1d03.dd8bd9bfd3664b530ea4e645105f557769387b3da9f79bdb55ed556bdd80611d", "name_or_path": "outputDAQonly09/", "do_basic_tokenize": true, "never_split": null, "tokenizer_class": "BertTokenizer"}
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:17302ce11d091603c83b86a7a5def3a66dff26422fb364425e66dddee3fea6c2
3
  size 2991
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:026f94c540cddb31b4be833b9b0c56cbae0687b3c10a1507690edd4eeb528a96
3
  size 2991