AnaniyaX commited on
Commit
355c49c
1 Parent(s): a3c7fbd

Training in progress epoch 0

Browse files
Files changed (6) hide show
  1. README.md +8 -3
  2. special_tokens_map.json +7 -0
  3. tf_model.h5 +1 -1
  4. tokenizer.json +0 -0
  5. tokenizer_config.json +13 -0
  6. vocab.txt +0 -0
README.md CHANGED
@@ -3,18 +3,20 @@ license: apache-2.0
3
  tags:
4
  - generated_from_keras_callback
5
  model-index:
6
- - name: decision-distilbert-uncased
7
  results: []
8
  ---
9
 
10
  <!-- This model card has been generated automatically according to the information Keras had access to. You should
11
  probably proofread and complete it, then remove this comment. -->
12
 
13
- # decision-distilbert-uncased
14
 
15
  This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on an unknown dataset.
16
  It achieves the following results on the evaluation set:
17
-
 
 
18
 
19
  ## Model description
20
 
@@ -38,6 +40,9 @@ The following hyperparameters were used during training:
38
 
39
  ### Training results
40
 
 
 
 
41
 
42
 
43
  ### Framework versions
3
  tags:
4
  - generated_from_keras_callback
5
  model-index:
6
+ - name: AnaniyaX/decision-distilbert-uncased
7
  results: []
8
  ---
9
 
10
  <!-- This model card has been generated automatically according to the information Keras had access to. You should
11
  probably proofread and complete it, then remove this comment. -->
12
 
13
+ # AnaniyaX/decision-distilbert-uncased
14
 
15
  This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on an unknown dataset.
16
  It achieves the following results on the evaluation set:
17
+ - Train Loss: 0.5755
18
+ - Train Accuracy: 0.8520
19
+ - Epoch: 0
20
 
21
  ## Model description
22
 
40
 
41
  ### Training results
42
 
43
+ | Train Loss | Train Accuracy | Epoch |
44
+ |:----------:|:--------------:|:-----:|
45
+ | 0.5755 | 0.8520 | 0 |
46
 
47
 
48
  ### Framework versions
special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
tf_model.h5 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c86a6616d91abac1eeef92ad6928bf83eb976f90dc4277e49742e2810bad4a26
3
  size 267955144
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0db06b6b956f067528d0b542452d5ce3f509dad58f71df51065f9a13420d2684
3
  size 267955144
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
tokenizer_config.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "do_lower_case": true,
4
+ "mask_token": "[MASK]",
5
+ "model_max_length": 512,
6
+ "pad_token": "[PAD]",
7
+ "sep_token": "[SEP]",
8
+ "special_tokens_map_file": null,
9
+ "strip_accents": null,
10
+ "tokenize_chinese_chars": true,
11
+ "tokenizer_class": "DistilBertTokenizer",
12
+ "unk_token": "[UNK]"
13
+ }
vocab.txt ADDED
The diff for this file is too large to render. See raw diff