Sarmila commited on
Commit
68f4d4c
1 Parent(s): 0b966ab

End of training

Browse files
config.json CHANGED
@@ -18,7 +18,7 @@
18
  "pad_token_id": 0,
19
  "position_embedding_type": "absolute",
20
  "torch_dtype": "float32",
21
- "transformers_version": "4.27.3",
22
  "type_vocab_size": 2,
23
  "use_cache": true,
24
  "vocab_size": 31090
 
18
  "pad_token_id": 0,
19
  "position_embedding_type": "absolute",
20
  "torch_dtype": "float32",
21
+ "transformers_version": "4.26.1",
22
  "type_vocab_size": 2,
23
  "use_cache": true,
24
  "vocab_size": 31090
generation_config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
  "_from_model_config": true,
3
  "pad_token_id": 0,
4
- "transformers_version": "4.27.3"
5
  }
 
1
  {
2
  "_from_model_config": true,
3
  "pad_token_id": 0,
4
+ "transformers_version": "4.26.1"
5
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e480d5e06277f6cff3cb8a1cd04ed1958dd62b7dc1136e9b41c1cc5048f026c5
3
  size 439877817
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4330797dec4443d9e74f99267b285731b38f904c1974f3a3b215356807254941
3
  size 439877817
runs/Mar27_17-13-48_9de9c7ff8909/events.out.tfevents.1679937382.9de9c7ff8909.23.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:66cf911efc7002bc81ff449635d6210511ee5776bcbf1f9efde8be48f26e65cb
3
+ size 249
runs/Mar27_18-34-44_9de9c7ff8909/events.out.tfevents.1679942101.9de9c7ff8909.23.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:91f290d594b5f69b8a33311cf4ec14680c9aeb6d7cce0ddf325fd2e6dfa8afb4
3
+ size 249
runs/Mar27_18-34-44_9de9c7ff8909/events.out.tfevents.1679942282.9de9c7ff8909.23.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:891052df8fb1663af13191a63dbc5d7e7cfce6cff2209ee5030ebc0b19795d7a
3
+ size 249
runs/Mar27_18-43-29_9de9c7ff8909/events.out.tfevents.1679942618.9de9c7ff8909.23.3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:850ed277a1a1269d1b043a8196884bf78cd777bf837668903a75dd0909dd7208
3
+ size 667
runs/Mar27_18-44-35_9de9c7ff8909/1679942688.761012/events.out.tfevents.1679942688.9de9c7ff8909.23.5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:61c938650d80ebc105e03223cb7f930f8b0ceebe238396ce3e3a31813aa18985
3
+ size 5743
runs/Mar27_18-44-35_9de9c7ff8909/events.out.tfevents.1679942685.9de9c7ff8909.23.4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3661b26c4b3199dda799a0e23945ba88b8ad38e2f6e371b6a5173208d68f2f2f
3
+ size 5172
runs/Mar27_18-44-35_9de9c7ff8909/events.out.tfevents.1679942775.9de9c7ff8909.23.6 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:766dd0ba60d14b1d8fb3cea467edbc9bdc7d36127e00a8cad6bde84c077663f1
3
+ size 306
tokenizer_config.json CHANGED
@@ -4,6 +4,7 @@
4
  "do_lower_case": true,
5
  "mask_token": "[MASK]",
6
  "model_max_length": 1000000000000000019884624838656,
 
7
  "never_split": null,
8
  "pad_token": "[PAD]",
9
  "sep_token": "[SEP]",
 
4
  "do_lower_case": true,
5
  "mask_token": "[MASK]",
6
  "model_max_length": 1000000000000000019884624838656,
7
+ "name_or_path": "allenai/scibert_scivocab_uncased",
8
  "never_split": null,
9
  "pad_token": "[PAD]",
10
  "sep_token": "[SEP]",
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:da38d984790897734a9de16014a32a8eea159213d0ce8b00d5d2c11f6f9966af
3
- size 3643
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d62e013d9367e24ad6bf51903dff3daae861e6fee5d9782bae38823464eb40bd
3
+ size 3579