ngchuchi commited on
Commit
c7c0930
1 Parent(s): 27016d9

Training in progress, epoch 1

Browse files
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "deepset/roberta-base-squad2",
3
  "architectures": [
4
  "RobertaForQuestionAnswering"
5
  ],
@@ -13,16 +13,15 @@
13
  "hidden_size": 768,
14
  "initializer_range": 0.02,
15
  "intermediate_size": 3072,
16
- "language": "english",
17
  "layer_norm_eps": 1e-05,
18
  "max_position_embeddings": 514,
19
  "model_type": "roberta",
20
- "name": "Roberta",
21
  "num_attention_heads": 12,
22
  "num_hidden_layers": 12,
23
  "pad_token_id": 1,
24
  "position_embedding_type": "absolute",
25
  "torch_dtype": "float32",
 
26
  "transformers_version": "4.38.2",
27
  "type_vocab_size": 1,
28
  "use_cache": true,
 
1
  {
2
+ "_name_or_path": "abhi1nandy2/EManuals_RoBERTa",
3
  "architectures": [
4
  "RobertaForQuestionAnswering"
5
  ],
 
13
  "hidden_size": 768,
14
  "initializer_range": 0.02,
15
  "intermediate_size": 3072,
 
16
  "layer_norm_eps": 1e-05,
17
  "max_position_embeddings": 514,
18
  "model_type": "roberta",
 
19
  "num_attention_heads": 12,
20
  "num_hidden_layers": 12,
21
  "pad_token_id": 1,
22
  "position_embedding_type": "absolute",
23
  "torch_dtype": "float32",
24
+ "total_flos": 4462005349122048000,
25
  "transformers_version": "4.38.2",
26
  "type_vocab_size": 1,
27
  "use_cache": true,
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:90aefae1a9fe83b3c13079a1ae4f19a3300505cb2f26dcba9b52570303879ec6
3
  size 496250232
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:44cd16b8f6567631b329bd500bc1c2ede2b9eff6ff139733410b13a5b77e1220
3
  size 496250232
runs/Mar26_04-21-48_69c6919e0c2f/events.out.tfevents.1711426922.69c6919e0c2f.497.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:85b3c9dbba900ce89d88a1998529078fc6166aafdc12204bdf7e3cd65b3b777c
3
+ size 5148
special_tokens_map.json CHANGED
@@ -2,49 +2,49 @@
2
  "bos_token": {
3
  "content": "<s>",
4
  "lstrip": false,
5
- "normalized": true,
6
  "rstrip": false,
7
  "single_word": false
8
  },
9
  "cls_token": {
10
  "content": "<s>",
11
  "lstrip": false,
12
- "normalized": true,
13
  "rstrip": false,
14
  "single_word": false
15
  },
16
  "eos_token": {
17
  "content": "</s>",
18
  "lstrip": false,
19
- "normalized": true,
20
  "rstrip": false,
21
  "single_word": false
22
  },
23
  "mask_token": {
24
  "content": "<mask>",
25
  "lstrip": true,
26
- "normalized": true,
27
  "rstrip": false,
28
  "single_word": false
29
  },
30
  "pad_token": {
31
  "content": "<pad>",
32
  "lstrip": false,
33
- "normalized": true,
34
  "rstrip": false,
35
  "single_word": false
36
  },
37
  "sep_token": {
38
  "content": "</s>",
39
  "lstrip": false,
40
- "normalized": true,
41
  "rstrip": false,
42
  "single_word": false
43
  },
44
  "unk_token": {
45
  "content": "<unk>",
46
  "lstrip": false,
47
- "normalized": true,
48
  "rstrip": false,
49
  "single_word": false
50
  }
 
2
  "bos_token": {
3
  "content": "<s>",
4
  "lstrip": false,
5
+ "normalized": false,
6
  "rstrip": false,
7
  "single_word": false
8
  },
9
  "cls_token": {
10
  "content": "<s>",
11
  "lstrip": false,
12
+ "normalized": false,
13
  "rstrip": false,
14
  "single_word": false
15
  },
16
  "eos_token": {
17
  "content": "</s>",
18
  "lstrip": false,
19
+ "normalized": false,
20
  "rstrip": false,
21
  "single_word": false
22
  },
23
  "mask_token": {
24
  "content": "<mask>",
25
  "lstrip": true,
26
+ "normalized": false,
27
  "rstrip": false,
28
  "single_word": false
29
  },
30
  "pad_token": {
31
  "content": "<pad>",
32
  "lstrip": false,
33
+ "normalized": false,
34
  "rstrip": false,
35
  "single_word": false
36
  },
37
  "sep_token": {
38
  "content": "</s>",
39
  "lstrip": false,
40
+ "normalized": false,
41
  "rstrip": false,
42
  "single_word": false
43
  },
44
  "unk_token": {
45
  "content": "<unk>",
46
  "lstrip": false,
47
+ "normalized": false,
48
  "rstrip": false,
49
  "single_word": false
50
  }
tokenizer.json CHANGED
@@ -23,7 +23,7 @@
23
  "single_word": false,
24
  "lstrip": false,
25
  "rstrip": false,
26
- "normalized": true,
27
  "special": true
28
  },
29
  {
@@ -32,7 +32,7 @@
32
  "single_word": false,
33
  "lstrip": false,
34
  "rstrip": false,
35
- "normalized": true,
36
  "special": true
37
  },
38
  {
@@ -41,7 +41,7 @@
41
  "single_word": false,
42
  "lstrip": false,
43
  "rstrip": false,
44
- "normalized": true,
45
  "special": true
46
  },
47
  {
@@ -50,7 +50,7 @@
50
  "single_word": false,
51
  "lstrip": false,
52
  "rstrip": false,
53
- "normalized": true,
54
  "special": true
55
  },
56
  {
@@ -59,7 +59,7 @@
59
  "single_word": false,
60
  "lstrip": true,
61
  "rstrip": false,
62
- "normalized": true,
63
  "special": true
64
  }
65
  ],
 
23
  "single_word": false,
24
  "lstrip": false,
25
  "rstrip": false,
26
+ "normalized": false,
27
  "special": true
28
  },
29
  {
 
32
  "single_word": false,
33
  "lstrip": false,
34
  "rstrip": false,
35
+ "normalized": false,
36
  "special": true
37
  },
38
  {
 
41
  "single_word": false,
42
  "lstrip": false,
43
  "rstrip": false,
44
+ "normalized": false,
45
  "special": true
46
  },
47
  {
 
50
  "single_word": false,
51
  "lstrip": false,
52
  "rstrip": false,
53
+ "normalized": false,
54
  "special": true
55
  },
56
  {
 
59
  "single_word": false,
60
  "lstrip": true,
61
  "rstrip": false,
62
+ "normalized": false,
63
  "special": true
64
  }
65
  ],
tokenizer_config.json CHANGED
@@ -4,7 +4,7 @@
4
  "0": {
5
  "content": "<s>",
6
  "lstrip": false,
7
- "normalized": true,
8
  "rstrip": false,
9
  "single_word": false,
10
  "special": true
@@ -12,7 +12,7 @@
12
  "1": {
13
  "content": "<pad>",
14
  "lstrip": false,
15
- "normalized": true,
16
  "rstrip": false,
17
  "single_word": false,
18
  "special": true
@@ -20,7 +20,7 @@
20
  "2": {
21
  "content": "</s>",
22
  "lstrip": false,
23
- "normalized": true,
24
  "rstrip": false,
25
  "single_word": false,
26
  "special": true
@@ -28,7 +28,7 @@
28
  "3": {
29
  "content": "<unk>",
30
  "lstrip": false,
31
- "normalized": true,
32
  "rstrip": false,
33
  "single_word": false,
34
  "special": true
@@ -36,7 +36,7 @@
36
  "50264": {
37
  "content": "<mask>",
38
  "lstrip": true,
39
- "normalized": true,
40
  "rstrip": false,
41
  "single_word": false,
42
  "special": true
@@ -45,7 +45,6 @@
45
  "bos_token": "<s>",
46
  "clean_up_tokenization_spaces": true,
47
  "cls_token": "<s>",
48
- "do_lower_case": false,
49
  "eos_token": "</s>",
50
  "errors": "replace",
51
  "full_tokenizer_file": null,
 
4
  "0": {
5
  "content": "<s>",
6
  "lstrip": false,
7
+ "normalized": false,
8
  "rstrip": false,
9
  "single_word": false,
10
  "special": true
 
12
  "1": {
13
  "content": "<pad>",
14
  "lstrip": false,
15
+ "normalized": false,
16
  "rstrip": false,
17
  "single_word": false,
18
  "special": true
 
20
  "2": {
21
  "content": "</s>",
22
  "lstrip": false,
23
+ "normalized": false,
24
  "rstrip": false,
25
  "single_word": false,
26
  "special": true
 
28
  "3": {
29
  "content": "<unk>",
30
  "lstrip": false,
31
+ "normalized": false,
32
  "rstrip": false,
33
  "single_word": false,
34
  "special": true
 
36
  "50264": {
37
  "content": "<mask>",
38
  "lstrip": true,
39
+ "normalized": false,
40
  "rstrip": false,
41
  "single_word": false,
42
  "special": true
 
45
  "bos_token": "<s>",
46
  "clean_up_tokenization_spaces": true,
47
  "cls_token": "<s>",
 
48
  "eos_token": "</s>",
49
  "errors": "replace",
50
  "full_tokenizer_file": null,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9a71c738e871b491ece43589e807058ccb914178356e8f892c27c32939c8abfd
3
  size 4920
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:13194c1152585246ae754e11f2fb5200031f2cac7510c99af0d724b84a45893f
3
  size 4920