Sandipan1994 commited on
Commit
9542c80
1 Parent(s): 5d7bab1

Training in progress, step 500

Browse files
config.json CHANGED
@@ -1,45 +1,39 @@
1
  {
2
- "_name_or_path": "cross-encoder/nli-deberta-v3-large",
 
3
  "architectures": [
4
- "DebertaV2ForSequenceClassification"
5
  ],
6
  "attention_probs_dropout_prob": 0.1,
 
 
 
7
  "hidden_act": "gelu",
8
  "hidden_dropout_prob": 0.1,
9
  "hidden_size": 1024,
10
  "id2label": {
11
- "0": "contradiction",
12
- "1": "entailment",
13
- "2": "neutral"
14
  },
15
  "initializer_range": 0.02,
16
  "intermediate_size": 4096,
17
  "label2id": {
18
- "contradiction": 0,
19
- "entailment": 1,
20
- "neutral": 2
21
  },
22
- "layer_norm_eps": 1e-07,
23
- "max_position_embeddings": 512,
24
- "max_relative_positions": -1,
25
- "model_type": "deberta-v2",
26
- "norm_rel_ebd": "layer_norm",
27
  "num_attention_heads": 16,
28
  "num_hidden_layers": 24,
29
- "pad_token_id": 0,
30
- "pooler_dropout": 0,
31
- "pooler_hidden_act": "gelu",
32
- "pooler_hidden_size": 1024,
33
- "pos_att_type": [
34
- "p2c",
35
- "c2p"
36
- ],
37
- "position_biased_input": false,
38
- "position_buckets": 256,
39
- "relative_attention": true,
40
- "share_att_key": true,
41
  "torch_dtype": "float32",
42
  "transformers_version": "4.25.1",
43
- "type_vocab_size": 0,
44
- "vocab_size": 128100
 
45
  }
 
1
  {
2
+ "_name_or_path": "roberta-large-mnli",
3
+ "_num_labels": 3,
4
  "architectures": [
5
+ "RobertaForSequenceClassification"
6
  ],
7
  "attention_probs_dropout_prob": 0.1,
8
+ "bos_token_id": 0,
9
+ "classifier_dropout": null,
10
+ "eos_token_id": 2,
11
  "hidden_act": "gelu",
12
  "hidden_dropout_prob": 0.1,
13
  "hidden_size": 1024,
14
  "id2label": {
15
+ "0": "CONTRADICTION",
16
+ "1": "NEUTRAL",
17
+ "2": "ENTAILMENT"
18
  },
19
  "initializer_range": 0.02,
20
  "intermediate_size": 4096,
21
  "label2id": {
22
+ "CONTRADICTION": 0,
23
+ "ENTAILMENT": 2,
24
+ "NEUTRAL": 1
25
  },
26
+ "layer_norm_eps": 1e-05,
27
+ "max_position_embeddings": 514,
28
+ "model_type": "roberta",
 
 
29
  "num_attention_heads": 16,
30
  "num_hidden_layers": 24,
31
+ "pad_token_id": 1,
32
+ "position_embedding_type": "absolute",
33
+ "problem_type": "single_label_classification",
 
 
 
 
 
 
 
 
 
34
  "torch_dtype": "float32",
35
  "transformers_version": "4.25.1",
36
+ "type_vocab_size": 1,
37
+ "use_cache": true,
38
+ "vocab_size": 50265
39
  }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ae24e029aa5c6d4fc27f0a765d055f75b8bdf172a30cdd67645af05924c13d60
3
- size 1740400377
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eab7ca94f54f489b969c58d02d327a4259f260e7f67cd6553e006a5e405aa539
3
+ size 1421591285
runs/Dec09_21-28-29_a03200a6b5c6/events.out.tfevents.1670621316.a03200a6b5c6.340.2 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6c9c76ac660ebbb5147831433556821ae8458c8449c6a31bbee153ffadeaf93d
3
- size 4292
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bf3b519c12b97480800b8084123fa7c80d76fde92d898bda1db423dcb7416aca
3
+ size 4606
runs/Dec09_21-49-05_a03200a6b5c6/1670622556.690126/events.out.tfevents.1670622556.a03200a6b5c6.340.5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2e0412c0442b80f99c85bc4dedf7b5601d191736c5fb1db3114d1b679819bbf9
3
+ size 5494
runs/Dec09_21-49-05_a03200a6b5c6/events.out.tfevents.1670622556.a03200a6b5c6.340.4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:752954e1f8d251018c3e75375009e4b0c2c66c5a3a28aee35776a705338f7a75
3
+ size 40
runs/Dec09_21-50-54_a03200a6b5c6/1670622665.1440845/events.out.tfevents.1670622665.a03200a6b5c6.1013.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e8bac52941a42a774dfbdc8a64552600f24b655644eeaaa7c1623ed03a444c87
3
+ size 5494
runs/Dec09_21-50-54_a03200a6b5c6/events.out.tfevents.1670622665.a03200a6b5c6.1013.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4e48ec5b4390d9d7f1c520bfc61341a86b7e8c1a21664d8d4599b379cb637c03
3
+ size 4086
special_tokens_map.json CHANGED
@@ -1,9 +1,15 @@
1
  {
2
- "bos_token": "[CLS]",
3
- "cls_token": "[CLS]",
4
- "eos_token": "[SEP]",
5
- "mask_token": "[MASK]",
6
- "pad_token": "[PAD]",
7
- "sep_token": "[SEP]",
8
- "unk_token": "[UNK]"
 
 
 
 
 
 
9
  }
 
1
  {
2
+ "bos_token": "<s>",
3
+ "cls_token": "<s>",
4
+ "eos_token": "</s>",
5
+ "mask_token": {
6
+ "content": "<mask>",
7
+ "lstrip": true,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false
11
+ },
12
+ "pad_token": "<pad>",
13
+ "sep_token": "</s>",
14
+ "unk_token": "<unk>"
15
  }
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -1,17 +1,16 @@
1
  {
2
- "bos_token": "[CLS]",
3
- "cls_token": "[CLS]",
4
- "do_lower_case": false,
5
- "eos_token": "[SEP]",
6
- "mask_token": "[MASK]",
 
7
  "model_max_length": 512,
8
- "name_or_path": "cross-encoder/nli-deberta-v3-large",
9
- "pad_token": "[PAD]",
10
- "sep_token": "[SEP]",
11
- "sp_model_kwargs": {},
12
  "special_tokens_map_file": null,
13
- "split_by_punct": false,
14
- "tokenizer_class": "DebertaV2Tokenizer",
15
- "unk_token": "[UNK]",
16
- "vocab_type": "spm"
17
  }
 
1
  {
2
+ "add_prefix_space": false,
3
+ "bos_token": "<s>",
4
+ "cls_token": "<s>",
5
+ "eos_token": "</s>",
6
+ "errors": "replace",
7
+ "mask_token": "<mask>",
8
  "model_max_length": 512,
9
+ "name_or_path": "roberta-large-mnli",
10
+ "pad_token": "<pad>",
11
+ "sep_token": "</s>",
 
12
  "special_tokens_map_file": null,
13
+ "tokenizer_class": "RobertaTokenizer",
14
+ "trim_offsets": true,
15
+ "unk_token": "<unk>"
 
16
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fcfbf4f0b8c7852533b42f545f4c02c964dc386cfc55c34deed5e22fdc43711f
3
  size 3387
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7590f2d41797aed6227e5b8de15a7619827e3f5b09b1441d4560384fc53e98f5
3
  size 3387
vocab.json ADDED
The diff for this file is too large to render. See raw diff