dd123 commited on
Commit
2728779
1 Parent(s): 54b5d6d

Training in progress, epoch 1

Browse files
config.json CHANGED
@@ -1,13 +1,14 @@
1
  {
2
- "_name_or_path": "distilbert-base-uncased",
3
- "activation": "gelu",
4
  "architectures": [
5
- "DistilBertForSequenceClassification"
6
  ],
7
- "attention_dropout": 0.1,
8
- "dim": 768,
9
- "dropout": 0.1,
10
- "hidden_dim": 3072,
 
 
11
  "id2label": {
12
  "0": "none",
13
  "1": "like",
@@ -22,6 +23,7 @@
22
  "9": "meme"
23
  },
24
  "initializer_range": 0.02,
 
25
  "label2id": {
26
  "express_abashed": "10",
27
  "express_sad": "7",
@@ -35,17 +37,17 @@
35
  "tease": "8",
36
  "unlike": "2"
37
  },
 
38
  "max_position_embeddings": 512,
39
- "model_type": "distilbert",
40
- "n_heads": 12,
41
- "n_layers": 6,
42
  "pad_token_id": 0,
 
43
  "problem_type": "single_label_classification",
44
- "qa_dropout": 0.1,
45
- "seq_classif_dropout": 0.2,
46
- "sinusoidal_pos_embds": false,
47
- "tie_weights_": true,
48
  "torch_dtype": "float32",
49
  "transformers_version": "4.27.1",
 
 
50
  "vocab_size": 30522
51
  }
 
1
  {
2
+ "_name_or_path": "bert-base-uncased",
 
3
  "architectures": [
4
+ "BertForSequenceClassification"
5
  ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "gradient_checkpointing": false,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
  "id2label": {
13
  "0": "none",
14
  "1": "like",
 
23
  "9": "meme"
24
  },
25
  "initializer_range": 0.02,
26
+ "intermediate_size": 3072,
27
  "label2id": {
28
  "express_abashed": "10",
29
  "express_sad": "7",
 
37
  "tease": "8",
38
  "unlike": "2"
39
  },
40
+ "layer_norm_eps": 1e-12,
41
  "max_position_embeddings": 512,
42
+ "model_type": "bert",
43
+ "num_attention_heads": 12,
44
+ "num_hidden_layers": 12,
45
  "pad_token_id": 0,
46
+ "position_embedding_type": "absolute",
47
  "problem_type": "single_label_classification",
 
 
 
 
48
  "torch_dtype": "float32",
49
  "transformers_version": "4.27.1",
50
+ "type_vocab_size": 2,
51
+ "use_cache": true,
52
  "vocab_size": 30522
53
  }
logs/1686802202.292625/events.out.tfevents.1686802202.ls.3992184.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f96485e654ffb97b19c331f55e4fd866235ff4d15f194353c294d318ad4e32f5
3
+ size 5791
logs/events.out.tfevents.1686801844.ls.3984103.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4c313a1e38b81f8aeb2e09835ef4ef622461aa005c6d2dcc470ca9531138cf6c
3
- size 5439
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:52897bfb716fb465c7102bec650bf08b6da48267e94d6589d3ad6aeb320ea5ce
3
+ size 7529
logs/events.out.tfevents.1686802202.ls.3992184.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:378b25914e6d91b230d41b96a82c2c26c51728a271f30aa6bfc87539a586c549
3
+ size 4920
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4ce56f2d79f1e9f87740a5dd5e5b2cbac4f6124ca346c449b7e8a2a6017bb70b
3
- size 267883181
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4a4b6be2308653aa6d28764d1bdf1ef5f10d92e97471737b5b0bd1bd1ac6131d
3
+ size 438035573
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:758effd6e8891c593a864281f10a42e9c315404096144e88f495c543577d6278
3
  size 3579
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b869e18eefa4186d81617a24c9d960f9e705a743f15586a5ab46a5ae640402fa
3
  size 3579