fawern commited on
Commit
133f546
1 Parent(s): da49600

Training in progress, epoch 1

Browse files
Files changed (2) hide show
  1. config.json +2 -14
  2. model.safetensors +2 -2
config.json CHANGED
@@ -2,17 +2,12 @@
2
  "_name_or_path": "medicalai/ClinicalBERT",
3
  "activation": "gelu",
4
  "architectures": [
5
- "BertForSequenceClassification"
6
  ],
7
  "attention_dropout": 0.1,
8
- "attention_probs_dropout_prob": 0.1,
9
- "classifier_dropout": null,
10
  "dim": 768,
11
  "dropout": 0.1,
12
- "hidden_act": "gelu",
13
  "hidden_dim": 3072,
14
- "hidden_dropout_prob": 0.1,
15
- "hidden_size": 768,
16
  "id2label": {
17
  "0": "LABEL_0",
18
  "1": "LABEL_1",
@@ -56,7 +51,6 @@
56
  "39": "LABEL_39"
57
  },
58
  "initializer_range": 0.02,
59
- "intermediate_size": 3072,
60
  "label2id": {
61
  "LABEL_0": 0,
62
  "LABEL_1": 1,
@@ -99,16 +93,12 @@
99
  "LABEL_8": 8,
100
  "LABEL_9": 9
101
  },
102
- "layer_norm_eps": 1e-12,
103
  "max_position_embeddings": 512,
104
- "model_type": "bert",
105
  "n_heads": 12,
106
  "n_layers": 6,
107
- "num_attention_heads": 12,
108
- "num_hidden_layers": 12,
109
  "output_past": true,
110
  "pad_token_id": 0,
111
- "position_embedding_type": "absolute",
112
  "problem_type": "single_label_classification",
113
  "qa_dropout": 0.1,
114
  "seq_classif_dropout": 0.2,
@@ -116,7 +106,5 @@
116
  "tie_weights_": true,
117
  "torch_dtype": "float32",
118
  "transformers_version": "4.39.3",
119
- "type_vocab_size": 2,
120
- "use_cache": true,
121
  "vocab_size": 119547
122
  }
 
2
  "_name_or_path": "medicalai/ClinicalBERT",
3
  "activation": "gelu",
4
  "architectures": [
5
+ "DistilBertForSequenceClassification"
6
  ],
7
  "attention_dropout": 0.1,
 
 
8
  "dim": 768,
9
  "dropout": 0.1,
 
10
  "hidden_dim": 3072,
 
 
11
  "id2label": {
12
  "0": "LABEL_0",
13
  "1": "LABEL_1",
 
51
  "39": "LABEL_39"
52
  },
53
  "initializer_range": 0.02,
 
54
  "label2id": {
55
  "LABEL_0": 0,
56
  "LABEL_1": 1,
 
93
  "LABEL_8": 8,
94
  "LABEL_9": 9
95
  },
 
96
  "max_position_embeddings": 512,
97
+ "model_type": "distilbert",
98
  "n_heads": 12,
99
  "n_layers": 6,
 
 
100
  "output_past": true,
101
  "pad_token_id": 0,
 
102
  "problem_type": "single_label_classification",
103
  "qa_dropout": 0.1,
104
  "seq_classif_dropout": 0.2,
 
106
  "tie_weights_": true,
107
  "torch_dtype": "float32",
108
  "transformers_version": "4.39.3",
 
 
109
  "vocab_size": 119547
110
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5c9c95ba60c638a1ba5e39eb70f149fac8f4fd18b4dbf7479434502a0c8b432c
3
- size 711560352
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:56545c6bcbc5d885e4a1c70a642d7e9ab5bed87fb373c5b95d9d2a578608a6a5
3
+ size 541434272