ajrayman commited on
Commit
0c6cdc5
·
verified ·
1 Parent(s): 262f508

Training in progress, epoch 1

Browse files
Files changed (4) hide show
  1. README.md +13 -12
  2. config.json +5 -5
  3. model.safetensors +2 -2
  4. training_args.bin +1 -1
README.md CHANGED
@@ -1,7 +1,7 @@
1
  ---
2
  library_name: transformers
3
  license: mit
4
- base_model: roberta-base
5
  tags:
6
  - generated_from_trainer
7
  metrics:
@@ -10,22 +10,22 @@ metrics:
10
  - recall
11
  - f1
12
  model-index:
13
- - name: narcissism_binary
14
  results: []
15
  ---
16
 
17
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
18
  should probably proofread and complete it, then remove this comment. -->
19
 
20
- # narcissism_binary
21
 
22
- This model is a fine-tuned version of [roberta-base](https://huggingface.co/roberta-base) on the None dataset.
23
  It achieves the following results on the evaluation set:
24
- - Loss: 0.5930
25
- - Accuracy: 0.6893
26
- - Precision: 0.7057
27
- - Recall: 0.5057
28
- - F1: 0.5892
29
 
30
  ## Model description
31
 
@@ -50,14 +50,15 @@ The following hyperparameters were used during training:
50
  - seed: 42
51
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
52
  - lr_scheduler_type: linear
53
- - num_epochs: 2
54
 
55
  ### Training results
56
 
57
  | Training Loss | Epoch | Step | Validation Loss | Accuracy | Precision | Recall | F1 |
58
  |:-------------:|:-----:|:----:|:---------------:|:--------:|:---------:|:------:|:------:|
59
- | No log | 1.0 | 126 | 0.6328 | 0.6833 | 0.8407 | 0.3469 | 0.4912 |
60
- | No log | 2.0 | 252 | 0.5930 | 0.6893 | 0.7057 | 0.5057 | 0.5892 |
 
61
 
62
 
63
  ### Framework versions
 
1
  ---
2
  library_name: transformers
3
  license: mit
4
+ base_model: roberta-large
5
  tags:
6
  - generated_from_trainer
7
  metrics:
 
10
  - recall
11
  - f1
12
  model-index:
13
+ - name: psychopathy_binary
14
  results: []
15
  ---
16
 
17
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
18
  should probably proofread and complete it, then remove this comment. -->
19
 
20
+ # psychopathy_binary
21
 
22
+ This model is a fine-tuned version of [roberta-large](https://huggingface.co/roberta-large) on the None dataset.
23
  It achieves the following results on the evaluation set:
24
+ - Loss: 0.5869
25
+ - Accuracy: 0.7172
26
+ - Precision: 0.7090
27
+ - Recall: 0.5967
28
+ - F1: 0.6480
29
 
30
  ## Model description
31
 
 
50
  - seed: 42
51
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
52
  - lr_scheduler_type: linear
53
+ - num_epochs: 3
54
 
55
  ### Training results
56
 
57
  | Training Loss | Epoch | Step | Validation Loss | Accuracy | Precision | Recall | F1 |
58
  |:-------------:|:-----:|:----:|:---------------:|:--------:|:---------:|:------:|:------:|
59
+ | No log | 1.0 | 140 | 0.5912 | 0.7011 | 0.7542 | 0.4671 | 0.5769 |
60
+ | No log | 2.0 | 280 | 0.5426 | 0.7127 | 0.7107 | 0.5761 | 0.6364 |
61
+ | No log | 3.0 | 420 | 0.5869 | 0.7172 | 0.7090 | 0.5967 | 0.6480 |
62
 
63
 
64
  ### Framework versions
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "roberta-base",
3
  "architectures": [
4
  "RobertaForSequenceClassification"
5
  ],
@@ -9,14 +9,14 @@
9
  "eos_token_id": 2,
10
  "hidden_act": "gelu",
11
  "hidden_dropout_prob": 0.1,
12
- "hidden_size": 768,
13
  "initializer_range": 0.02,
14
- "intermediate_size": 3072,
15
  "layer_norm_eps": 1e-05,
16
  "max_position_embeddings": 514,
17
  "model_type": "roberta",
18
- "num_attention_heads": 12,
19
- "num_hidden_layers": 12,
20
  "pad_token_id": 1,
21
  "position_embedding_type": "absolute",
22
  "problem_type": "single_label_classification",
 
1
  {
2
+ "_name_or_path": "roberta-large",
3
  "architectures": [
4
  "RobertaForSequenceClassification"
5
  ],
 
9
  "eos_token_id": 2,
10
  "hidden_act": "gelu",
11
  "hidden_dropout_prob": 0.1,
12
+ "hidden_size": 1024,
13
  "initializer_range": 0.02,
14
+ "intermediate_size": 4096,
15
  "layer_norm_eps": 1e-05,
16
  "max_position_embeddings": 514,
17
  "model_type": "roberta",
18
+ "num_attention_heads": 16,
19
+ "num_hidden_layers": 24,
20
  "pad_token_id": 1,
21
  "position_embedding_type": "absolute",
22
  "problem_type": "single_label_classification",
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ed96541eedb00263c914fc3336fb848f786976d191c1a472c38b38812e2c70b0
3
- size 498612824
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7997871a6deb1a46885d008f088cfb54695d8d9c7184e2d04b4ab4c092ecd6b3
3
+ size 1421495416
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7fe09f39a37f7d140755b2dd5c5c4ae797819d308de58dd22a9c654a43318eb1
3
  size 4719
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1d87df5154e0ba6b462f31abee6edcde2f2f8c38f18242f1de72eac3a8bc565f
3
  size 4719