bob-bob-bob-3 commited on
Commit
c88dcda
1 Parent(s): 5da1bea

End of training

Browse files
README.md CHANGED
@@ -1,6 +1,6 @@
1
  ---
2
- license: mit
3
- base_model: indobenchmark/indobert-large-p2
4
  tags:
5
  - generated_from_trainer
6
  model-index:
@@ -13,12 +13,12 @@ should probably proofread and complete it, then remove this comment. -->
13
 
14
  # buburayam2024_p2_14_asli
15
 
16
- This model is a fine-tuned version of [indobenchmark/indobert-large-p2](https://huggingface.co/indobenchmark/indobert-large-p2) on the None dataset.
17
  It achieves the following results on the evaluation set:
18
- - Loss: 3.0210
19
- - F1 macro: 0.3791
20
- - Weighted: 0.5964
21
- - Balanced accuracy: 0.4977
22
 
23
  ## Model description
24
 
@@ -43,26 +43,19 @@ The following hyperparameters were used during training:
43
  - seed: 42
44
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
  - lr_scheduler_type: linear
46
- - num_epochs: 14
47
 
48
  ### Training results
49
 
50
  | Training Loss | Epoch | Step | Validation Loss | F1 macro | Weighted | Balanced accuracy |
51
  |:-------------:|:-----:|:----:|:---------------:|:--------:|:--------:|:-----------------:|
52
- | 1.1074 | 1.0 | 154 | 1.2377 | 0.3519 | 0.6293 | 0.4548 |
53
- | 0.7157 | 2.0 | 308 | 1.2606 | 0.3927 | 0.5965 | 0.4757 |
54
- | 0.348 | 3.0 | 462 | 1.7488 | 0.4201 | 0.5722 | 0.5450 |
55
- | 0.1433 | 4.0 | 616 | 2.1221 | 0.4589 | 0.6046 | 0.5166 |
56
- | 0.0748 | 5.0 | 770 | 2.4451 | 0.3833 | 0.5914 | 0.5034 |
57
- | 0.0016 | 6.0 | 924 | 2.8787 | 0.3869 | 0.5620 | 0.5239 |
58
- | 0.0004 | 7.0 | 1078 | 2.0919 | 0.4192 | 0.6759 | 0.4867 |
59
- | 0.0003 | 8.0 | 1232 | 2.8603 | 0.3797 | 0.5930 | 0.5068 |
60
- | 0.0485 | 9.0 | 1386 | 2.6217 | 0.3914 | 0.6340 | 0.4941 |
61
- | 0.0002 | 10.0 | 1540 | 3.1652 | 0.3623 | 0.5676 | 0.4882 |
62
- | 0.0002 | 11.0 | 1694 | 3.0986 | 0.3719 | 0.5822 | 0.4951 |
63
- | 0.0002 | 12.0 | 1848 | 3.0331 | 0.3763 | 0.5917 | 0.4968 |
64
- | 0.0002 | 13.0 | 2002 | 3.0254 | 0.3778 | 0.5940 | 0.4973 |
65
- | 0.0001 | 14.0 | 2156 | 3.0210 | 0.3791 | 0.5964 | 0.4977 |
66
 
67
 
68
  ### Framework versions
 
1
  ---
2
+ license: apache-2.0
3
+ base_model: indolem/indobertweet-base-uncased
4
  tags:
5
  - generated_from_trainer
6
  model-index:
 
13
 
14
  # buburayam2024_p2_14_asli
15
 
16
+ This model is a fine-tuned version of [indolem/indobertweet-base-uncased](https://huggingface.co/indolem/indobertweet-base-uncased) on the None dataset.
17
  It achieves the following results on the evaluation set:
18
+ - Loss: 2.3478
19
+ - F1 macro: 0.3466
20
+ - Weighted: 0.5394
21
+ - Balanced accuracy: 0.4857
22
 
23
  ## Model description
24
 
 
43
  - seed: 42
44
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
  - lr_scheduler_type: linear
46
+ - num_epochs: 7
47
 
48
  ### Training results
49
 
50
  | Training Loss | Epoch | Step | Validation Loss | F1 macro | Weighted | Balanced accuracy |
51
  |:-------------:|:-----:|:----:|:---------------:|:--------:|:--------:|:-----------------:|
52
+ | 1.3473 | 1.0 | 93 | 1.4538 | 0.3261 | 0.5837 | 0.4352 |
53
+ | 0.9979 | 2.0 | 186 | 1.4648 | 0.3461 | 0.5112 | 0.4916 |
54
+ | 0.5325 | 3.0 | 279 | 1.8786 | 0.3325 | 0.5012 | 0.4817 |
55
+ | 0.2097 | 4.0 | 372 | 2.2614 | 0.3081 | 0.4708 | 0.4584 |
56
+ | 0.1208 | 5.0 | 465 | 2.0293 | 0.3620 | 0.5624 | 0.4896 |
57
+ | 0.0695 | 6.0 | 558 | 2.4287 | 0.3513 | 0.5322 | 0.4986 |
58
+ | 0.0129 | 7.0 | 651 | 2.3478 | 0.3466 | 0.5394 | 0.4857 |
 
 
 
 
 
 
 
59
 
60
 
61
  ### Framework versions
config.json CHANGED
@@ -1,15 +1,16 @@
1
  {
2
- "_name_or_path": "indobenchmark/indobert-large-p2",
3
- "_num_labels": 5,
4
  "architectures": [
5
  "BertForSequenceClassification"
6
  ],
7
  "attention_probs_dropout_prob": 0.1,
 
8
  "classifier_dropout": null,
9
- "directionality": "bidi",
 
10
  "hidden_act": "gelu",
11
  "hidden_dropout_prob": 0.1,
12
- "hidden_size": 1024,
13
  "id2label": {
14
  "0": "Sumber Daya Alam",
15
  "1": "Politik",
@@ -21,31 +22,19 @@
21
  "7": "Geografi"
22
  },
23
  "initializer_range": 0.02,
24
- "intermediate_size": 4096,
25
- "label2id": {
26
- "LABEL_0": 0,
27
- "LABEL_1": 1,
28
- "LABEL_2": 2,
29
- "LABEL_3": 3,
30
- "LABEL_4": 4
31
- },
32
  "layer_norm_eps": 1e-12,
33
  "max_position_embeddings": 512,
34
  "model_type": "bert",
35
- "num_attention_heads": 16,
36
- "num_hidden_layers": 24,
37
  "output_past": true,
38
  "pad_token_id": 0,
39
- "pooler_fc_size": 768,
40
- "pooler_num_attention_heads": 12,
41
- "pooler_num_fc_layers": 3,
42
- "pooler_size_per_head": 128,
43
- "pooler_type": "first_token_transform",
44
  "position_embedding_type": "absolute",
45
  "problem_type": "single_label_classification",
46
  "torch_dtype": "float32",
47
  "transformers_version": "4.41.2",
48
  "type_vocab_size": 2,
49
  "use_cache": true,
50
- "vocab_size": 30522
51
  }
 
1
  {
2
+ "_name_or_path": "indolem/indobertweet-base-uncased",
 
3
  "architectures": [
4
  "BertForSequenceClassification"
5
  ],
6
  "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": 0,
8
  "classifier_dropout": null,
9
+ "eos_token_ids": 0,
10
+ "gradient_checkpointing": false,
11
  "hidden_act": "gelu",
12
  "hidden_dropout_prob": 0.1,
13
+ "hidden_size": 768,
14
  "id2label": {
15
  "0": "Sumber Daya Alam",
16
  "1": "Politik",
 
22
  "7": "Geografi"
23
  },
24
  "initializer_range": 0.02,
25
+ "intermediate_size": 3072,
 
 
 
 
 
 
 
26
  "layer_norm_eps": 1e-12,
27
  "max_position_embeddings": 512,
28
  "model_type": "bert",
29
+ "num_attention_heads": 12,
30
+ "num_hidden_layers": 12,
31
  "output_past": true,
32
  "pad_token_id": 0,
 
 
 
 
 
33
  "position_embedding_type": "absolute",
34
  "problem_type": "single_label_classification",
35
  "torch_dtype": "float32",
36
  "transformers_version": "4.41.2",
37
  "type_vocab_size": 2,
38
  "use_cache": true,
39
+ "vocab_size": 31923
40
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:af84cf29983426fb96b1b4f406412895d86e8a8227c72cce5008d3383ec7e961
3
- size 1340647360
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:27150179b56d170b0e3b4170d41cdf2416a4b8edbd0bb2fccd887b254fcc7175
3
+ size 442280976
runs/Jun19_18-13-12_80516ee61c48/events.out.tfevents.1718822265.80516ee61c48.721.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3078d3fa058cc35404d1d8eb2917c279e3ed2c75afb453290c2507b8cd23493c
3
+ size 524
runs/Jun19_18-54-03_80516ee61c48/events.out.tfevents.1718823244.80516ee61c48.721.3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:15e404b693842b6bcc6b545d839d01803cc68b94a52bc7bbf20e06e8009393d8
3
+ size 22242
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c48c71a90575444fdb62b1414f5699cdadbf8f76cc9895ca2a8097e3d4d103ea
3
  size 5176
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bd1e83d2019953804d961e797d89dcf8be3d35e7975273b0ed258b96884a5d92
3
  size 5176