deandrasetya commited on
Commit
5cdcd73
1 Parent(s): acdf52f

Upload TFBertForSequenceClassification

Browse files
Files changed (3) hide show
  1. README.md +44 -6
  2. config.json +1 -5
  3. tf_model.h5 +2 -2
README.md CHANGED
@@ -1,19 +1,57 @@
1
  ---
2
  license: mit
 
 
3
  model-index:
4
  - name: indobert-abusive-language-classifier
5
  results: []
6
- tags:
7
- - text-classification
8
- - bert
9
- - indobert
10
  ---
11
 
 
 
 
12
  # indobert-abusive-language-classifier
13
 
 
 
 
 
 
 
 
14
 
15
  ## Model description
16
 
17
- This model is an abusive language classifier for Bahasa Indonesia based on [indolem/indobert-base-uncased](https://huggingface.co/indolem/indobert-base-uncased).
18
- It was trained on a modified [Multi-label Hate Speech and Abusive Language Detection](https://github.com/okkyibrohim/id-multi-label-hate-speech-and-abusive-language-detection) dataset and achieves an evaluation accuracy of 90.17%.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
19
 
 
 
 
 
 
1
  ---
2
  license: mit
3
+ tags:
4
+ - generated_from_keras_callback
5
  model-index:
6
  - name: indobert-abusive-language-classifier
7
  results: []
 
 
 
 
8
  ---
9
 
10
+ <!-- This model card has been generated automatically according to the information Keras had access to. You should
11
+ probably proofread and complete it, then remove this comment. -->
12
+
13
  # indobert-abusive-language-classifier
14
 
15
+ This model is a fine-tuned version of [indolem/indobert-base-uncased](https://huggingface.co/indolem/indobert-base-uncased) on an unknown dataset.
16
+ It achieves the following results on the evaluation set:
17
+ - Train Loss: 0.1613
18
+ - Train Sparse Categorical Accuracy: 0.9417
19
+ - Validation Loss: 0.2973
20
+ - Validation Sparse Categorical Accuracy: 0.8857
21
+ - Epoch: 2
22
 
23
  ## Model description
24
 
25
+ More information needed
26
+
27
+ ## Intended uses & limitations
28
+
29
+ More information needed
30
+
31
+ ## Training and evaluation data
32
+
33
+ More information needed
34
+
35
+ ## Training procedure
36
+
37
+ ### Training hyperparameters
38
+
39
+ The following hyperparameters were used during training:
40
+ - optimizer: {'name': 'Adam', 'learning_rate': 5e-05, 'decay': 0.0, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-07, 'amsgrad': False}
41
+ - training_precision: float32
42
+
43
+ ### Training results
44
+
45
+ | Train Loss | Train Sparse Categorical Accuracy | Validation Loss | Validation Sparse Categorical Accuracy | Epoch |
46
+ |:----------:|:---------------------------------:|:---------------:|:--------------------------------------:|:-----:|
47
+ | 0.4496 | 0.7811 | 0.3146 | 0.8671 | 0 |
48
+ | 0.2437 | 0.9026 | 0.2959 | 0.8888 | 1 |
49
+ | 0.1613 | 0.9417 | 0.2973 | 0.8857 | 2 |
50
+
51
+
52
+ ### Framework versions
53
 
54
+ - Transformers 4.30.2
55
+ - TensorFlow 2.12.0
56
+ - Datasets 2.13.1
57
+ - Tokenizers 0.13.3
config.json CHANGED
@@ -3,10 +3,6 @@
3
  "architectures": [
4
  "BertForSequenceClassification"
5
  ],
6
- "id2label": {
7
- "0": "SAFE",
8
- "1": "ABUSIVE"
9
- },
10
  "attention_probs_dropout_prob": 0.1,
11
  "bos_token_id": 0,
12
  "classifier_dropout": null,
@@ -24,7 +20,7 @@
24
  "output_past": true,
25
  "pad_token_id": 0,
26
  "position_embedding_type": "absolute",
27
- "transformers_version": "4.26.1",
28
  "type_vocab_size": 2,
29
  "use_cache": true,
30
  "vocab_size": 31923
 
3
  "architectures": [
4
  "BertForSequenceClassification"
5
  ],
 
 
 
 
6
  "attention_probs_dropout_prob": 0.1,
7
  "bos_token_id": 0,
8
  "classifier_dropout": null,
 
20
  "output_past": true,
21
  "pad_token_id": 0,
22
  "position_embedding_type": "absolute",
23
+ "transformers_version": "4.30.2",
24
  "type_vocab_size": 2,
25
  "use_cache": true,
26
  "vocab_size": 31923
tf_model.h5 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ac3258a051f140c5177d259514901171b920f1afc1c9e405a59a1999315f82cf
3
- size 442527064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e3bc34ac5ee91dfa5142272b62d098cfa9f480ffbc329ddafd22d2674b4ffc4f
3
+ size 442527000