KIZervus commited on
Commit
10cbb4b
1 Parent(s): f9dc484
Files changed (3) hide show
  1. README.md +26 -27
  2. config.json +1 -1
  3. tf_model.h5 +2 -2
README.md CHANGED
@@ -3,55 +3,54 @@ license: apache-2.0
3
  tags:
4
  - generated_from_keras_callback
5
  model-index:
6
- - name: tmp3y468_8j
7
  results: []
8
- widget:
9
- - text: "Ich liebe dich"
10
- example_title: "Non-vulgar"
11
- - text: "Leck mich am arsch"
12
- example_title: "Vulgar"
13
  ---
14
 
15
- # KIZervus
 
16
 
17
- This model is a fine-tuned version of [distilbert-base-german-cased](https://huggingface.co/distilbert-base-german-cased).
18
- It is trained to classify german text into the classes "vulgar" speech and "non-vulgar" speech.
19
- The data set is a collection of other labeled sources in german. For an overview, see the github repository here: https://github.com/NKDataConv/KIZervus
20
- Both data and training procedure are documented in the GitHub repo. Your are welcome to contribute.
21
 
 
22
  It achieves the following results on the evaluation set:
23
- - Train Loss: 0.4221
24
- - Train Accuracy: 0.8025
25
- - Validation Loss: 0.4418
26
- - Validation Accuracy: 0.8094
27
- - Epoch: 2
 
 
 
 
 
 
 
 
 
 
 
 
28
 
29
  ## Training procedure
30
- For details, see the repo and documentation here: https://github.com/NKDataConv/KIZervus
31
 
32
  ### Training hyperparameters
33
 
34
  The following hyperparameters were used during training:
35
- - optimizer: {'name': 'Adam', 'learning_rate': {'class_name': 'PolynomialDecay', 'config': {'initial_learning_rate': 5e-05, 'decay_steps': 1233, 'end_learning_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}}, 'decay': 0.0, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-08, 'amsgrad': False}
36
  - training_precision: float32
37
 
38
  ### Training results
39
 
40
  | Train Loss | Train Accuracy | Validation Loss | Validation Accuracy | Epoch |
41
  |:----------:|:--------------:|:---------------:|:-------------------:|:-----:|
42
- | 0.4524 | 0.7813 | 0.4397 | 0.7969 | 0 |
43
- | 0.4215 | 0.8030 | 0.4838 | 0.7781 | 1 |
44
- | 0.4221 | 0.8025 | 0.4418 | 0.8094 | 2 |
45
 
46
 
47
  ### Framework versions
48
 
49
- - Transformers 4.21.1
50
  - TensorFlow 2.8.2
51
  - Datasets 2.2.2
52
  - Tokenizers 0.12.1
53
-
54
- ### Supporter
55
-
56
- ![BMBF Logo](./BMBF_Logo.png)
57
-
 
3
  tags:
4
  - generated_from_keras_callback
5
  model-index:
6
+ - name: tmpnhxhsble
7
  results: []
 
 
 
 
 
8
  ---
9
 
10
+ <!-- This model card has been generated automatically according to the information Keras had access to. You should
11
+ probably proofread and complete it, then remove this comment. -->
12
 
13
+ # tmpnhxhsble
 
 
 
14
 
15
+ This model is a fine-tuned version of [distilbert-base-german-cased](https://huggingface.co/distilbert-base-german-cased) on an unknown dataset.
16
  It achieves the following results on the evaluation set:
17
+ - Train Loss: 0.4640
18
+ - Train Accuracy: 0.7744
19
+ - Validation Loss: 0.4852
20
+ - Validation Accuracy: 0.7937
21
+ - Epoch: 1
22
+
23
+ ## Model description
24
+
25
+ More information needed
26
+
27
+ ## Intended uses & limitations
28
+
29
+ More information needed
30
+
31
+ ## Training and evaluation data
32
+
33
+ More information needed
34
 
35
  ## Training procedure
 
36
 
37
  ### Training hyperparameters
38
 
39
  The following hyperparameters were used during training:
40
+ - optimizer: {'name': 'Adam', 'learning_rate': {'class_name': 'PolynomialDecay', 'config': {'initial_learning_rate': 5e-05, 'decay_steps': 822, 'end_learning_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}}, 'decay': 0.0, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-08, 'amsgrad': False}
41
  - training_precision: float32
42
 
43
  ### Training results
44
 
45
  | Train Loss | Train Accuracy | Validation Loss | Validation Accuracy | Epoch |
46
  |:----------:|:--------------:|:---------------:|:-------------------:|:-----:|
47
+ | 0.4830 | 0.7617 | 0.5061 | 0.7406 | 0 |
48
+ | 0.4640 | 0.7744 | 0.4852 | 0.7937 | 1 |
 
49
 
50
 
51
  ### Framework versions
52
 
53
+ - Transformers 4.21.2
54
  - TensorFlow 2.8.2
55
  - Datasets 2.2.2
56
  - Tokenizers 0.12.1
 
 
 
 
 
config.json CHANGED
@@ -19,6 +19,6 @@
19
  "seq_classif_dropout": 0.2,
20
  "sinusoidal_pos_embds": true,
21
  "tie_weights_": true,
22
- "transformers_version": "4.21.1",
23
  "vocab_size": 31102
24
  }
 
19
  "seq_classif_dropout": 0.2,
20
  "sinusoidal_pos_embds": true,
21
  "tie_weights_": true,
22
+ "transformers_version": "4.21.2",
23
  "vocab_size": 31102
24
  }
tf_model.h5 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cb2593312c37ef57c22d2f93b17735ceb7c872234e17ed2134b73b98a3322e46
3
- size 269738152
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8362d35f35b804c48d5cd1381a5315ba84d22039e390a3cb7c3ae1bafa36498a
3
+ size 269733656