Commit
·
ec1b9f9
1
Parent(s):
05b03da
Update README.md
Browse files
README.md
CHANGED
@@ -12,10 +12,10 @@ The model almost achieves SOTA results while being 20x smaller, on a test set co
|
|
12 |
|
13 |
| **Model** | **Precision** | **Recall** | **F1-score** | **F2-score** |
|
14 |
| :-------- | :------------ | :--------- | :----------- | :----------- |
|
15 |
-
| [`alexandrainst/
|
16 |
-
| `alexandrainst/
|
17 |
| [`A&ttack`](https://github.com/ogtal/A-ttack) | **97.32%** | 50.70% | 66.67% | 56.07% |
|
18 |
-
| [`alexandrainst/da-
|
19 |
| [`Guscode/DKbert-hatespeech-detection`](https://huggingface.co/Guscode/DKbert-hatespeech-detection) | 75.41% | 42.79% | 54.60% | 46.84% |
|
20 |
|
21 |
|
@@ -25,7 +25,7 @@ You can use the model simply by running the following:
|
|
25 |
|
26 |
```python
|
27 |
>>> from transformers import pipeline
|
28 |
-
>>> offensive_text_pipeline = pipeline(model="alexandrainst/
|
29 |
>>> offensive_text_pipeline("Din store idiot")
|
30 |
[{'label': 'Offensive', 'score': 0.9997463822364807}]
|
31 |
```
|
|
|
12 |
|
13 |
| **Model** | **Precision** | **Recall** | **F1-score** | **F2-score** |
|
14 |
| :-------- | :------------ | :--------- | :----------- | :----------- |
|
15 |
+
| [`alexandrainst/da-offensive-detection-base`](https://huggingface.co/alexandrainst/danoff-base) | 74.81% | **89.77%** | **81.61%** | **86.32%** |
|
16 |
+
| `alexandrainst/da-offensive-detection-small` (this) | 74.13% | 89.30% | 81.01% | 85.79% |
|
17 |
| [`A&ttack`](https://github.com/ogtal/A-ttack) | **97.32%** | 50.70% | 66.67% | 56.07% |
|
18 |
+
| [`alexandrainst/da-hatespeech-detection-small`](https://huggingface.co/alexandrainst/da-hatespeech-detection-small) | 86.43% | 56.28% | 68.17% | 60.50% |
|
19 |
| [`Guscode/DKbert-hatespeech-detection`](https://huggingface.co/Guscode/DKbert-hatespeech-detection) | 75.41% | 42.79% | 54.60% | 46.84% |
|
20 |
|
21 |
|
|
|
25 |
|
26 |
```python
|
27 |
>>> from transformers import pipeline
|
28 |
+
>>> offensive_text_pipeline = pipeline(model="alexandrainst/da-offensive-detection-small")
|
29 |
>>> offensive_text_pipeline("Din store idiot")
|
30 |
[{'label': 'Offensive', 'score': 0.9997463822364807}]
|
31 |
```
|