LukeGPT88 commited on
Commit
5fe3c89
1 Parent(s): 5ee0d49

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +68 -11
README.md CHANGED
@@ -1,21 +1,31 @@
1
  ---
2
- license: apache-2.0
3
- base_model: distilbert-base-uncased
4
  tags:
5
- - generated_from_trainer
6
- metrics:
7
- - accuracy
8
- model-index:
9
- - name: distilbert-base-uncased-finetuned-text-classification
10
- results: []
 
 
 
11
  ---
12
 
13
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
14
- should probably proofread and complete it, then remove this comment. -->
15
-
16
  # distilbert-base-uncased-finetuned-text-classification
17
 
18
  This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on an unknown dataset.
 
 
 
 
 
 
 
 
 
 
 
19
  It achieves the following results on the evaluation set:
20
  - Loss: 0.0501
21
  - Accuracy: 0.9861
@@ -59,3 +69,50 @@ The following hyperparameters were used during training:
59
  - Pytorch 2.1.2
60
  - Datasets 2.1.0
61
  - Tokenizers 0.15.1
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ language: "en"
 
3
  tags:
4
+ - distilbert-base-uncased
5
+ - text-classification
6
+ - patient
7
+ - doctor
8
+
9
+ widget:
10
+ - text: "I've got flu"
11
+ - text: "I prescribe you some drugs and you need to stay at home for a couple of days"
12
+ - text: "Let's move to the theatre this evening!"
13
  ---
14
 
 
 
 
15
  # distilbert-base-uncased-finetuned-text-classification
16
 
17
  This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on an unknown dataset.
18
+
19
+ # Fine-tuned DistilBERT-base-uncased for Patient-Doctor Classification
20
+
21
+ # Model Description
22
+
23
+ DistilBERT is a transformer model that performs text classification. I fine-tuned the model on with the purpose of classifying patient, doctor or neutral content, specifically when text is related to the supposed context. The model predicts 3 classes, which are Patient, Doctor or Neutral.
24
+
25
+ The model is a fine-tuned version of [DistilBERT](https://huggingface.co/docs/transformers/model_doc/distilbert).
26
+
27
+ It was fine-tuned on the prepared dataset (https://huggingface.co/datasets/LukeGPT88/text-classification-dataset).
28
+
29
  It achieves the following results on the evaluation set:
30
  - Loss: 0.0501
31
  - Accuracy: 0.9861
 
69
  - Pytorch 2.1.2
70
  - Datasets 2.1.0
71
  - Tokenizers 0.15.1
72
+
73
+ ---
74
+ language: "en"
75
+ tags:
76
+ - distilroberta
77
+ - sentiment
78
+ - NSFW
79
+ - inappropriate
80
+ - spam
81
+ - twitter
82
+ - reddit
83
+
84
+ widget:
85
+ - text: "I like you. You remind me of me when I was young and stupid."
86
+ - text: "I see you’ve set aside this special time to humiliate yourself in public."
87
+ - text: "Have a great weekend! See you next week!"
88
+
89
+ ---
90
+
91
+ # Fine-tuned DistilRoBERTa-base for NSFW Classification
92
+
93
+ # Model Description
94
+
95
+ DistilBERT is a transformer model that performs sentiment analysis. I fine-tuned the model on Reddit posts with the purpose of classifying not safe for work (NSFW) content, specifically text that is considered inappropriate and unprofessional. The model predicts 2 classes, which are NSFW or safe for work (SFW).
96
+
97
+ The model is a fine-tuned version of [DistilBERT](https://huggingface.co/docs/transformers/model_doc/distilbert).
98
+
99
+ It was fine-tuned on 14317 Reddit posts pulled from the (Reddit API) [https://praw.readthedocs.io/en/stable/].
100
+
101
+ # How to Use
102
+
103
+ ```python
104
+ from transformers import pipeline
105
+ classifier = pipeline("sentiment-analysis", model="michellejieli/NSFW_text_classification")
106
+ classifier("I see you’ve set aside this special time to humiliate yourself in public.")
107
+ ```
108
+
109
+ ```python
110
+ Output:
111
+ [{'label': 'NSFW', 'score': 0.998853325843811}]
112
+ ```
113
+
114
+ # Contact
115
+
116
+ Please reach out to [luca.flammia@gmail.com](luca.flammia@gmail.com) if you have any questions or feedback.
117
+
118
+ ---