Zakia's picture
Updated the examples sections of the model card
de3f434
---
license: apache-2.0
datasets:
- Zakia/drugscom_reviews
language:
- en
metrics:
- training loss
library_name: transformers
pipeline_tag: text-generation
tags:
- health
- medicine
- patient reviews
- drug reviews
- depression
- text generation
widget:
- text: After starting this new treatment, I felt
example_title: Example 1
- text: I was apprehensive about the side effects of
example_title: Example 2
- text: This medication has changed my life for the better
example_title: Example 3
- text: I've had a terrible experience with this medication
example_title: Example 4
- text: Since I began taking L-methylfolate, my experience has been
example_title: Example 5
---
# Model Card for Zakia/gpt2-drugscom_depression_reviews
This model is a GPT-2-based language model fine-tuned on drug reviews for the depression medical condition from Drugs.com.
The dataset used for fine-tuning is the [Zakia/drugscom_reviews](https://huggingface.co/datasets/Zakia/drugscom_reviews) dataset, which is filtered for the condition 'Depression'.
The base model for fine-tuning was the [gpt2](https://huggingface.co/gpt2).
## Model Details
### Model Description
- Developed by: [Zakia](https://huggingface.co/Zakia)
- Model type: Text Generation
- Language(s) (NLP): English
- License: Apache 2.0
- Finetuned from model: gpt2
## Uses
### Direct Use
This model is intended to generate text that mimics patient reviews of depression medications, useful for understanding patient sentiments and experiences.
### Out-of-Scope Use
This model is not designed to diagnose or treat depression or to replace professional medical advice.
## Bias, Risks, and Limitations
The model may inherit biases present in the dataset and should be used with caution in decision-making processes.
### Recommendations
Use the model as a tool for generating synthetic patient reviews and for NLP research.
## How to Get Started with the Model
Use the code below to generate synthetic reviews with the model.
```python
from transformers import GPT2LMHeadModel, GPT2Tokenizer
import torch
model_name = "Zakia/gpt2-drugscom_depression_reviews"
model = GPT2LMHeadModel.from_pretrained(model_name)
tokenizer = GPT2Tokenizer.from_pretrained(model_name)
# Function to generate text
def generate_review(prompt, model, tokenizer):
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs)
return tokenizer.decode(outputs[0], skip_special_tokens=True)
# Example usage for various scenarios
prompts = [
"After starting this new treatment, I felt",
"I was apprehensive about the side effects of",
"This medication has changed my life for the better",
"I've had a terrible experience with this medication",
"Since I began taking L-methylfolate, my experience has been"
]
for prompt in prompts:
print(f"Prompt: {prompt}")
print(generate_review(prompt, model, tokenizer))
print()
```
## Training Details
### Training Data
The model was fine-tuned on patient reviews related to depression, filtered from Drugs.com.
This dataset is accessible from [Zakia/drugscom_reviews](https://huggingface.co/datasets/Zakia/drugscom_reviews) on Hugging Face datasets (condition = 'Depression') for 'train'.
Number of records in train dataset: 9069 rows.
### Training Procedure
#### Preprocessing
The reviews were cleaned and preprocessed to remove quotes, HTML tags and decode HTML entities.
#### Training Hyperparameters
- Batch Size: 2
- Epochs: 5
## Evaluation
- Training Loss
#### Metrics
The model's performance was evaluated based on Training Loss.
### Results
The fine-tuning process yielded the following results:
| Epoch | Training Loss | Training Runtime | Training Samples | Training Samples per Second | Training Steps per Second |
|-------|---------------|------------------|------------------|-----------------------------|---------------------------|
| 5.0 | 0.5944 | 2:15:40.11 | 4308 | 2.646 | 1.323 |
The fine-tuning process achieved a final training loss of 0.5944 after 5 epochs, with the model processing
approximately 2.646 samples per second and completing 1.323 training steps per second over a training runtime
of 2 hours, 15 minutes, and 40 seconds.
## Technical Specifications
### Model Architecture and Objective
GPT-2 model architecture was used, with the objective of generating coherent and contextually relevant text based on patient reviews.
### Compute Infrastructure
The model was trained using a T4 GPU on Google Colab.
#### Hardware
T4 GPU via Google Colab.
## Citation
If you use this model, please cite the original GPT-2 paper:
**BibTeX:**
```bibtex
@article{radford2019language,
title={Language Models are Unsupervised Multitask Learners},
author={Radford, Alec and others},
year={2019}
}
```
**APA:**
Radford, A., et al. (2019). Language Models are Unsupervised Multitask Learners.
## More Information
For further queries or issues with the model, please use the [discussions section on this model's Hugging Face page](https://huggingface.co/Zakia/gpt2-drugscom_depression_reviews/discussions).
## Model Card Authors
- [Zakia](https://huggingface.co/Zakia)
## Model Card Contact
For more information or inquiries regarding this model, please use the [discussions section on this model's Hugging Face page](https://huggingface.co/Zakia/gpt2-drugscom_depression_reviews/discussions).