Back to all models
text-generation mask_token:
Query this model
πŸ”₯ This model is currently loaded and running on the Inference API. ⚠️ This model could not be loaded by the inference API. ⚠️ This model can be loaded on the Inference API on-demand.
JSON Output
API endpoint
								$
								curl -X POST \
-H "Authorization: Bearer YOUR_ORG_OR_USER_API_TOKEN" \
-H "Content-Type: application/json" \
-d '"json encoded string"' \
https://api-inference.huggingface.co/models/mrm8488/gpt2-imdb-neg
Share Copied link to clipboard

Monthly model downloads

mrm8488/gpt2-imdb-neg mrm8488/gpt2-imdb-neg
24 downloads
last 30 days

pytorch

tf

Contributed by

mrm8488 Manuel Romero
116 models

How to use this model directly from the πŸ€—/transformers library:

			
Copy to clipboard
from transformers import AutoTokenizer, AutoModelWithLMHead tokenizer = AutoTokenizer.from_pretrained("mrm8488/gpt2-imdb-neg") model = AutoModelWithLMHead.from_pretrained("mrm8488/gpt2-imdb-neg")

GPT2-IMDB-neg (LM + RL) 🎞😑✍

All credits to @lvwerra

What is it?

A small GPT2 (lvwerra/gpt2-imdb) language model fine-tuned to produce negative movie reviews based the IMDB dataset. The model is trained with rewards from a BERT sentiment classifier (lvwerra/gpt2-imdb) via PPO.

Why?

I wanted to reproduce the experiment lvwerra/gpt2-imdb-pos but for generating negative movie reviews.

Training setting

The model was trained for 100 optimisation steps with a batch size of 256 which corresponds to 25600 training samples. The full experiment setup (for positive samples) in trl repo.

Examples

A few examples of the model response to a query before and after optimisation:

query response (before) response (after) rewards (before) rewards (after)
This movie is a fine attempt as far as live action is concerned, n... example of how bad Hollywood in theatrics pla... 2.118391 -3.31625
I have watched 3 episodes with this guy and he is such a talented actor... but the show is just plain awful and there ne... 2.681171 -4.512792
We know that firefighters and police officers are forced to become populari... other chains have going to get this disaster ... 1.367811 -3.34017

Training logs and metrics

Watch the whole training logs and metrics on W&B

Created by Manuel Romero/@mrm8488

Made with in Spain