Text Generation
English
telmo000's picture
Add sample working space
6a1a405
---
datasets:
- SALT-NLP/positive_reframing
language:
- en
license: bigscience-bloom-rail-1.0
pipeline_tag: text-generation
---
# Model Card for Model ID
This model is a BLOOM-base adjusted to the sentiment transfer task, developed as part of a FourthBrain workshop on GenerativeAI.
## Model Details
### Model Description
This model is a BLOOM-base adjusted to the sentiment transfer task, where the objective is to reverse the sentiment polarity of a text without contradicting
the original meaning. Positive reframing induces a complementary positive viewpoint (e.g. glass-half-full) escaping negative patterns.
Based on the article [Ziems at. al (2022)](https://arxiv.org/abs/2204.02952).
Sample working space [here](https://huggingface.co/spaces/telmo000/bloom-positive-reframing).
### Input
`### Negative sentence:\n{original_text}\n\n### Reframing strategy: \n{reframing_strategy}\n\n### Reframing sentence:\n`
- **Developed by:** Telmo Correa
- **Model type:** LLM
- **Language(s) (NLP):** English
- **License:** [bigscience-bloom-rail-1.0](https://bigscience.huggingface.co/blog/the-bigscience-rail-license)
- **Finetuned from model :** [bigscience/bloom-1b7](https://huggingface.co/bigscience/bloom-1b7)
## Uses
Model trained as a proof-of-concept fine tuning of BLOOM for sentence rewriting.
### Direct Use
Model is intended to be directly used to rewrite sentences with the provided strategy.
### Out-of-Scope Use
Any uses of the model must abide by the terms of both the original BLOOM model and the Salt-NLP/positive-reframing dataset.
## Bias, Risks, and Limitations
As a fine-tuned version of BLOOM, this model carries all the biases, risks, and limitations. of its original training.
## Training Details
### Training Data
[Salt-NLP/positive-reframing](https://huggingface.co/datasets/SALT-NLP/positive_reframing)
### Training Procedure
The baseline model [bigscience/bloom-1b7](https://huggingface.co/bigscience/bloom-1b7) was trained through 100 steps over the training split of the training data, with its prompt engineered to request explicit positive sentence reframing:
```
Below is a negative sentence, please select a reframing strategy and write the positive reframed sentence.
### Negative sentence:
NEGATIVE SENTENCE HERE
### Reframing strategy:
STRATEGY HERE
### Reframed sentence:
REFRAMED SENTENCE HERE
```
#### Training Hyperparameters
- **Training regime:** fp16 non-mixed precision, using PEFT and LoRA
## Evaluation
Evaluation not performed.
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** Coogle Colab PRO GPU
- **Hours used:** 10 min
- **Cloud Provider:** GCP
- **Compute Region:** us-west-1
- **Carbon Emitted:** 10g