Text Generation
English
telmo000 commited on
Commit
6bf55d2
1 Parent(s): 8a12990

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +87 -0
README.md ADDED
@@ -0,0 +1,87 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ datasets:
3
+ - SALT-NLP/positive_reframing
4
+ language:
5
+ - en
6
+ license: bigscience-bloom-rail-1.0
7
+ ---
8
+
9
+ # Model Card for Model ID
10
+
11
+ This model is a BLOOM-base adjusted to the sentiment transfer task, developed as part of a FourthBrain workshop on GenerativeAI.
12
+
13
+ ## Model Details
14
+
15
+ ### Model Description
16
+
17
+ This model is a BLOOM-base adjusted to the sentiment transfer task, where the objective is to reverse the sentiment polarity of a text without contradicting the original meaning. Positive reframing induces a complementary positive viewpoint (e.g. glass-half-full) escaping negative patterns. Based on the article arXiv:2204.02952.
18
+
19
+ ### Input
20
+ `### Negative sentence:\n{original_text}\n\n### Reframing strategy: \n{reframing_strategy}\n\n### Reframing sentence:\n`
21
+
22
+
23
+ - **Developed by:** Telmo Correa
24
+ - **Model type:** LLM
25
+ - **Language(s) (NLP):** English
26
+ - **License:** [bigscience-bloom-rail-1.0](https://bigscience.huggingface.co/blog/the-bigscience-rail-license)
27
+ - **Finetuned from model :** [https://huggingface.co/bigscience/bloom-7b1](bigscience/bloom-1b7)
28
+
29
+ ## Uses
30
+
31
+ Model trained as a proof-of-concept fine tuning of BLOOM for sentence rewriting.
32
+
33
+ ### Direct Use
34
+
35
+ Model is intended to be directly used to rewrite sentences with the provided strategy.
36
+
37
+ ### Out-of-Scope Use
38
+
39
+ Any uses of the model must abide by the terms of both the original BLOOM model and the Salt-NLP/positive-reframing dataset.
40
+
41
+ ## Bias, Risks, and Limitations
42
+
43
+ As a fine-tuned version of BLOOM, this model carries all the biases, risks, and limitations. of its original training.
44
+
45
+ ## Training Details
46
+
47
+ ### Training Data
48
+
49
+ [Salt-NLP/positive-reframing](https://huggingface.co/datasets/SALT-NLP/positive_reframing)
50
+
51
+ ### Training Procedure
52
+
53
+ The baseline model [https://huggingface.co/bigscience/bloom-7b1](bigscience/bloom-1b7) was trained through 100 steps over the training split of the training data, with its prompt engineered to request explicit positive sentence reframing:
54
+
55
+ ```
56
+ Below is a negative sentence, please select a reframing strategy and write the positive reframed sentence.
57
+
58
+ ### Negative sentence:
59
+ NEGATIVE SENTENCE HERE
60
+
61
+ ### Reframing strategy:
62
+ STRATEGY HERE
63
+
64
+ ### Reframed sentence:
65
+ REFRAMED SENTENCE HERE
66
+ ```
67
+
68
+
69
+ #### Training Hyperparameters
70
+
71
+ - **Training regime:** fp16 non-mixed precision, using PEFT and LoRA
72
+
73
+ ## Evaluation
74
+
75
+ Evaluation not performed.
76
+
77
+ ## Environmental Impact
78
+
79
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
80
+
81
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
82
+
83
+ - **Hardware Type:** Coogle Colab PRO GPU
84
+ - **Hours used:** 10 min
85
+ - **Cloud Provider:** GCP
86
+ - **Compute Region:** us-west-1
87
+ - **Carbon Emitted:** 10g