justtherightsize
commited on
Commit
•
0171e49
1
Parent(s):
d09a192
Upload README.md
Browse files
README.md
CHANGED
@@ -1,199 +1,94 @@
|
|
1 |
---
|
2 |
-
|
3 |
-
|
4 |
-
|
5 |
-
|
6 |
-
|
7 |
-
|
8 |
-
|
9 |
-
|
10 |
-
|
11 |
-
|
12 |
-
|
13 |
-
|
14 |
-
|
15 |
-
|
16 |
-
|
17 |
-
|
18 |
-
|
19 |
-
|
20 |
-
|
21 |
-
|
22 |
-
|
23 |
-
-
|
24 |
-
-
|
25 |
-
-
|
26 |
-
|
27 |
-
|
28 |
-
|
29 |
-
|
30 |
-
|
31 |
-
|
32 |
-
|
33 |
-
|
34 |
-
|
35 |
-
|
36 |
-
|
37 |
-
|
38 |
-
|
39 |
-
|
40 |
-
|
41 |
-
|
42 |
-
|
43 |
-
|
44 |
-
|
45 |
-
|
46 |
-
|
47 |
-
|
48 |
-
|
49 |
-
|
50 |
-
|
51 |
-
|
52 |
-
|
53 |
-
|
54 |
-
|
55 |
-
|
56 |
-
|
57 |
-
|
58 |
-
|
59 |
-
|
60 |
-
|
61 |
-
|
62 |
-
|
63 |
-
|
64 |
-
### Recommendations
|
65 |
-
|
66 |
-
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
|
67 |
-
|
68 |
-
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
|
69 |
-
|
70 |
-
## How to Get Started with the Model
|
71 |
-
|
72 |
-
Use the code below to get started with the model.
|
73 |
-
|
74 |
-
[More Information Needed]
|
75 |
-
|
76 |
-
## Training Details
|
77 |
-
|
78 |
-
### Training Data
|
79 |
-
|
80 |
-
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
|
81 |
-
|
82 |
-
[More Information Needed]
|
83 |
-
|
84 |
-
### Training Procedure
|
85 |
-
|
86 |
-
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
|
87 |
-
|
88 |
-
#### Preprocessing [optional]
|
89 |
-
|
90 |
-
[More Information Needed]
|
91 |
-
|
92 |
|
93 |
-
#### Training Hyperparameters
|
94 |
|
95 |
-
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
|
96 |
|
97 |
-
|
98 |
-
|
99 |
-
|
100 |
-
|
101 |
-
[More Information Needed]
|
102 |
-
|
103 |
-
## Evaluation
|
104 |
-
|
105 |
-
<!-- This section describes the evaluation protocols and provides the results. -->
|
106 |
-
|
107 |
-
### Testing Data, Factors & Metrics
|
108 |
-
|
109 |
-
#### Testing Data
|
110 |
-
|
111 |
-
<!-- This should link to a Dataset Card if possible. -->
|
112 |
-
|
113 |
-
[More Information Needed]
|
114 |
-
|
115 |
-
#### Factors
|
116 |
-
|
117 |
-
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
|
118 |
-
|
119 |
-
[More Information Needed]
|
120 |
-
|
121 |
-
#### Metrics
|
122 |
-
|
123 |
-
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
|
124 |
-
|
125 |
-
[More Information Needed]
|
126 |
-
|
127 |
-
### Results
|
128 |
-
|
129 |
-
[More Information Needed]
|
130 |
-
|
131 |
-
#### Summary
|
132 |
-
|
133 |
-
|
134 |
-
|
135 |
-
## Model Examination [optional]
|
136 |
-
|
137 |
-
<!-- Relevant interpretability work for the model goes here -->
|
138 |
-
|
139 |
-
[More Information Needed]
|
140 |
-
|
141 |
-
## Environmental Impact
|
142 |
-
|
143 |
-
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
|
144 |
-
|
145 |
-
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
|
146 |
-
|
147 |
-
- **Hardware Type:** [More Information Needed]
|
148 |
-
- **Hours used:** [More Information Needed]
|
149 |
-
- **Cloud Provider:** [More Information Needed]
|
150 |
-
- **Compute Region:** [More Information Needed]
|
151 |
-
- **Carbon Emitted:** [More Information Needed]
|
152 |
-
|
153 |
-
## Technical Specifications [optional]
|
154 |
-
|
155 |
-
### Model Architecture and Objective
|
156 |
-
|
157 |
-
[More Information Needed]
|
158 |
-
|
159 |
-
### Compute Infrastructure
|
160 |
-
|
161 |
-
[More Information Needed]
|
162 |
-
|
163 |
-
#### Hardware
|
164 |
-
|
165 |
-
[More Information Needed]
|
166 |
-
|
167 |
-
#### Software
|
168 |
-
|
169 |
-
[More Information Needed]
|
170 |
-
|
171 |
-
## Citation [optional]
|
172 |
-
|
173 |
-
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
|
174 |
-
|
175 |
-
**BibTeX:**
|
176 |
-
|
177 |
-
[More Information Needed]
|
178 |
-
|
179 |
-
**APA:**
|
180 |
-
|
181 |
-
[More Information Needed]
|
182 |
|
183 |
-
## Glossary [optional]
|
184 |
|
185 |
-
|
|
|
186 |
|
187 |
-
|
|
|
|
|
|
|
188 |
|
189 |
-
##
|
|
|
|
|
190 |
|
191 |
-
|
|
|
192 |
|
193 |
-
## Model Card Authors [optional]
|
194 |
|
195 |
-
|
|
|
196 |
|
197 |
-
##
|
|
|
198 |
|
199 |
-
|
|
|
|
1 |
---
|
2 |
+
license: mit
|
3 |
+
datasets:
|
4 |
+
- facebook/empathetic_dialogues
|
5 |
+
language:
|
6 |
+
- en
|
7 |
+
base_model: alignment-handbook/zephyr-7b-sft-full
|
8 |
+
widget:
|
9 |
+
- example_title: Pirate!
|
10 |
+
messages:
|
11 |
+
- role: system
|
12 |
+
content: You are a friendly assistant, who provides empathetic responses to the user. The input contains previous turn of the dialog, where each utterance is prefaced with tags <|user>, or <|assistant|>. Be empathetic and precise. Make sure to give responses that make the dialogue flow. Avoid repeating the prompt. Please respond creatively and expressively to make the responses longer. You can offer advice.
|
13 |
+
- role: user
|
14 |
+
content: Yeah about 10 years ago I had a horrifying experience. It was 100% their fault but they hit the water barrels and survived. They had no injuries but they almost ran me off the road.
|
15 |
+
- role: assistant
|
16 |
+
content: Did you suffer any injuries?
|
17 |
+
- role: user
|
18 |
+
content: No I wasn't hit. It turned out they were drunk. I felt guilty but realized it was his fault.
|
19 |
+
output:
|
20 |
+
text: >-
|
21 |
+
That's good that you didn't get hurt. I hope they got in trouble for driving drunk.
|
22 |
+
|
23 |
+
pipeline_tag: text-generation
|
24 |
+
model-index:
|
25 |
+
- name: justtherightsize/zephyr-7b-sft-full124_d270
|
26 |
+
results:
|
27 |
+
- task:
|
28 |
+
type: text-generation
|
29 |
+
name: Text Generation
|
30 |
+
dataset:
|
31 |
+
name: Open LLM Leaderboard
|
32 |
+
type: various
|
33 |
+
config: various
|
34 |
+
split: various
|
35 |
+
args:
|
36 |
+
num_few_shot: 5
|
37 |
+
metrics:
|
38 |
+
- type: acc
|
39 |
+
name: accuracy
|
40 |
+
value: 0.2665
|
41 |
+
source:
|
42 |
+
name: Open LLM Leaderboard
|
43 |
+
url: >-
|
44 |
+
https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard
|
45 |
+
- task:
|
46 |
+
type: text-generation
|
47 |
+
name: Text Generation
|
48 |
+
dataset:
|
49 |
+
name: MMLU (5-Shot)
|
50 |
+
type: cais/mmlu
|
51 |
+
config: all
|
52 |
+
split: test
|
53 |
+
args:
|
54 |
+
num_few_shot: 5
|
55 |
+
metrics:
|
56 |
+
- type: acc
|
57 |
+
name: accuracy
|
58 |
+
value: 58.38
|
59 |
+
source:
|
60 |
+
name: MMLU
|
61 |
+
url: >-
|
62 |
+
https://github.com/huggingface/lm-evaluation-harness.git
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
63 |
|
|
|
64 |
|
|
|
65 |
|
66 |
+
---
|
67 |
+
# Model Card for zephyr-7b-sft-full124_d270
|
68 |
+
This model paricipated in multi-turn dialogues and responses empathetically.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
69 |
|
|
|
70 |
|
71 |
+
## Model Description
|
72 |
+
We propose a data-driven solution for Empathetic Response Generation with LLMs: aligning LLMs via preference optimization algorithms. First, we build a preference dataset using the benchmark dataset EmpatheticDialogues (Rashkin et al., 2019). It contains short multi-turn human-to-human dialogues grounded by emotion labels. We leverage this emotion grounding to sample dialog completions labeled with polar opposite emotions using Plutchik’s wheel (Plutchik, 2001) such that each prompt is paired with preferred and non-preferred completions. We then fine-tune a foundational LLM using Direct Preference Optimization (DPO) (Rafailov et al., 2024) to generate responses aligned with the preferred candidate response.
|
73 |
|
74 |
+
- **Developed by:** TBA
|
75 |
+
- **Model type:** Autoregressive Encoder-Decoder
|
76 |
+
- **Language(s):** en
|
77 |
+
- **Finetuned from:** alignment-handbook/zephyr-7b-sft-full
|
78 |
|
79 |
+
## Sources
|
80 |
+
- **Repository:** <https://github.com/justtherightsize/empo>
|
81 |
+
- **(*non-anonymized*) Paper preprint:** <https://arxiv.org/abs/2406.19071>
|
82 |
|
83 |
+
## Usage
|
84 |
+
TODO
|
85 |
|
|
|
86 |
|
87 |
+
## Out-of-Scope Usage
|
88 |
+
Note that fine-tuning on the EmpatheticDialogues caused some specialization.
|
89 |
|
90 |
+
## Training
|
91 |
+
TODO
|
92 |
|
93 |
+
## Cite
|
94 |
+
TBA, now please cite the **non-anonymized** [preprint](https://arxiv.org/abs/2305.15017)
|