alonzogarbanzo commited on
Commit
4105f1c
1 Parent(s): 83285c3

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +25 -5
README.md CHANGED
@@ -8,12 +8,9 @@ model-index:
8
  results: []
9
  ---
10
 
11
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
12
- should probably proofread and complete it, then remove this comment. -->
13
-
14
  # Bloom-1b7-dialogsum-IT
15
 
16
- This model is a fine-tuned version of [bigscience/bloom-1b7](https://huggingface.co/bigscience/bloom-1b7) on an unknown dataset.
17
 
18
  ## Model description
19
 
@@ -25,10 +22,31 @@ More information needed
25
 
26
  ## Training and evaluation data
27
 
28
- More information needed
29
 
30
  ## Training procedure
31
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
32
  ### Training hyperparameters
33
 
34
  The following hyperparameters were used during training:
@@ -45,7 +63,9 @@ The following hyperparameters were used during training:
45
 
46
  ### Training results
47
 
 
48
 
 
49
 
50
  ### Framework versions
51
 
 
8
  results: []
9
  ---
10
 
 
 
 
11
  # Bloom-1b7-dialogsum-IT
12
 
13
+ This model is a instruction-tuned version of [bigscience/bloom-1b7](https://huggingface.co/bigscience/bloom-1b7) on a dialog summation dataset.
14
 
15
  ## Model description
16
 
 
22
 
23
  ## Training and evaluation data
24
 
25
+ Instruction Tuned on the dialog summation task here: https://huggingface.co/datasets/adambjorn/UnrelatedForgettingOverhead/viewer/dialogsum/train
26
 
27
  ## Training procedure
28
 
29
+ Given a set of prompts:
30
+
31
+ ``` python
32
+ prompts = [
33
+ "Provide a concise summary for the following dialogue:",
34
+ "Summarize this conversation in a few sentences:",
35
+ "Here is a dialogue. Can you summarize it briefly?",
36
+ "Read the following dialogue and write a short summary:",
37
+ "Condense the essence of this conversation into a summary:"
38
+ ]
39
+ ```
40
+
41
+ Each example is concatenated with the prompt, the dialogue, and the summary as so:
42
+
43
+ ``` python
44
+ concatenated_texts = [
45
+ random.choice(prompts) + " " + dialogue + "<\s>" + " Summary:" + summary
46
+ for dialogue, summary in zip(examples['dialogue'], examples['summary'])
47
+ ]
48
+ ```
49
+
50
  ### Training hyperparameters
51
 
52
  The following hyperparameters were used during training:
 
63
 
64
  ### Training results
65
 
66
+ Final epoch results: {'loss': 0.0137, 'grad_norm': 0.6599154472351074, 'learning_rate': 7.000000000000001e-07, 'epoch': 10.0}
67
 
68
+ Average results: {'train_runtime': 1142.1524, 'train_samples_per_second': 1.751, 'train_steps_per_second': 0.438, 'train_loss': 0.37129621666669843, 'epoch': 10.0}
69
 
70
  ### Framework versions
71