File size: 3,721 Bytes
12ca3b8 d1abd57 12ca3b8 d1abd57 4284c9b d1abd57 cd7449f 397cbf3 cd7449f 397cbf3 cd7449f 12ca3b8 4284c9b 12ca3b8 d1abd57 12ca3b8 622444d 12ca3b8 c1b01be 4284c9b 12ca3b8 e225df6 12ca3b8 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 |
---
license: bigscience-bloom-rail-1.0
tags:
- text generation
- generated_from_trainer
- email generation
- email
- emailgen
datasets:
- aeslc
- postbot/multi-emails-100k
widget:
- text: 'Good Morning Professor Beans,
Hope you are doing well. I just wanted to reach out and ask if differential calculus
will be on the exam'
example_title: email to prof
- text: '嘿<NAME>
感谢你注册我的每周通讯。在我们开始之前,你必须确认你的电子邮件地址。.'
example_title: 通讯
- text: 'Hi <NAME>,
I hope this email finds you well. I wanted to reach out and ask about office hours'
example_title: office hours
- text: 'Grüße <NAME>,
Ich hoffe, du hattest einen schönen Abend beim Wurstessen der Firma. Ich melde
mich, weil'
example_title: Wurstessen festival
- text: 'Guten Morgen Harold,
ich habe mich gefragt, wann die nächste'
example_title: event
- text: URGENT - I need the TPS reports
example_title: URGENT
- text: 'Hoi Archibald,
ik hoop dat deze e-mail je goed doet.'
example_title: e-mails die je vinden
- text: 'Hello there.
I just wanted to reach out and check in to'
example_title: checking in
- text: 'Hello <NAME>,
I hope this email finds you well. I wanted to reach out and see if you''ve enjoyed
your time with us'
example_title: work well
- text: 'Hi <NAME>,
I hope this email finds you well. I wanted to reach out and see if we could catch
up'
example_title: catch up
- text: 'Jestem <NAME>,
Właśnie wprowadziłem się do obszaru i chciałem dotrzeć i uzyskać kilka szczegółów
na temat tego, gdzie mogę dostać artykuły spożywcze i'
example_title: zakupy spożywcze
parameters:
min_length: 32
max_length: 128
no_repeat_ngram_size: 2
do_sample: true
temperature: 0.2
top_k: 20
top_p: 0.95
repetition_penalty: 3.5
length_penalty: 0.9
base_model: bigscience/bloom-1b1
---
# bloom-1b1-emailgen - v1
This model is a fine-tuned version of [bigscience/bloom-1b1](https://huggingface.co/bigscience/bloom-1b1) on the ` postbot/multi-emails-100k` dataset.
It achieves the following results on the evaluation set:
- Loss: 1.7397
## Model description
More information needed
## Intended uses & limitations
⚠️ **this model did not have any of the original layers frozen during training** ⚠️
- while this is still an area of investigation, the model likely needs to have some layers frozen during fine-tuning to retain the multilingual capabilities in balance with learning how to write emails.
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 7e-05
- train_batch_size: 2
- eval_batch_size: 1
- seed: 42
- distributed_type: multi-GPU
- gradient_accumulation_steps: 64
- total_train_batch_size: 128
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: cosine
- lr_scheduler_warmup_ratio: 0.03
- num_epochs: 2.0
### Training results
| Training Loss | Epoch | Step | Validation Loss |
|:-------------:|:-----:|:----:|:---------------:|
| 1.8465 | 1.0 | 256 | 1.8656 |
| 1.4903 | 2.0 | 512 | 1.7396 |
### details
```md
***** eval metrics *****
epoch = 2.0
eval_loss = 1.7397
eval_runtime = 0:04:27.41
eval_samples = 4216
eval_samples_per_second = 15.766
eval_steps_per_second = 15.766
perplexity = 5.6956
```
### Framework versions
- Transformers 4.25.0.dev0
- Pytorch 1.13.0+cu117
- Datasets 2.6.1
- Tokenizers 0.13.1
|