mav23 commited on
Commit
be222cc
1 Parent(s): 28c09cc

Upload folder using huggingface_hub

Browse files
Files changed (3) hide show
  1. .gitattributes +1 -0
  2. README.md +144 -0
  3. distilgpt2-emailgen-v2.Q4_0.gguf +3 -0
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ distilgpt2-emailgen-v2.Q4_0.gguf filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,144 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ tags:
4
+ - generated_from_trainer
5
+ - distilgpt2
6
+ - email generation
7
+ - email
8
+ datasets:
9
+ - aeslc
10
+ - postbot/multi-emails-100k
11
+
12
+ widget:
13
+ - text: "Good Morning Professor Beans,
14
+
15
+ Hope you are doing well. I just wanted to reach out and ask if differential calculus will be on the exam"
16
+ example_title: "email to prof"
17
+ - text: "Hey <NAME>,\n\nThank you for signing up for my weekly newsletter. Before we get started, you'll have to confirm your email address."
18
+ example_title: "newsletter"
19
+ - text: "Hi <NAME>,\n\nI hope this email finds you well. I wanted to reach out and ask about office hours"
20
+ example_title: "office hours"
21
+ - text: "Greetings <NAME>,\n\nI hope you had a splendid evening at the Company sausage eating festival. I am reaching out because"
22
+ example_title: "festival"
23
+ - text: "Good Morning Harold,\n\nI was wondering when the next"
24
+ example_title: "event"
25
+ - text: "URGENT - I need the TPS reports"
26
+ example_title: "URGENT"
27
+ - text: "Hi Archibald,\n\nI hope this email finds you extremely well."
28
+ example_title: "emails that find you"
29
+ - text: "Hello there.\n\nI just wanted to reach out and check in to"
30
+ example_title: "checking in"
31
+ - text: "Hello <NAME>,\n\nI hope this email finds you well. I wanted to reach out and see if you've enjoyed your time with us"
32
+ example_title: "work well"
33
+ - text: "Hi <NAME>,\n\nI hope this email finds you well. I wanted to reach out and see if we could catch up"
34
+ example_title: "catch up"
35
+ - text: "I'm <NAME> and I just moved into the area and wanted to reach out and get some details on where I could get groceries and"
36
+ example_title: "grocery"
37
+ parameters:
38
+ min_length: 4
39
+ max_length: 128
40
+ length_penalty: 0.8
41
+ no_repeat_ngram_size: 2
42
+ do_sample: False
43
+ num_beams: 8
44
+ early_stopping: True
45
+ repetition_penalty: 5.5
46
+ ---
47
+
48
+
49
+ # distilgpt2-emailgen: V2
50
+
51
+
52
+ [![colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/gist/pszemraj/d1c2d88b6120cca4ca7df078ea1d1e50/scratchpad.ipynb)
53
+
54
+ Why write the rest of your email when you can generate it?
55
+
56
+ ```python
57
+ from transformers import pipeline
58
+
59
+ model_tag = "postbot/distilgpt2-emailgen-V2"
60
+ generator = pipeline(
61
+ 'text-generation',
62
+ model=model_tag,
63
+ )
64
+
65
+ prompt = """
66
+ Hello,
67
+
68
+ Following up on the bubblegum shipment."""
69
+
70
+ result = generator(
71
+ prompt,
72
+ max_length=64,
73
+ do_sample=False,
74
+ early_stopping=True,
75
+ ) # generate
76
+ print(result[0]['generated_text'])
77
+ ```
78
+
79
+ ## Model description
80
+
81
+ This model is a fine-tuned version of `distilgpt2` on the postbot/multi-emails-100k dataset.
82
+ It achieves the following results on the evaluation set:
83
+ - Loss: 1.9126
84
+
85
+
86
+ ## Intended uses & limitations
87
+
88
+ More information needed
89
+
90
+ ## Training and evaluation data
91
+
92
+ More information needed
93
+
94
+ ## Training procedure
95
+
96
+ ### Training hyperparameters (run 1/2)
97
+
98
+ TODO
99
+
100
+ ### Training hyperparameters (run 2/2)
101
+
102
+ The following hyperparameters were used during training:
103
+ - learning_rate: 0.0006
104
+ - train_batch_size: 16
105
+ - eval_batch_size: 16
106
+ - seed: 42
107
+ - distributed_type: multi-GPU
108
+ - gradient_accumulation_steps: 8
109
+ - total_train_batch_size: 128
110
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
111
+ - lr_scheduler_type: cosine
112
+ - lr_scheduler_warmup_ratio: 0.01
113
+ - num_epochs: 4
114
+
115
+ ### Training results
116
+
117
+ | Training Loss | Epoch | Step | Validation Loss |
118
+ |:-------------:|:-----:|:----:|:---------------:|
119
+ | 1.9045 | 1.0 | 789 | 2.0006 |
120
+ | 1.8115 | 2.0 | 1578 | 1.9557 |
121
+ | 1.8501 | 3.0 | 2367 | 1.9110 |
122
+ | 1.7376 | 4.0 | 3156 | 1.9126 |
123
+
124
+
125
+ ### Framework versions
126
+
127
+ - Transformers 4.22.2
128
+ - Pytorch 1.10.0+cu113
129
+ - Datasets 2.5.1
130
+ - Tokenizers 0.12.1
131
+
132
+ # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
133
+ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_postbot__distilgpt2-emailgen-V2)
134
+
135
+ | Metric | Value |
136
+ |-----------------------|---------------------------|
137
+ | Avg. | 24.59 |
138
+ | ARC (25-shot) | 20.99 |
139
+ | HellaSwag (10-shot) | 26.78 |
140
+ | MMLU (5-shot) | 25.53 |
141
+ | TruthfulQA (0-shot) | 46.51 |
142
+ | Winogrande (5-shot) | 52.01 |
143
+ | GSM8K (5-shot) | 0.0 |
144
+ | DROP (3-shot) | 0.31 |
distilgpt2-emailgen-v2.Q4_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3e80662d74d3221c9a343206fdf7f0a6c3dc731506f62e1693f01e8917769a51
3
+ size 82423936