Text Generation
PEFT
Safetensors
English
code
SilvioLima commited on
Commit
43f5ed2
1 Parent(s): 062a08d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +67 -12
README.md CHANGED
@@ -137,18 +137,18 @@ A partir desses valores é calculada a Precisão, o Recall e F1-score do modelo.
137
 
138
  ### Domain
139
 
140
- - Restaurant: 82.418006
141
- - Laptop: 75.098765
142
- - book: 40.076923
143
- - home: 39.204082
144
- - pet: 35.227273
145
- - beauty: 33.675676
146
- - grocery: 33.384615
147
- - electronics: 32.947368
148
- - fashion: 32.480000
149
- - toy: 30.520000
150
-
151
- ### Training Details:
152
 
153
  - Model: "google/flan-t5-base"
154
  - Tokenizer: T5Tokenizer / T5ForConditionalGeneration
@@ -159,6 +159,61 @@ A partir desses valores é calculada a Precisão, o Recall e F1-score do modelo.
159
  - Max_length: 512
160
  - Num_Epochs: 10
161
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
162
  ### Lora config:
163
 
164
  - lora_config =
 
137
 
138
  ### Domain
139
 
140
+ - Restaurant: 82.42
141
+ - Laptop: 75.10
142
+ - book: 40.08
143
+ - home: 39.21
144
+ - pet: 35.23
145
+ - beauty: 33.68
146
+ - grocery: 33.39
147
+ - electronics: 32.95
148
+ - fashion: 32.49
149
+ - toy: 30.52
150
+
151
+ ### Training Details
152
 
153
  - Model: "google/flan-t5-base"
154
  - Tokenizer: T5Tokenizer / T5ForConditionalGeneration
 
159
  - Max_length: 512
160
  - Num_Epochs: 10
161
 
162
+ ### Global Parameters
163
+
164
+ L_RATE = 3e-4
165
+
166
+ BATCH_SIZE = batch_size
167
+
168
+ PER_DEVICE_EVAL_BATCH = batch_size
169
+
170
+ WEIGHT_DECAY = 0.01
171
+
172
+ SAVE_TOTAL_LIM = 1
173
+
174
+ NUM_EPOCHS = num_epochs
175
+
176
+ MAX_NEW_TOKENS = max_length
177
+
178
+ ### Set up training arguments
179
+
180
+ training_args = Seq2SeqTrainingArguments(
181
+
182
+ output_dir="./results/absa_domain_parafrase_v512_2",
183
+
184
+ evaluation_strategy="epoch",
185
+
186
+ save_strategy="epoch",
187
+
188
+ learning_rate=L_RATE,
189
+
190
+ per_device_train_batch_size=BATCH_SIZE,
191
+
192
+ per_device_eval_batch_size=PER_DEVICE_EVAL_BATCH,
193
+
194
+ gradient_accumulation_steps=8,
195
+
196
+ weight_decay=WEIGHT_DECAY,
197
+
198
+ save_total_limit=SAVE_TOTAL_LIM,
199
+
200
+ load_best_model_at_end=True,
201
+
202
+ num_train_epochs=NUM_EPOCHS,
203
+
204
+ predict_with_generate=True,
205
+
206
+ logging_dir="./logs",
207
+
208
+ logging_strategy="epoch",
209
+
210
+ logging_steps=100,
211
+
212
+ report_to = 'wandb',
213
+
214
+ push_to_hub=True
215
+
216
+ )
217
  ### Lora config:
218
 
219
  - lora_config =