Commit
•
43f5ed2
1
Parent(s):
062a08d
Update README.md
Browse files
README.md
CHANGED
@@ -137,18 +137,18 @@ A partir desses valores é calculada a Precisão, o Recall e F1-score do modelo.
|
|
137 |
|
138 |
### Domain
|
139 |
|
140 |
-
- Restaurant: 82.
|
141 |
-
- Laptop: 75.
|
142 |
-
- book: 40.
|
143 |
-
- home: 39.
|
144 |
-
- pet: 35.
|
145 |
-
- beauty: 33.
|
146 |
-
- grocery: 33.
|
147 |
-
- electronics: 32.
|
148 |
-
- fashion: 32.
|
149 |
-
- toy: 30.
|
150 |
-
|
151 |
-
### Training Details
|
152 |
|
153 |
- Model: "google/flan-t5-base"
|
154 |
- Tokenizer: T5Tokenizer / T5ForConditionalGeneration
|
@@ -159,6 +159,61 @@ A partir desses valores é calculada a Precisão, o Recall e F1-score do modelo.
|
|
159 |
- Max_length: 512
|
160 |
- Num_Epochs: 10
|
161 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
162 |
### Lora config:
|
163 |
|
164 |
- lora_config =
|
|
|
137 |
|
138 |
### Domain
|
139 |
|
140 |
+
- Restaurant: 82.42
|
141 |
+
- Laptop: 75.10
|
142 |
+
- book: 40.08
|
143 |
+
- home: 39.21
|
144 |
+
- pet: 35.23
|
145 |
+
- beauty: 33.68
|
146 |
+
- grocery: 33.39
|
147 |
+
- electronics: 32.95
|
148 |
+
- fashion: 32.49
|
149 |
+
- toy: 30.52
|
150 |
+
|
151 |
+
### Training Details
|
152 |
|
153 |
- Model: "google/flan-t5-base"
|
154 |
- Tokenizer: T5Tokenizer / T5ForConditionalGeneration
|
|
|
159 |
- Max_length: 512
|
160 |
- Num_Epochs: 10
|
161 |
|
162 |
+
### Global Parameters
|
163 |
+
|
164 |
+
L_RATE = 3e-4
|
165 |
+
|
166 |
+
BATCH_SIZE = batch_size
|
167 |
+
|
168 |
+
PER_DEVICE_EVAL_BATCH = batch_size
|
169 |
+
|
170 |
+
WEIGHT_DECAY = 0.01
|
171 |
+
|
172 |
+
SAVE_TOTAL_LIM = 1
|
173 |
+
|
174 |
+
NUM_EPOCHS = num_epochs
|
175 |
+
|
176 |
+
MAX_NEW_TOKENS = max_length
|
177 |
+
|
178 |
+
### Set up training arguments
|
179 |
+
|
180 |
+
training_args = Seq2SeqTrainingArguments(
|
181 |
+
|
182 |
+
output_dir="./results/absa_domain_parafrase_v512_2",
|
183 |
+
|
184 |
+
evaluation_strategy="epoch",
|
185 |
+
|
186 |
+
save_strategy="epoch",
|
187 |
+
|
188 |
+
learning_rate=L_RATE,
|
189 |
+
|
190 |
+
per_device_train_batch_size=BATCH_SIZE,
|
191 |
+
|
192 |
+
per_device_eval_batch_size=PER_DEVICE_EVAL_BATCH,
|
193 |
+
|
194 |
+
gradient_accumulation_steps=8,
|
195 |
+
|
196 |
+
weight_decay=WEIGHT_DECAY,
|
197 |
+
|
198 |
+
save_total_limit=SAVE_TOTAL_LIM,
|
199 |
+
|
200 |
+
load_best_model_at_end=True,
|
201 |
+
|
202 |
+
num_train_epochs=NUM_EPOCHS,
|
203 |
+
|
204 |
+
predict_with_generate=True,
|
205 |
+
|
206 |
+
logging_dir="./logs",
|
207 |
+
|
208 |
+
logging_strategy="epoch",
|
209 |
+
|
210 |
+
logging_steps=100,
|
211 |
+
|
212 |
+
report_to = 'wandb',
|
213 |
+
|
214 |
+
push_to_hub=True
|
215 |
+
|
216 |
+
)
|
217 |
### Lora config:
|
218 |
|
219 |
- lora_config =
|