Update distillation.py
Browse files- distillation.py +2 -2
distillation.py
CHANGED
@@ -36,8 +36,8 @@ teacher_model_name = '/Users/fujidai/sinTED/paraphrase-mpnet-base-v2' #Our mon
|
|
36 |
student_model_name = '/Users/fujidai/dataseigen/09-MarginMSELoss-finetuning-7-5' #Multilingual base model we use to imitate the teacher model
|
37 |
|
38 |
max_seq_length = 128 #Student model max. lengths for inputs (number of word pieces)
|
39 |
-
train_batch_size =
|
40 |
-
inference_batch_size =
|
41 |
max_sentences_per_language = 500000 #Maximum number of parallel sentences for training
|
42 |
train_max_sentence_length = 250 #Maximum length (characters) for parallel training sentences
|
43 |
|
|
|
36 |
student_model_name = '/Users/fujidai/dataseigen/09-MarginMSELoss-finetuning-7-5' #Multilingual base model we use to imitate the teacher model
|
37 |
|
38 |
max_seq_length = 128 #Student model max. lengths for inputs (number of word pieces)
|
39 |
+
train_batch_size = 64 #Batch size for training
|
40 |
+
inference_batch_size = 64 #Batch size at inference
|
41 |
max_sentences_per_language = 500000 #Maximum number of parallel sentences for training
|
42 |
train_max_sentence_length = 250 #Maximum length (characters) for parallel training sentences
|
43 |
|