kkuramitsu commited on
Commit
01929dd
1 Parent(s): 1de64b9

mini4 AdamW

Browse files
Files changed (4) hide show
  1. config.json +1 -1
  2. pytorch_model.bin +1 -1
  3. tokenizer_config.json +1 -1
  4. train_log.txt +320 -0
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "mini2",
3
  "architectures": [
4
  "MT5ForConditionalGeneration"
5
  ],
 
1
  {
2
+ "_name_or_path": "mini4",
3
  "architectures": [
4
  "MT5ForConditionalGeneration"
5
  ],
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6a31e76db56aec04c81affe569cfb952c62ce5dea9f9c59c8593fdc08122d556
3
  size 321795553
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:85380d1e920ecc7d145c62509e086def661ad714cbf23139e6f6d3250c8384a9
3
  size 321795553
tokenizer_config.json CHANGED
@@ -3,7 +3,7 @@
3
  "eos_token": "</s>",
4
  "extra_ids": 0,
5
  "model_max_length": 1000000000000000019884624838656,
6
- "name_or_path": "mini2",
7
  "pad_token": "<pad>",
8
  "sp_model_kwargs": {},
9
  "special_tokens_map_file": "/home/acc12952oa/.cache/huggingface/hub/models--kkuramitsu--mt5np_mini12L/snapshots/e66bd8feec1522ea93ed176acb765f0c44f81526/special_tokens_map.json",
 
3
  "eos_token": "</s>",
4
  "extra_ids": 0,
5
  "model_max_length": 1000000000000000019884624838656,
6
+ "name_or_path": "mini4",
7
  "pad_token": "<pad>",
8
  "sp_model_kwargs": {},
9
  "special_tokens_map_file": "/home/acc12952oa/.cache/huggingface/hub/models--kkuramitsu--mt5np_mini12L/snapshots/e66bd8feec1522ea93ed176acb765f0c44f81526/special_tokens_map.json",
train_log.txt CHANGED
@@ -285,3 +285,323 @@ train epoch=1 loss=1.05271 PPL=2.86539
285
  [batch_size] 256
286
  [accumulate_grad_batches] 2
287
  [failed] ['/groups/gcc50582/MSP/mc4_ja_msp_00.jsonl'] [Errno 28] No space left on device
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
285
  [batch_size] 256
286
  [accumulate_grad_batches] 2
287
  [failed] ['/groups/gcc50582/MSP/mc4_ja_msp_00.jsonl'] [Errno 28] No space left on device
288
+ [hparams] Namespace(accelerator=None, adam_epsilon=1e-08, batch_size=256, cache=False, checkpoint_path=None, desc='', devices=1, early_stopping=False, fast_dev_run=False, files=['/groups/gcc50582/MSP/mc4_en_msp_09.jsonl'], float32_matmul_precision=None, gradient_accumulation_steps=1, learning_rate=0.0003, max_epochs=1, max_grad_norm=1.0, max_hours=None, max_length=128, model_path='mini3', num_workers=4, output_path='mini3', precision='bf16', pretrain=False, score=None, score_file=None, seed=42, solver='adamw', source_max_length=128, step_batch_size=128, target_max_length=128, tokenizer_path='mini3', top_k=0, warmup_steps=1, weight_decay=0.0)
289
+ [train] ['/groups/gcc50582/MSP/mc4_en_msp_09.jsonl']
290
+ [batch_size] 256
291
+ [accumulate_grad_batches] 2
292
+ val epoch=1 loss=1.95896 PPL=7.09197
293
+ val epoch=1 loss=1.90212 PPL=6.70007
294
+ train epoch=1 loss=1.02537 PPL=2.78812
295
+ [trained] 0.0[H] 41.157410267988844[M] 2469.445[sec]
296
+ [hparams] Namespace(accelerator=None, adam_epsilon=1e-08, batch_size=256, cache=False, checkpoint_path=None, desc='', devices=1, early_stopping=False, fast_dev_run=False, files=['/groups/gcc50582/MSP/mc4_ja_msp_09.jsonl'], float32_matmul_precision=None, gradient_accumulation_steps=1, learning_rate=0.0003, max_epochs=1, max_grad_norm=1.0, max_hours=None, max_length=128, model_path='mini3', num_workers=4, output_path='mini3', precision='bf16', pretrain=False, score=None, score_file=None, seed=42, solver='adamw', source_max_length=128, step_batch_size=128, target_max_length=128, tokenizer_path='mini3', top_k=0, warmup_steps=1, weight_decay=0.0)
297
+ [train] ['/groups/gcc50582/MSP/mc4_ja_msp_09.jsonl']
298
+ [batch_size] 256
299
+ [accumulate_grad_batches] 2
300
+ val epoch=1 loss=3.27824 PPL=26.52905
301
+ val epoch=1 loss=2.65349 PPL=14.20354
302
+ train epoch=1 loss=1.20492 PPL=3.33650
303
+ [trained] 0.0[H] 45.46110556125641[M] 2727.666[sec]
304
+ [hparams] Namespace(accelerator=None, adam_epsilon=1e-08, batch_size=256, cache=False, checkpoint_path=None, desc='', devices=1, early_stopping=False, fast_dev_run=False, files=['/groups/gcc50582/MSP/mc4_en_msp_08.jsonl'], float32_matmul_precision=None, gradient_accumulation_steps=1, learning_rate=0.0003, max_epochs=1, max_grad_norm=1.0, max_hours=None, max_length=128, model_path='mini3', num_workers=4, output_path='mini3', precision='bf16', pretrain=False, score=None, score_file=None, seed=42, solver='adamw', source_max_length=128, step_batch_size=128, target_max_length=128, tokenizer_path='mini3', top_k=0, warmup_steps=1, weight_decay=0.0)
305
+ [train] ['/groups/gcc50582/MSP/mc4_en_msp_08.jsonl']
306
+ [batch_size] 256
307
+ [accumulate_grad_batches] 2
308
+ val epoch=1 loss=2.57661 PPL=13.15245
309
+ val epoch=1 loss=1.92245 PPL=6.83766
310
+ train epoch=1 loss=1.02811 PPL=2.79576
311
+ [trained] 0.0[H] 41.425186324119565[M] 2485.511[sec]
312
+ [hparams] Namespace(accelerator=None, adam_epsilon=1e-08, batch_size=256, cache=False, checkpoint_path=None, desc='', devices=1, early_stopping=False, fast_dev_run=False, files=['/groups/gcc50582/MSP/mc4_ja_msp_08.jsonl'], float32_matmul_precision=None, gradient_accumulation_steps=1, learning_rate=0.0003, max_epochs=1, max_grad_norm=1.0, max_hours=None, max_length=128, model_path='mini3', num_workers=4, output_path='mini3', precision='bf16', pretrain=False, score=None, score_file=None, seed=42, solver='adamw', source_max_length=128, step_batch_size=128, target_max_length=128, tokenizer_path='mini3', top_k=0, warmup_steps=1, weight_decay=0.0)
313
+ [train] ['/groups/gcc50582/MSP/mc4_ja_msp_08.jsonl']
314
+ [batch_size] 256
315
+ [accumulate_grad_batches] 2
316
+ val epoch=1 loss=3.20961 PPL=24.76951
317
+ val epoch=1 loss=2.68262 PPL=14.62342
318
+ train epoch=1 loss=1.19228 PPL=3.29459
319
+ [trained] 0.0[H] 45.62885602712631[M] 2737.731[sec]
320
+ [hparams] Namespace(accelerator=None, adam_epsilon=1e-08, batch_size=256, cache=False, checkpoint_path=None, desc='', devices=1, early_stopping=False, fast_dev_run=False, files=['/groups/gcc50582/MSP/mc4_en_msp_07.jsonl'], float32_matmul_precision=None, gradient_accumulation_steps=1, learning_rate=0.0003, max_epochs=1, max_grad_norm=1.0, max_hours=None, max_length=128, model_path='mini3', num_workers=4, output_path='mini3', precision='bf16', pretrain=False, score=None, score_file=None, seed=42, solver='adamw', source_max_length=128, step_batch_size=128, target_max_length=128, tokenizer_path='mini3', top_k=0, warmup_steps=1, weight_decay=0.0)
321
+ [train] ['/groups/gcc50582/MSP/mc4_en_msp_07.jsonl']
322
+ [batch_size] 256
323
+ [accumulate_grad_batches] 2
324
+ val epoch=1 loss=2.46379 PPL=11.74931
325
+ val epoch=1 loss=1.88820 PPL=6.60748
326
+ train epoch=1 loss=1.02569 PPL=2.78903
327
+ [trained] 0.0[H] 41.40251029332479[M] 2484.151[sec]
328
+ [hparams] Namespace(accelerator=None, adam_epsilon=1e-08, batch_size=256, cache=False, checkpoint_path=None, desc='', devices=1, early_stopping=False, fast_dev_run=False, files=['/groups/gcc50582/MSP/mc4_ja_msp_07.jsonl'], float32_matmul_precision=None, gradient_accumulation_steps=1, learning_rate=0.0003, max_epochs=1, max_grad_norm=1.0, max_hours=None, max_length=128, model_path='mini3', num_workers=4, output_path='mini3', precision='bf16', pretrain=False, score=None, score_file=None, seed=42, solver='adamw', source_max_length=128, step_batch_size=128, target_max_length=128, tokenizer_path='mini3', top_k=0, warmup_steps=1, weight_decay=0.0)
329
+ [train] ['/groups/gcc50582/MSP/mc4_ja_msp_07.jsonl']
330
+ [batch_size] 256
331
+ [accumulate_grad_batches] 2
332
+ val epoch=1 loss=3.24453 PPL=25.64962
333
+ val epoch=1 loss=2.63306 PPL=13.91623
334
+ train epoch=1 loss=1.18417 PPL=3.26797
335
+ [trained] 0.0[H] 45.55619955062866[M] 2733.372[sec]
336
+ [hparams] Namespace(accelerator=None, adam_epsilon=1e-08, batch_size=256, cache=False, checkpoint_path=None, desc='', devices=1, early_stopping=False, fast_dev_run=False, files=['/groups/gcc50582/MSP/mc4_en_msp_06.jsonl'], float32_matmul_precision=None, gradient_accumulation_steps=1, learning_rate=0.0003, max_epochs=1, max_grad_norm=1.0, max_hours=None, max_length=128, model_path='mini3', num_workers=4, output_path='mini3', precision='bf16', pretrain=False, score=None, score_file=None, seed=42, solver='adamw', source_max_length=128, step_batch_size=128, target_max_length=128, tokenizer_path='mini3', top_k=0, warmup_steps=1, weight_decay=0.0)
337
+ [train] ['/groups/gcc50582/MSP/mc4_en_msp_06.jsonl']
338
+ [batch_size] 256
339
+ [accumulate_grad_batches] 2
340
+ val epoch=1 loss=2.47150 PPL=11.84021
341
+ val epoch=1 loss=1.87957 PPL=6.55069
342
+ train epoch=1 loss=1.02233 PPL=2.77967
343
+ [trained] 0.0[H] 41.72278196414312[M] 2503.367[sec]
344
+ [hparams] Namespace(accelerator=None, adam_epsilon=1e-08, batch_size=256, cache=False, checkpoint_path=None, desc='', devices=1, early_stopping=False, fast_dev_run=False, files=['/groups/gcc50582/MSP/mc4_ja_msp_06.jsonl'], float32_matmul_precision=None, gradient_accumulation_steps=1, learning_rate=0.0003, max_epochs=1, max_grad_norm=1.0, max_hours=None, max_length=128, model_path='mini3', num_workers=4, output_path='mini3', precision='bf16', pretrain=False, score=None, score_file=None, seed=42, solver='adamw', source_max_length=128, step_batch_size=128, target_max_length=128, tokenizer_path='mini3', top_k=0, warmup_steps=1, weight_decay=0.0)
345
+ [train] ['/groups/gcc50582/MSP/mc4_ja_msp_06.jsonl']
346
+ [batch_size] 256
347
+ [accumulate_grad_batches] 2
348
+ val epoch=1 loss=3.07940 PPL=21.74541
349
+ val epoch=1 loss=2.38436 PPL=10.85216
350
+ train epoch=1 loss=1.17423 PPL=3.23566
351
+ [trained] 0.0[H] 45.18255339066187[M] 2710.953[sec]
352
+ [hparams] Namespace(accelerator=None, adam_epsilon=1e-08, batch_size=256, cache=False, checkpoint_path=None, desc='', devices=1, early_stopping=False, fast_dev_run=False, files=['/groups/gcc50582/MSP/mc4_en_msp_05.jsonl'], float32_matmul_precision=None, gradient_accumulation_steps=1, learning_rate=0.0003, max_epochs=1, max_grad_norm=1.0, max_hours=None, max_length=128, model_path='mini3', num_workers=4, output_path='mini3', precision='bf16', pretrain=False, score=None, score_file=None, seed=42, solver='adamw', source_max_length=128, step_batch_size=128, target_max_length=128, tokenizer_path='mini3', top_k=0, warmup_steps=1, weight_decay=0.0)
353
+ [train] ['/groups/gcc50582/MSP/mc4_en_msp_05.jsonl']
354
+ [batch_size] 256
355
+ [accumulate_grad_batches] 2
356
+ val epoch=1 loss=2.48200 PPL=11.96519
357
+ val epoch=1 loss=1.88596 PPL=6.59266
358
+ train epoch=1 loss=1.01817 PPL=2.76812
359
+ [trained] 0.0[H] 41.02570736805598[M] 2461.542[sec]
360
+ [hparams] Namespace(accelerator=None, adam_epsilon=1e-08, batch_size=256, cache=False, checkpoint_path=None, desc='', devices=1, early_stopping=False, fast_dev_run=False, files=['/groups/gcc50582/MSP/mc4_ja_msp_05.jsonl'], float32_matmul_precision=None, gradient_accumulation_steps=1, learning_rate=0.0003, max_epochs=1, max_grad_norm=1.0, max_hours=None, max_length=128, model_path='mini3', num_workers=4, output_path='mini3', precision='bf16', pretrain=False, score=None, score_file=None, seed=42, solver='adamw', source_max_length=128, step_batch_size=128, target_max_length=128, tokenizer_path='mini3', top_k=0, warmup_steps=1, weight_decay=0.0)
361
+ [train] ['/groups/gcc50582/MSP/mc4_ja_msp_05.jsonl']
362
+ [batch_size] 256
363
+ [accumulate_grad_batches] 2
364
+ val epoch=1 loss=3.28042 PPL=26.58687
365
+ val epoch=1 loss=2.53102 PPL=12.56630
366
+ train epoch=1 loss=1.17467 PPL=3.23707
367
+ [trained] 0.0[H] 45.04093019167582[M] 2702.456[sec]
368
+ [hparams] Namespace(accelerator=None, adam_epsilon=1e-08, batch_size=256, cache=False, checkpoint_path=None, desc='', devices=1, early_stopping=False, fast_dev_run=False, files=['/groups/gcc50582/MSP/mc4_en_msp_04.jsonl'], float32_matmul_precision=None, gradient_accumulation_steps=1, learning_rate=0.0003, max_epochs=1, max_grad_norm=1.0, max_hours=None, max_length=128, model_path='mini3', num_workers=4, output_path='mini3', precision='bf16', pretrain=False, score=None, score_file=None, seed=42, solver='adamw', source_max_length=128, step_batch_size=128, target_max_length=128, tokenizer_path='mini3', top_k=0, warmup_steps=1, weight_decay=0.0)
369
+ [train] ['/groups/gcc50582/MSP/mc4_en_msp_04.jsonl']
370
+ [batch_size] 256
371
+ [accumulate_grad_batches] 2
372
+ val epoch=1 loss=2.34633 PPL=10.44713
373
+ val epoch=1 loss=1.78546 PPL=5.96234
374
+ train epoch=1 loss=1.02161 PPL=2.77766
375
+ [trained] 0.0[H] 40.93621168136597[M] 2456.173[sec]
376
+ [hparams] Namespace(accelerator=None, adam_epsilon=1e-08, batch_size=256, cache=False, checkpoint_path=None, desc='', devices=1, early_stopping=False, fast_dev_run=False, files=['/groups/gcc50582/MSP/mc4_ja_msp_04.jsonl'], float32_matmul_precision=None, gradient_accumulation_steps=1, learning_rate=0.0003, max_epochs=1, max_grad_norm=1.0, max_hours=None, max_length=128, model_path='mini3', num_workers=4, output_path='mini3', precision='bf16', pretrain=False, score=None, score_file=None, seed=42, solver='adamw', source_max_length=128, step_batch_size=128, target_max_length=128, tokenizer_path='mini3', top_k=0, warmup_steps=1, weight_decay=0.0)
377
+ [train] ['/groups/gcc50582/MSP/mc4_ja_msp_04.jsonl']
378
+ [batch_size] 256
379
+ [accumulate_grad_batches] 2
380
+ val epoch=1 loss=2.90283 PPL=18.22571
381
+ val epoch=1 loss=2.43509 PPL=11.41683
382
+ train epoch=1 loss=1.16996 PPL=3.22188
383
+ [trained] 0.0[H] 45.014744373162586[M] 2700.885[sec]
384
+ [hparams] Namespace(accelerator=None, adam_epsilon=1e-08, batch_size=256, cache=False, checkpoint_path=None, desc='', devices=1, early_stopping=False, fast_dev_run=False, files=['/groups/gcc50582/MSP/mc4_en_msp_03.jsonl'], float32_matmul_precision=None, gradient_accumulation_steps=1, learning_rate=0.0003, max_epochs=1, max_grad_norm=1.0, max_hours=None, max_length=128, model_path='mini3', num_workers=4, output_path='mini3', precision='bf16', pretrain=False, score=None, score_file=None, seed=42, solver='adamw', source_max_length=128, step_batch_size=128, target_max_length=128, tokenizer_path='mini3', top_k=0, warmup_steps=1, weight_decay=0.0)
385
+ [train] ['/groups/gcc50582/MSP/mc4_en_msp_03.jsonl']
386
+ [batch_size] 256
387
+ [accumulate_grad_batches] 2
388
+ val epoch=1 loss=2.38792 PPL=10.89079
389
+ val epoch=1 loss=1.87885 PPL=6.54596
390
+ train epoch=1 loss=1.01714 PPL=2.76527
391
+ [trained] 0.0[H] 40.9392077644666[M] 2456.352[sec]
392
+ [hparams] Namespace(accelerator=None, adam_epsilon=1e-08, batch_size=256, cache=False, checkpoint_path=None, desc='', devices=1, early_stopping=False, fast_dev_run=False, files=['/groups/gcc50582/MSP/mc4_ja_msp_03.jsonl'], float32_matmul_precision=None, gradient_accumulation_steps=1, learning_rate=0.0003, max_epochs=1, max_grad_norm=1.0, max_hours=None, max_length=128, model_path='mini3', num_workers=4, output_path='mini3', precision='bf16', pretrain=False, score=None, score_file=None, seed=42, solver='adamw', source_max_length=128, step_batch_size=128, target_max_length=128, tokenizer_path='mini3', top_k=0, warmup_steps=1, weight_decay=0.0)
393
+ [train] ['/groups/gcc50582/MSP/mc4_ja_msp_03.jsonl']
394
+ [batch_size] 256
395
+ [accumulate_grad_batches] 2
396
+ val epoch=1 loss=3.02356 PPL=20.56429
397
+ val epoch=1 loss=2.60609 PPL=13.54600
398
+ train epoch=1 loss=1.16236 PPL=3.19748
399
+ [trained] 0.0[H] 44.92984497149785[M] 2695.791[sec]
400
+ [hparams] Namespace(accelerator=None, adam_epsilon=1e-08, batch_size=256, cache=False, checkpoint_path=None, desc='', devices=1, early_stopping=False, fast_dev_run=False, files=['/groups/gcc50582/MSP/mc4_en_msp_02.jsonl'], float32_matmul_precision=None, gradient_accumulation_steps=1, learning_rate=0.0003, max_epochs=1, max_grad_norm=1.0, max_hours=None, max_length=128, model_path='mini3', num_workers=4, output_path='mini3', precision='bf16', pretrain=False, score=None, score_file=None, seed=42, solver='adamw', source_max_length=128, step_batch_size=128, target_max_length=128, tokenizer_path='mini3', top_k=0, warmup_steps=1, weight_decay=0.0)
401
+ [train] ['/groups/gcc50582/MSP/mc4_en_msp_02.jsonl']
402
+ [batch_size] 256
403
+ [accumulate_grad_batches] 2
404
+ val epoch=1 loss=2.38416 PPL=10.84993
405
+ val epoch=1 loss=1.85154 PPL=6.36963
406
+ train epoch=1 loss=1.01303 PPL=2.75394
407
+ [trained] 0.0[H] 40.939160716533664[M] 2456.350[sec]
408
+ [hparams] Namespace(accelerator=None, adam_epsilon=1e-08, batch_size=256, cache=False, checkpoint_path=None, desc='', devices=1, early_stopping=False, fast_dev_run=False, files=['/groups/gcc50582/MSP/mc4_ja_msp_02.jsonl'], float32_matmul_precision=None, gradient_accumulation_steps=1, learning_rate=0.0003, max_epochs=1, max_grad_norm=1.0, max_hours=None, max_length=128, model_path='mini3', num_workers=4, output_path='mini3', precision='bf16', pretrain=False, score=None, score_file=None, seed=42, solver='adamw', source_max_length=128, step_batch_size=128, target_max_length=128, tokenizer_path='mini3', top_k=0, warmup_steps=1, weight_decay=0.0)
409
+ [train] ['/groups/gcc50582/MSP/mc4_ja_msp_02.jsonl']
410
+ [batch_size] 256
411
+ [accumulate_grad_batches] 2
412
+ val epoch=1 loss=2.99403 PPL=19.96596
413
+ val epoch=1 loss=2.40907 PPL=11.12363
414
+ train epoch=1 loss=1.15822 PPL=3.18425
415
+ [trained] 0.0[H] 45.009803128242496[M] 2700.588[sec]
416
+ [hparams] Namespace(accelerator=None, adam_epsilon=1e-08, batch_size=256, cache=False, checkpoint_path=None, desc='', devices=1, early_stopping=False, fast_dev_run=False, files=['/groups/gcc50582/MSP/mc4_en_msp_01.jsonl'], float32_matmul_precision=None, gradient_accumulation_steps=1, learning_rate=0.0003, max_epochs=1, max_grad_norm=1.0, max_hours=None, max_length=128, model_path='mini3', num_workers=4, output_path='mini3', precision='bf16', pretrain=False, score=None, score_file=None, seed=42, solver='adamw', source_max_length=128, step_batch_size=128, target_max_length=128, tokenizer_path='mini3', top_k=0, warmup_steps=1, weight_decay=0.0)
417
+ [train] ['/groups/gcc50582/MSP/mc4_en_msp_01.jsonl']
418
+ [batch_size] 256
419
+ [accumulate_grad_batches] 2
420
+ val epoch=1 loss=2.31114 PPL=10.08587
421
+ val epoch=1 loss=1.80426 PPL=6.07548
422
+ train epoch=1 loss=1.00900 PPL=2.74285
423
+ [trained] 0.0[H] 40.89026815891266[M] 2453.416[sec]
424
+ [hparams] Namespace(accelerator=None, adam_epsilon=1e-08, batch_size=256, cache=False, checkpoint_path=None, desc='', devices=1, early_stopping=False, fast_dev_run=False, files=['/groups/gcc50582/MSP/mc4_ja_msp_01.jsonl'], float32_matmul_precision=None, gradient_accumulation_steps=1, learning_rate=0.0003, max_epochs=1, max_grad_norm=1.0, max_hours=None, max_length=128, model_path='mini3', num_workers=4, output_path='mini3', precision='bf16', pretrain=False, score=None, score_file=None, seed=42, solver='adamw', source_max_length=128, step_batch_size=128, target_max_length=128, tokenizer_path='mini3', top_k=0, warmup_steps=1, weight_decay=0.0)
425
+ [train] ['/groups/gcc50582/MSP/mc4_ja_msp_01.jsonl']
426
+ [batch_size] 256
427
+ [accumulate_grad_batches] 2
428
+ val epoch=1 loss=2.62659 PPL=13.82649
429
+ val epoch=1 loss=2.22096 PPL=9.21613
430
+ train epoch=1 loss=1.15455 PPL=3.17260
431
+ [trained] 0.0[H] 45.08568317492803[M] 2705.141[sec]
432
+ [hparams] Namespace(accelerator=None, adam_epsilon=1e-08, batch_size=256, cache=False, checkpoint_path=None, desc='', devices=1, early_stopping=False, fast_dev_run=False, files=['/groups/gcc50582/MSP/mc4_en_msp_00.jsonl'], float32_matmul_precision=None, gradient_accumulation_steps=1, learning_rate=0.0003, max_epochs=1, max_grad_norm=1.0, max_hours=None, max_length=128, model_path='mini3', num_workers=4, output_path='mini3', precision='bf16', pretrain=False, score=None, score_file=None, seed=42, solver='adamw', source_max_length=128, step_batch_size=128, target_max_length=128, tokenizer_path='mini3', top_k=0, warmup_steps=1, weight_decay=0.0)
433
+ [train] ['/groups/gcc50582/MSP/mc4_en_msp_00.jsonl']
434
+ [batch_size] 256
435
+ [accumulate_grad_batches] 2
436
+ val epoch=1 loss=2.33120 PPL=10.29027
437
+ val epoch=1 loss=1.82923 PPL=6.22906
438
+ train epoch=1 loss=1.00533 PPL=2.73282
439
+ [trained] 0.0[H] 41.071046415964766[M] 2464.263[sec]
440
+ [hparams] Namespace(accelerator=None, adam_epsilon=1e-08, batch_size=256, cache=False, checkpoint_path=None, desc='', devices=1, early_stopping=False, fast_dev_run=False, files=['/groups/gcc50582/MSP/mc4_ja_msp_00.jsonl'], float32_matmul_precision=None, gradient_accumulation_steps=1, learning_rate=0.0003, max_epochs=1, max_grad_norm=1.0, max_hours=None, max_length=128, model_path='mini3', num_workers=4, output_path='mini3', precision='bf16', pretrain=False, score=None, score_file=None, seed=42, solver='adamw', source_max_length=128, step_batch_size=128, target_max_length=128, tokenizer_path='mini3', top_k=0, warmup_steps=1, weight_decay=0.0)
441
+ [train] ['/groups/gcc50582/MSP/mc4_ja_msp_00.jsonl']
442
+ [batch_size] 256
443
+ [accumulate_grad_batches] 2
444
+ val epoch=1 loss=2.96448 PPL=19.38460
445
+ val epoch=1 loss=2.56859 PPL=13.04741
446
+ train epoch=1 loss=1.14901 PPL=3.15508
447
+ [trained] 0.0[H] 45.149318718910216[M] 2708.959[sec]
448
+ [hparams] Namespace(accelerator=None, adam_epsilon=1e-08, batch_size=256, cache=False, checkpoint_path=None, desc='', devices=1, early_stopping=False, fast_dev_run=False, files=['/groups/gcc50582/MSP/mc4_en_msp_09.jsonl'], float32_matmul_precision=None, gradient_accumulation_steps=1, learning_rate=0.0002, max_epochs=1, max_grad_norm=1.0, max_hours=None, max_length=128, model_path='mini4', num_workers=4, output_path='mini4', precision='bf16', pretrain=False, score=None, score_file=None, seed=42, solver='adamw', source_max_length=128, step_batch_size=128, target_max_length=128, tokenizer_path='mini4', top_k=0, warmup_steps=1, weight_decay=0.0)
449
+ [train] ['/groups/gcc50582/MSP/mc4_en_msp_09.jsonl']
450
+ [batch_size] 256
451
+ [accumulate_grad_batches] 2
452
+ val epoch=1 loss=2.31578 PPL=10.13282
453
+ val epoch=1 loss=1.84484 PPL=6.32710
454
+ train epoch=1 loss=0.99815 PPL=2.71326
455
+ [trained] 0.0[H] 41.47976658344269[M] 2488.786[sec]
456
+ [hparams] Namespace(accelerator=None, adam_epsilon=1e-08, batch_size=256, cache=False, checkpoint_path=None, desc='', devices=1, early_stopping=False, fast_dev_run=False, files=['/groups/gcc50582/MSP/mc4_ja_msp_09.jsonl'], float32_matmul_precision=None, gradient_accumulation_steps=1, learning_rate=0.0002, max_epochs=1, max_grad_norm=1.0, max_hours=None, max_length=128, model_path='mini4', num_workers=4, output_path='mini4', precision='bf16', pretrain=False, score=None, score_file=None, seed=42, solver='adamw', source_max_length=128, step_batch_size=128, target_max_length=128, tokenizer_path='mini4', top_k=0, warmup_steps=1, weight_decay=0.0)
457
+ [train] ['/groups/gcc50582/MSP/mc4_ja_msp_09.jsonl']
458
+ [batch_size] 256
459
+ [accumulate_grad_batches] 2
460
+ val epoch=1 loss=2.79162 PPL=16.30747
461
+ val epoch=1 loss=2.55381 PPL=12.85597
462
+ train epoch=1 loss=1.13298 PPL=3.10490
463
+ [trained] 0.0[H] 44.85936383008957[M] 2691.562[sec]
464
+ [hparams] Namespace(accelerator=None, adam_epsilon=1e-08, batch_size=256, cache=False, checkpoint_path=None, desc='', devices=1, early_stopping=False, fast_dev_run=False, files=['/groups/gcc50582/MSP/mc4_en_msp_08.jsonl'], float32_matmul_precision=None, gradient_accumulation_steps=1, learning_rate=0.0002, max_epochs=1, max_grad_norm=1.0, max_hours=None, max_length=128, model_path='mini4', num_workers=4, output_path='mini4', precision='bf16', pretrain=False, score=None, score_file=None, seed=42, solver='adamw', source_max_length=128, step_batch_size=128, target_max_length=128, tokenizer_path='mini4', top_k=0, warmup_steps=1, weight_decay=0.0)
465
+ [train] ['/groups/gcc50582/MSP/mc4_en_msp_08.jsonl']
466
+ [batch_size] 256
467
+ [accumulate_grad_batches] 2
468
+ val epoch=1 loss=2.19775 PPL=9.00472
469
+ val epoch=1 loss=1.86080 PPL=6.42891
470
+ train epoch=1 loss=0.99177 PPL=2.69602
471
+ [trained] 0.0[H] 41.247025799751285[M] 2474.822[sec]
472
+ [hparams] Namespace(accelerator=None, adam_epsilon=1e-08, batch_size=256, cache=False, checkpoint_path=None, desc='', devices=1, early_stopping=False, fast_dev_run=False, files=['/groups/gcc50582/MSP/mc4_ja_msp_08.jsonl'], float32_matmul_precision=None, gradient_accumulation_steps=1, learning_rate=0.0002, max_epochs=1, max_grad_norm=1.0, max_hours=None, max_length=128, model_path='mini4', num_workers=4, output_path='mini4', precision='bf16', pretrain=False, score=None, score_file=None, seed=42, solver='adamw', source_max_length=128, step_batch_size=128, target_max_length=128, tokenizer_path='mini4', top_k=0, warmup_steps=1, weight_decay=0.0)
473
+ [train] ['/groups/gcc50582/MSP/mc4_ja_msp_08.jsonl']
474
+ [batch_size] 256
475
+ [accumulate_grad_batches] 2
476
+ val epoch=1 loss=2.88427 PPL=17.89051
477
+ val epoch=1 loss=2.61712 PPL=13.69621
478
+ train epoch=1 loss=1.13177 PPL=3.10115
479
+ [trained] 0.0[H] 44.970662931601204[M] 2698.240[sec]
480
+ [hparams] Namespace(accelerator=None, adam_epsilon=1e-08, batch_size=256, cache=False, checkpoint_path=None, desc='', devices=1, early_stopping=False, fast_dev_run=False, files=['/groups/gcc50582/MSP/mc4_en_msp_07.jsonl'], float32_matmul_precision=None, gradient_accumulation_steps=1, learning_rate=0.0002, max_epochs=1, max_grad_norm=1.0, max_hours=None, max_length=128, model_path='mini4', num_workers=4, output_path='mini4', precision='bf16', pretrain=False, score=None, score_file=None, seed=42, solver='adamw', source_max_length=128, step_batch_size=128, target_max_length=128, tokenizer_path='mini4', top_k=0, warmup_steps=1, weight_decay=0.0)
481
+ [train] ['/groups/gcc50582/MSP/mc4_en_msp_07.jsonl']
482
+ [batch_size] 256
483
+ [accumulate_grad_batches] 2
484
+ val epoch=1 loss=2.13605 PPL=8.46593
485
+ val epoch=1 loss=1.82511 PPL=6.20346
486
+ train epoch=1 loss=0.98874 PPL=2.68785
487
+ [trained] 0.0[H] 41.13867333332698[M] 2468.320[sec]
488
+ [hparams] Namespace(accelerator=None, adam_epsilon=1e-08, batch_size=256, cache=False, checkpoint_path=None, desc='', devices=1, early_stopping=False, fast_dev_run=False, files=['/groups/gcc50582/MSP/mc4_ja_msp_07.jsonl'], float32_matmul_precision=None, gradient_accumulation_steps=1, learning_rate=0.0002, max_epochs=1, max_grad_norm=1.0, max_hours=None, max_length=128, model_path='mini4', num_workers=4, output_path='mini4', precision='bf16', pretrain=False, score=None, score_file=None, seed=42, solver='adamw', source_max_length=128, step_batch_size=128, target_max_length=128, tokenizer_path='mini4', top_k=0, warmup_steps=1, weight_decay=0.0)
489
+ [train] ['/groups/gcc50582/MSP/mc4_ja_msp_07.jsonl']
490
+ [batch_size] 256
491
+ [accumulate_grad_batches] 2
492
+ val epoch=1 loss=2.80929 PPL=16.59817
493
+ val epoch=1 loss=2.57963 PPL=13.19225
494
+ train epoch=1 loss=1.12952 PPL=3.09418
495
+ [trained] 0.0[H] 44.85550689299901[M] 2691.330[sec]
496
+ [hparams] Namespace(accelerator=None, adam_epsilon=1e-08, batch_size=256, cache=False, checkpoint_path=None, desc='', devices=1, early_stopping=False, fast_dev_run=False, files=['/groups/gcc50582/MSP/mc4_en_msp_06.jsonl'], float32_matmul_precision=None, gradient_accumulation_steps=1, learning_rate=0.0002, max_epochs=1, max_grad_norm=1.0, max_hours=None, max_length=128, model_path='mini4', num_workers=4, output_path='mini4', precision='bf16', pretrain=False, score=None, score_file=None, seed=42, solver='adamw', source_max_length=128, step_batch_size=128, target_max_length=128, tokenizer_path='mini4', top_k=0, warmup_steps=1, weight_decay=0.0)
497
+ [train] ['/groups/gcc50582/MSP/mc4_en_msp_06.jsonl']
498
+ [batch_size] 256
499
+ [accumulate_grad_batches] 2
500
+ val epoch=1 loss=2.10408 PPL=8.19958
501
+ val epoch=1 loss=1.82278 PPL=6.18904
502
+ train epoch=1 loss=0.98625 PPL=2.68116
503
+ [trained] 0.0[H] 41.238258417447405[M] 2474.296[sec]
504
+ [hparams] Namespace(accelerator=None, adam_epsilon=1e-08, batch_size=256, cache=False, checkpoint_path=None, desc='', devices=1, early_stopping=False, fast_dev_run=False, files=['/groups/gcc50582/MSP/mc4_ja_msp_06.jsonl'], float32_matmul_precision=None, gradient_accumulation_steps=1, learning_rate=0.0002, max_epochs=1, max_grad_norm=1.0, max_hours=None, max_length=128, model_path='mini4', num_workers=4, output_path='mini4', precision='bf16', pretrain=False, score=None, score_file=None, seed=42, solver='adamw', source_max_length=128, step_batch_size=128, target_max_length=128, tokenizer_path='mini4', top_k=0, warmup_steps=1, weight_decay=0.0)
505
+ [train] ['/groups/gcc50582/MSP/mc4_ja_msp_06.jsonl']
506
+ [batch_size] 256
507
+ [accumulate_grad_batches] 2
508
+ val epoch=1 loss=2.54932 PPL=12.79840
509
+ val epoch=1 loss=2.30628 PPL=10.03706
510
+ train epoch=1 loss=1.12400 PPL=3.07713
511
+ [trained] 0.0[H] 45.42449986537297[M] 2725.470[sec]
512
+ [hparams] Namespace(accelerator=None, adam_epsilon=1e-08, batch_size=256, cache=False, checkpoint_path=None, desc='', devices=1, early_stopping=False, fast_dev_run=False, files=['/groups/gcc50582/MSP/mc4_en_msp_05.jsonl'], float32_matmul_precision=None, gradient_accumulation_steps=1, learning_rate=0.0002, max_epochs=1, max_grad_norm=1.0, max_hours=None, max_length=128, model_path='mini4', num_workers=4, output_path='mini4', precision='bf16', pretrain=False, score=None, score_file=None, seed=42, solver='adamw', source_max_length=128, step_batch_size=128, target_max_length=128, tokenizer_path='mini4', top_k=0, warmup_steps=1, weight_decay=0.0)
513
+ [train] ['/groups/gcc50582/MSP/mc4_en_msp_05.jsonl']
514
+ [batch_size] 256
515
+ [accumulate_grad_batches] 2
516
+ val epoch=1 loss=2.11911 PPL=8.32372
517
+ val epoch=1 loss=1.83168 PPL=6.24436
518
+ train epoch=1 loss=0.98346 PPL=2.67370
519
+ [trained] 0.0[H] 41.16821654637655[M] 2470.093[sec]
520
+ [hparams] Namespace(accelerator=None, adam_epsilon=1e-08, batch_size=256, cache=False, checkpoint_path=None, desc='', devices=1, early_stopping=False, fast_dev_run=False, files=['/groups/gcc50582/MSP/mc4_ja_msp_05.jsonl'], float32_matmul_precision=None, gradient_accumulation_steps=1, learning_rate=0.0002, max_epochs=1, max_grad_norm=1.0, max_hours=None, max_length=128, model_path='mini4', num_workers=4, output_path='mini4', precision='bf16', pretrain=False, score=None, score_file=None, seed=42, solver='adamw', source_max_length=128, step_batch_size=128, target_max_length=128, tokenizer_path='mini4', top_k=0, warmup_steps=1, weight_decay=0.0)
521
+ [train] ['/groups/gcc50582/MSP/mc4_ja_msp_05.jsonl']
522
+ [batch_size] 256
523
+ [accumulate_grad_batches] 2
524
+ val epoch=1 loss=2.69997 PPL=14.87932
525
+ val epoch=1 loss=2.48527 PPL=12.00436
526
+ train epoch=1 loss=1.12319 PPL=3.07466
527
+ [trained] 0.0[H] 44.44852333863576[M] 2666.911[sec]
528
+ [hparams] Namespace(accelerator=None, adam_epsilon=1e-08, batch_size=256, cache=False, checkpoint_path=None, desc='', devices=1, early_stopping=False, fast_dev_run=False, files=['/groups/gcc50582/MSP/mc4_en_msp_04.jsonl'], float32_matmul_precision=None, gradient_accumulation_steps=1, learning_rate=0.0002, max_epochs=1, max_grad_norm=1.0, max_hours=None, max_length=128, model_path='mini4', num_workers=4, output_path='mini4', precision='bf16', pretrain=False, score=None, score_file=None, seed=42, solver='adamw', source_max_length=128, step_batch_size=128, target_max_length=128, tokenizer_path='mini4', top_k=0, warmup_steps=1, weight_decay=0.0)
529
+ [train] ['/groups/gcc50582/MSP/mc4_en_msp_04.jsonl']
530
+ [batch_size] 256
531
+ [accumulate_grad_batches] 2
532
+ val epoch=1 loss=1.98369 PPL=7.26949
533
+ val epoch=1 loss=1.72509 PPL=5.61302
534
+ train epoch=1 loss=0.98371 PPL=2.67435
535
+ [trained] 0.0[H] 41.144457550843555[M] 2468.667[sec]
536
+ [hparams] Namespace(accelerator=None, adam_epsilon=1e-08, batch_size=256, cache=False, checkpoint_path=None, desc='', devices=1, early_stopping=False, fast_dev_run=False, files=['/groups/gcc50582/MSP/mc4_ja_msp_04.jsonl'], float32_matmul_precision=None, gradient_accumulation_steps=1, learning_rate=0.0002, max_epochs=1, max_grad_norm=1.0, max_hours=None, max_length=128, model_path='mini4', num_workers=4, output_path='mini4', precision='bf16', pretrain=False, score=None, score_file=None, seed=42, solver='adamw', source_max_length=128, step_batch_size=128, target_max_length=128, tokenizer_path='mini4', top_k=0, warmup_steps=1, weight_decay=0.0)
537
+ [train] ['/groups/gcc50582/MSP/mc4_ja_msp_04.jsonl']
538
+ [batch_size] 256
539
+ [accumulate_grad_batches] 2
540
+ val epoch=1 loss=2.53665 PPL=12.63722
541
+ val epoch=1 loss=2.40133 PPL=11.03790
542
+ train epoch=1 loss=1.12251 PPL=3.07255
543
+ [trained] 0.0[H] 44.89564416805903[M] 2693.739[sec]
544
+ [hparams] Namespace(accelerator=None, adam_epsilon=1e-08, batch_size=256, cache=False, checkpoint_path=None, desc='', devices=1, early_stopping=False, fast_dev_run=False, files=['/groups/gcc50582/MSP/mc4_en_msp_13.jsonl'], float32_matmul_precision=None, gradient_accumulation_steps=1, learning_rate=0.0002, max_epochs=1, max_grad_norm=1.0, max_hours=None, max_length=128, model_path='mini4', num_workers=4, output_path='mini4', precision='bf16', pretrain=False, score=None, score_file=None, seed=42, solver='adamw', source_max_length=128, step_batch_size=128, target_max_length=128, tokenizer_path='mini4', top_k=0, warmup_steps=1, weight_decay=0.0)
545
+ [train] ['/groups/gcc50582/MSP/mc4_en_msp_13.jsonl']
546
+ [batch_size] 256
547
+ [accumulate_grad_batches] 2
548
+ val epoch=1 loss=2.04189 PPL=7.70517
549
+ val epoch=1 loss=1.78558 PPL=5.96303
550
+ train epoch=1 loss=0.98738 PPL=2.68419
551
+ [trained] 0.0[H] 41.18701983690262[M] 2471.221[sec]
552
+ [hparams] Namespace(accelerator=None, adam_epsilon=1e-08, batch_size=256, cache=False, checkpoint_path=None, desc='', devices=1, early_stopping=False, fast_dev_run=False, files=['/groups/gcc50582/MSP/mc4_ja_msp_13.jsonl'], float32_matmul_precision=None, gradient_accumulation_steps=1, learning_rate=0.0002, max_epochs=1, max_grad_norm=1.0, max_hours=None, max_length=128, model_path='mini4', num_workers=4, output_path='mini4', precision='bf16', pretrain=False, score=None, score_file=None, seed=42, solver='adamw', source_max_length=128, step_batch_size=128, target_max_length=128, tokenizer_path='mini4', top_k=0, warmup_steps=1, weight_decay=0.0)
553
+ [train] ['/groups/gcc50582/MSP/mc4_ja_msp_13.jsonl']
554
+ [batch_size] 256
555
+ [accumulate_grad_batches] 2
556
+ val epoch=1 loss=2.55545 PPL=12.87711
557
+ val epoch=1 loss=2.23607 PPL=9.35652
558
+ train epoch=1 loss=1.12599 PPL=3.08326
559
+ [trained] 0.0[H] 44.80764791965485[M] 2688.459[sec]
560
+ [hparams] Namespace(accelerator=None, adam_epsilon=1e-08, batch_size=256, cache=False, checkpoint_path=None, desc='', devices=1, early_stopping=False, fast_dev_run=False, files=['/groups/gcc50582/MSP/mc4_en_msp_12.jsonl'], float32_matmul_precision=None, gradient_accumulation_steps=1, learning_rate=0.0002, max_epochs=1, max_grad_norm=1.0, max_hours=None, max_length=128, model_path='mini4', num_workers=4, output_path='mini4', precision='bf16', pretrain=False, score=None, score_file=None, seed=42, solver='adamw', source_max_length=128, step_batch_size=128, target_max_length=128, tokenizer_path='mini4', top_k=0, warmup_steps=1, weight_decay=0.0)
561
+ [train] ['/groups/gcc50582/MSP/mc4_en_msp_12.jsonl']
562
+ [batch_size] 256
563
+ [accumulate_grad_batches] 2
564
+ val epoch=1 loss=1.90556 PPL=6.72317
565
+ val epoch=1 loss=1.68472 PPL=5.39096
566
+ train epoch=1 loss=0.98539 PPL=2.67885
567
+ [trained] 0.0[H] 41.29294394652049[M] 2477.577[sec]
568
+ [hparams] Namespace(accelerator=None, adam_epsilon=1e-08, batch_size=256, cache=False, checkpoint_path=None, desc='', devices=1, early_stopping=False, fast_dev_run=False, files=['/groups/gcc50582/MSP/mc4_ja_msp_12.jsonl'], float32_matmul_precision=None, gradient_accumulation_steps=1, learning_rate=0.0002, max_epochs=1, max_grad_norm=1.0, max_hours=None, max_length=128, model_path='mini4', num_workers=4, output_path='mini4', precision='bf16', pretrain=False, score=None, score_file=None, seed=42, solver='adamw', source_max_length=128, step_batch_size=128, target_max_length=128, tokenizer_path='mini4', top_k=0, warmup_steps=1, weight_decay=0.0)
569
+ [train] ['/groups/gcc50582/MSP/mc4_ja_msp_12.jsonl']
570
+ [batch_size] 256
571
+ [accumulate_grad_batches] 2
572
+ val epoch=1 loss=2.60585 PPL=13.54276
573
+ val epoch=1 loss=2.41645 PPL=11.20597
574
+ train epoch=1 loss=1.12524 PPL=3.08097
575
+ [trained] 0.0[H] 47.45244402488073[M] 2847.147[sec]
576
+ [hparams] Namespace(accelerator=None, adam_epsilon=1e-08, batch_size=256, cache=False, checkpoint_path=None, desc='', devices=1, early_stopping=False, fast_dev_run=False, files=['/groups/gcc50582/MSP/mc4_en_msp_11.jsonl'], float32_matmul_precision=None, gradient_accumulation_steps=1, learning_rate=0.0002, max_epochs=1, max_grad_norm=1.0, max_hours=None, max_length=128, model_path='mini4', num_workers=4, output_path='mini4', precision='bf16', pretrain=False, score=None, score_file=None, seed=42, solver='adamw', source_max_length=128, step_batch_size=128, target_max_length=128, tokenizer_path='mini4', top_k=0, warmup_steps=1, weight_decay=0.0)
577
+ [train] ['/groups/gcc50582/MSP/mc4_en_msp_11.jsonl']
578
+ [batch_size] 256
579
+ [accumulate_grad_batches] 2
580
+ val epoch=1 loss=1.99018 PPL=7.31686
581
+ val epoch=1 loss=1.75003 PPL=5.75478
582
+ train epoch=1 loss=0.98361 PPL=2.67409
583
+ [trained] 0.0[H] 41.130028867721556[M] 2467.802[sec]
584
+ [hparams] Namespace(accelerator=None, adam_epsilon=1e-08, batch_size=256, cache=False, checkpoint_path=None, desc='', devices=1, early_stopping=False, fast_dev_run=False, files=['/groups/gcc50582/MSP/mc4_ja_msp_11.jsonl'], float32_matmul_precision=None, gradient_accumulation_steps=1, learning_rate=0.0002, max_epochs=1, max_grad_norm=1.0, max_hours=None, max_length=128, model_path='mini4', num_workers=4, output_path='mini4', precision='bf16', pretrain=False, score=None, score_file=None, seed=42, solver='adamw', source_max_length=128, step_batch_size=128, target_max_length=128, tokenizer_path='mini4', top_k=0, warmup_steps=1, weight_decay=0.0)
585
+ [train] ['/groups/gcc50582/MSP/mc4_ja_msp_11.jsonl']
586
+ [batch_size] 256
587
+ [accumulate_grad_batches] 2
588
+ val epoch=1 loss=2.69124 PPL=14.74992
589
+ val epoch=1 loss=2.52559 PPL=12.49825
590
+ train epoch=1 loss=1.12024 PPL=3.06559
591
+ [trained] 0.0[H] 44.83212472200394[M] 2689.927[sec]
592
+ [hparams] Namespace(accelerator=None, adam_epsilon=1e-08, batch_size=256, cache=False, checkpoint_path=None, desc='', devices=1, early_stopping=False, fast_dev_run=False, files=['/groups/gcc50582/MSP/mc4_en_msp_10.jsonl'], float32_matmul_precision=None, gradient_accumulation_steps=1, learning_rate=0.0002, max_epochs=1, max_grad_norm=1.0, max_hours=None, max_length=128, model_path='mini4', num_workers=4, output_path='mini4', precision='bf16', pretrain=False, score=None, score_file=None, seed=42, solver='adamw', source_max_length=128, step_batch_size=128, target_max_length=128, tokenizer_path='mini4', top_k=0, warmup_steps=1, weight_decay=0.0)
593
+ [train] ['/groups/gcc50582/MSP/mc4_en_msp_10.jsonl']
594
+ [batch_size] 256
595
+ [accumulate_grad_batches] 2
596
+ val epoch=1 loss=2.01462 PPL=7.49789
597
+ val epoch=1 loss=1.79355 PPL=6.01077
598
+ train epoch=1 loss=0.98182 PPL=2.66931
599
+ [trained] 0.0[H] 41.305239562193556[M] 2478.314[sec]
600
+ [hparams] Namespace(accelerator=None, adam_epsilon=1e-08, batch_size=256, cache=False, checkpoint_path=None, desc='', devices=1, early_stopping=False, fast_dev_run=False, files=['/groups/gcc50582/MSP/mc4_ja_msp_10.jsonl'], float32_matmul_precision=None, gradient_accumulation_steps=1, learning_rate=0.0002, max_epochs=1, max_grad_norm=1.0, max_hours=None, max_length=128, model_path='mini4', num_workers=4, output_path='mini4', precision='bf16', pretrain=False, score=None, score_file=None, seed=42, solver='adamw', source_max_length=128, step_batch_size=128, target_max_length=128, tokenizer_path='mini4', top_k=0, warmup_steps=1, weight_decay=0.0)
601
+ [train] ['/groups/gcc50582/MSP/mc4_ja_msp_10.jsonl']
602
+ [batch_size] 256
603
+ [accumulate_grad_batches] 2
604
+ val epoch=1 loss=2.34977 PPL=10.48312
605
+ val epoch=1 loss=2.21273 PPL=9.14063
606
+ train epoch=1 loss=1.11990 PPL=3.06456
607
+ [trained] 0.0[H] 44.55307694673538[M] 2673.185[sec]