elsayedissa
commited on
Commit
•
f8064d0
1
Parent(s):
e5e85ab
Training in progress, step 11000
Browse files- last-checkpoint/optimizer.pt +1 -1
- last-checkpoint/pytorch_model.bin +1 -1
- last-checkpoint/rng_state.pth +2 -2
- last-checkpoint/scaler.pt +1 -1
- last-checkpoint/scheduler.pt +1 -1
- last-checkpoint/trainer_state.json +252 -3
- pytorch_model.bin +1 -1
- runs/Dec26_19-57-19_gpu07.cyverse.org/events.out.tfevents.1672109970.gpu07.cyverse.org.126369.0 +2 -2
last-checkpoint/optimizer.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 12347192855
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b15938f7058276e6bdfc2fddcc7a120da5e6a0c7c9c19c71e16cb873fc6886b3
|
3 |
size 12347192855
|
last-checkpoint/pytorch_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 6173655480
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6bdd501f6153777369e4dc961faa1bf60f1f7535ded5d73c87f2ff06b5aa8064
|
3 |
size 6173655480
|
last-checkpoint/rng_state.pth
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:069bc79c0b97b8bebabb6777e867a3617901ff0e79b431f2ba4a791a90815e7f
|
3 |
+
size 14639
|
last-checkpoint/scaler.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 557
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1b4b38bb567a4c3bd8aafad6c492b89fde494be62ffddc238da23107b19a9945
|
3 |
size 557
|
last-checkpoint/scheduler.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 627
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7e9a74567933ed8c342d4d89869156924da4e6db37efcc7105e4c34e754e42a4
|
3 |
size 627
|
last-checkpoint/trainer_state.json
CHANGED
@@ -1,8 +1,8 @@
|
|
1 |
{
|
2 |
"best_metric": 0.11536111854119403,
|
3 |
"best_model_checkpoint": "/storage/elsayedissa/whisper-large-v2-spanish/checkpoint-1000",
|
4 |
-
"epoch": 0.
|
5 |
-
"global_step":
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
@@ -2496,11 +2496,260 @@
|
|
2496 |
"eval_steps_per_second": 0.062,
|
2497 |
"eval_wer": 0.10322861869805133,
|
2498 |
"step": 10000
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
2499 |
}
|
2500 |
],
|
2501 |
"max_steps": 25000,
|
2502 |
"num_train_epochs": 1,
|
2503 |
-
"total_flos": 3.
|
2504 |
"trial_name": null,
|
2505 |
"trial_params": null
|
2506 |
}
|
|
|
1 |
{
|
2 |
"best_metric": 0.11536111854119403,
|
3 |
"best_model_checkpoint": "/storage/elsayedissa/whisper-large-v2-spanish/checkpoint-1000",
|
4 |
+
"epoch": 0.3818251240931653,
|
5 |
+
"global_step": 11000,
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
|
|
2496 |
"eval_steps_per_second": 0.062,
|
2497 |
"eval_wer": 0.10322861869805133,
|
2498 |
"step": 10000
|
2499 |
+
},
|
2500 |
+
{
|
2501 |
+
"epoch": 0.35,
|
2502 |
+
"learning_rate": 6.115510204081633e-06,
|
2503 |
+
"loss": 0.1827,
|
2504 |
+
"step": 10025
|
2505 |
+
},
|
2506 |
+
{
|
2507 |
+
"epoch": 0.35,
|
2508 |
+
"learning_rate": 6.10530612244898e-06,
|
2509 |
+
"loss": 0.1653,
|
2510 |
+
"step": 10050
|
2511 |
+
},
|
2512 |
+
{
|
2513 |
+
"epoch": 0.35,
|
2514 |
+
"learning_rate": 6.095102040816327e-06,
|
2515 |
+
"loss": 0.1726,
|
2516 |
+
"step": 10075
|
2517 |
+
},
|
2518 |
+
{
|
2519 |
+
"epoch": 0.35,
|
2520 |
+
"learning_rate": 6.084897959183675e-06,
|
2521 |
+
"loss": 0.1569,
|
2522 |
+
"step": 10100
|
2523 |
+
},
|
2524 |
+
{
|
2525 |
+
"epoch": 0.35,
|
2526 |
+
"learning_rate": 6.074693877551021e-06,
|
2527 |
+
"loss": 0.1796,
|
2528 |
+
"step": 10125
|
2529 |
+
},
|
2530 |
+
{
|
2531 |
+
"epoch": 0.35,
|
2532 |
+
"learning_rate": 6.064489795918368e-06,
|
2533 |
+
"loss": 0.1457,
|
2534 |
+
"step": 10150
|
2535 |
+
},
|
2536 |
+
{
|
2537 |
+
"epoch": 0.35,
|
2538 |
+
"learning_rate": 6.054285714285715e-06,
|
2539 |
+
"loss": 0.1797,
|
2540 |
+
"step": 10175
|
2541 |
+
},
|
2542 |
+
{
|
2543 |
+
"epoch": 0.35,
|
2544 |
+
"learning_rate": 6.044081632653061e-06,
|
2545 |
+
"loss": 0.1596,
|
2546 |
+
"step": 10200
|
2547 |
+
},
|
2548 |
+
{
|
2549 |
+
"epoch": 0.35,
|
2550 |
+
"learning_rate": 6.033877551020409e-06,
|
2551 |
+
"loss": 0.1855,
|
2552 |
+
"step": 10225
|
2553 |
+
},
|
2554 |
+
{
|
2555 |
+
"epoch": 0.36,
|
2556 |
+
"learning_rate": 6.023673469387755e-06,
|
2557 |
+
"loss": 0.1495,
|
2558 |
+
"step": 10250
|
2559 |
+
},
|
2560 |
+
{
|
2561 |
+
"epoch": 0.36,
|
2562 |
+
"learning_rate": 6.013469387755102e-06,
|
2563 |
+
"loss": 0.1741,
|
2564 |
+
"step": 10275
|
2565 |
+
},
|
2566 |
+
{
|
2567 |
+
"epoch": 0.36,
|
2568 |
+
"learning_rate": 6.0032653061224495e-06,
|
2569 |
+
"loss": 0.1859,
|
2570 |
+
"step": 10300
|
2571 |
+
},
|
2572 |
+
{
|
2573 |
+
"epoch": 0.36,
|
2574 |
+
"learning_rate": 5.993061224489797e-06,
|
2575 |
+
"loss": 0.1874,
|
2576 |
+
"step": 10325
|
2577 |
+
},
|
2578 |
+
{
|
2579 |
+
"epoch": 0.36,
|
2580 |
+
"learning_rate": 5.982857142857144e-06,
|
2581 |
+
"loss": 0.159,
|
2582 |
+
"step": 10350
|
2583 |
+
},
|
2584 |
+
{
|
2585 |
+
"epoch": 0.36,
|
2586 |
+
"learning_rate": 5.97265306122449e-06,
|
2587 |
+
"loss": 0.1653,
|
2588 |
+
"step": 10375
|
2589 |
+
},
|
2590 |
+
{
|
2591 |
+
"epoch": 0.36,
|
2592 |
+
"learning_rate": 5.962448979591837e-06,
|
2593 |
+
"loss": 0.1417,
|
2594 |
+
"step": 10400
|
2595 |
+
},
|
2596 |
+
{
|
2597 |
+
"epoch": 0.36,
|
2598 |
+
"learning_rate": 5.952244897959184e-06,
|
2599 |
+
"loss": 0.1819,
|
2600 |
+
"step": 10425
|
2601 |
+
},
|
2602 |
+
{
|
2603 |
+
"epoch": 0.36,
|
2604 |
+
"learning_rate": 5.942040816326531e-06,
|
2605 |
+
"loss": 0.1605,
|
2606 |
+
"step": 10450
|
2607 |
+
},
|
2608 |
+
{
|
2609 |
+
"epoch": 0.36,
|
2610 |
+
"learning_rate": 5.931836734693878e-06,
|
2611 |
+
"loss": 0.1886,
|
2612 |
+
"step": 10475
|
2613 |
+
},
|
2614 |
+
{
|
2615 |
+
"epoch": 0.36,
|
2616 |
+
"learning_rate": 5.921632653061224e-06,
|
2617 |
+
"loss": 0.1667,
|
2618 |
+
"step": 10500
|
2619 |
+
},
|
2620 |
+
{
|
2621 |
+
"epoch": 0.37,
|
2622 |
+
"learning_rate": 5.911428571428573e-06,
|
2623 |
+
"loss": 0.1765,
|
2624 |
+
"step": 10525
|
2625 |
+
},
|
2626 |
+
{
|
2627 |
+
"epoch": 0.37,
|
2628 |
+
"learning_rate": 5.901224489795919e-06,
|
2629 |
+
"loss": 0.1484,
|
2630 |
+
"step": 10550
|
2631 |
+
},
|
2632 |
+
{
|
2633 |
+
"epoch": 0.37,
|
2634 |
+
"learning_rate": 5.891020408163266e-06,
|
2635 |
+
"loss": 0.1748,
|
2636 |
+
"step": 10575
|
2637 |
+
},
|
2638 |
+
{
|
2639 |
+
"epoch": 0.37,
|
2640 |
+
"learning_rate": 5.880816326530613e-06,
|
2641 |
+
"loss": 0.1514,
|
2642 |
+
"step": 10600
|
2643 |
+
},
|
2644 |
+
{
|
2645 |
+
"epoch": 0.37,
|
2646 |
+
"learning_rate": 5.87061224489796e-06,
|
2647 |
+
"loss": 0.191,
|
2648 |
+
"step": 10625
|
2649 |
+
},
|
2650 |
+
{
|
2651 |
+
"epoch": 0.37,
|
2652 |
+
"learning_rate": 5.860408163265307e-06,
|
2653 |
+
"loss": 0.154,
|
2654 |
+
"step": 10650
|
2655 |
+
},
|
2656 |
+
{
|
2657 |
+
"epoch": 0.37,
|
2658 |
+
"learning_rate": 5.850204081632653e-06,
|
2659 |
+
"loss": 0.1758,
|
2660 |
+
"step": 10675
|
2661 |
+
},
|
2662 |
+
{
|
2663 |
+
"epoch": 0.37,
|
2664 |
+
"learning_rate": 5.84e-06,
|
2665 |
+
"loss": 0.1675,
|
2666 |
+
"step": 10700
|
2667 |
+
},
|
2668 |
+
{
|
2669 |
+
"epoch": 0.37,
|
2670 |
+
"learning_rate": 5.829795918367347e-06,
|
2671 |
+
"loss": 0.1946,
|
2672 |
+
"step": 10725
|
2673 |
+
},
|
2674 |
+
{
|
2675 |
+
"epoch": 0.37,
|
2676 |
+
"learning_rate": 5.819591836734695e-06,
|
2677 |
+
"loss": 0.1563,
|
2678 |
+
"step": 10750
|
2679 |
+
},
|
2680 |
+
{
|
2681 |
+
"epoch": 0.37,
|
2682 |
+
"learning_rate": 5.809387755102042e-06,
|
2683 |
+
"loss": 0.177,
|
2684 |
+
"step": 10775
|
2685 |
+
},
|
2686 |
+
{
|
2687 |
+
"epoch": 0.37,
|
2688 |
+
"learning_rate": 5.799183673469388e-06,
|
2689 |
+
"loss": 0.1467,
|
2690 |
+
"step": 10800
|
2691 |
+
},
|
2692 |
+
{
|
2693 |
+
"epoch": 0.38,
|
2694 |
+
"learning_rate": 5.788979591836735e-06,
|
2695 |
+
"loss": 0.1677,
|
2696 |
+
"step": 10825
|
2697 |
+
},
|
2698 |
+
{
|
2699 |
+
"epoch": 0.38,
|
2700 |
+
"learning_rate": 5.778775510204082e-06,
|
2701 |
+
"loss": 0.1638,
|
2702 |
+
"step": 10850
|
2703 |
+
},
|
2704 |
+
{
|
2705 |
+
"epoch": 0.38,
|
2706 |
+
"learning_rate": 5.768571428571429e-06,
|
2707 |
+
"loss": 0.2078,
|
2708 |
+
"step": 10875
|
2709 |
+
},
|
2710 |
+
{
|
2711 |
+
"epoch": 0.38,
|
2712 |
+
"learning_rate": 5.758367346938776e-06,
|
2713 |
+
"loss": 0.1474,
|
2714 |
+
"step": 10900
|
2715 |
+
},
|
2716 |
+
{
|
2717 |
+
"epoch": 0.38,
|
2718 |
+
"learning_rate": 5.748163265306122e-06,
|
2719 |
+
"loss": 0.1758,
|
2720 |
+
"step": 10925
|
2721 |
+
},
|
2722 |
+
{
|
2723 |
+
"epoch": 0.38,
|
2724 |
+
"learning_rate": 5.73795918367347e-06,
|
2725 |
+
"loss": 0.1453,
|
2726 |
+
"step": 10950
|
2727 |
+
},
|
2728 |
+
{
|
2729 |
+
"epoch": 0.38,
|
2730 |
+
"learning_rate": 5.727755102040817e-06,
|
2731 |
+
"loss": 0.1806,
|
2732 |
+
"step": 10975
|
2733 |
+
},
|
2734 |
+
{
|
2735 |
+
"epoch": 0.38,
|
2736 |
+
"learning_rate": 5.717551020408164e-06,
|
2737 |
+
"loss": 0.1457,
|
2738 |
+
"step": 11000
|
2739 |
+
},
|
2740 |
+
{
|
2741 |
+
"epoch": 0.38,
|
2742 |
+
"eval_loss": 0.17772968113422394,
|
2743 |
+
"eval_runtime": 31110.5279,
|
2744 |
+
"eval_samples_per_second": 0.998,
|
2745 |
+
"eval_steps_per_second": 0.062,
|
2746 |
+
"eval_wer": 0.09984380314573199,
|
2747 |
+
"step": 11000
|
2748 |
}
|
2749 |
],
|
2750 |
"max_steps": 25000,
|
2751 |
"num_train_epochs": 1,
|
2752 |
+
"total_flos": 3.736780996608e+20,
|
2753 |
"trial_name": null,
|
2754 |
"trial_params": null
|
2755 |
}
|
pytorch_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 6173655480
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6bdd501f6153777369e4dc961faa1bf60f1f7535ded5d73c87f2ff06b5aa8064
|
3 |
size 6173655480
|
runs/Dec26_19-57-19_gpu07.cyverse.org/events.out.tfevents.1672109970.gpu07.cyverse.org.126369.0
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9c954f05ee34517b3eb4c4f0da7bbc163f490b1b0ac3836bb3e4f0baf08b45f5
|
3 |
+
size 76967
|