sanchit-gandhi HF staff commited on
Commit
bda8fc9
1 Parent(s): 58efde9

2hx8pk65: saving weights and logs of step 50k

Browse files
flax_model.msgpack CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8387a279a508da48c4b9b41b29ad79013e3f168689d639a6ed6588c9f0a4ed1d
3
  size 2353616717
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:372221fdf608b28a94812360d641b0852fa54d1adc2d7c546610d3200abb28d8
3
  size 2353616717
nohup.out CHANGED
The diff for this file is too large to render. See raw diff
 
wandb/run-20220828_085247-2hx8pk65/files/output.log CHANGED
@@ -43966,5 +43966,10170 @@ To disable this warning, you can either:
43966
  - Avoid using `tokenizers` before the fork if possible
43967
  - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
43968
  huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
43969
  To disable this warning, you can either:
43970
  - Avoid using `tokenizers` before the fork if possible
 
43966
  - Avoid using `tokenizers` before the fork if possible
43967
  - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
43968
  huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
43969
+ To disable this warning, you can either:
43970
+ - Avoid using `tokenizers` before the fork if possible
43971
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
43972
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
43973
+ To disable this warning, you can either:
43974
+ - Avoid using `tokenizers` before the fork if possible
43975
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
43976
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
43977
+ To disable this warning, you can either:
43978
+ - Avoid using `tokenizers` before the fork if possible
43979
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
43980
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
43981
+ To disable this warning, you can either:
43982
+ - Avoid using `tokenizers` before the fork if possible
43983
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
43984
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
43985
+ To disable this warning, you can either:
43986
+ - Avoid using `tokenizers` before the fork if possible
43987
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
43988
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
43989
+ To disable this warning, you can either:
43990
+ - Avoid using `tokenizers` before the fork if possible
43991
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
43992
+ Training...: 11% 463/4393 [45:57<182:26:10, 167.12s/it]
43993
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
43994
+ To disable this warning, you can either:
43995
+ - Avoid using `tokenizers` before the fork if possible
43996
+ return jax.tree_map(4393 [45:57<182:26:10, 167.12s/it]
43997
+
43998
+
43999
+
44000
+
44001
+
44002
+
44003
+
44004
+
44005
+
44006
+
44007
+
44008
+
44009
+
44010
+
44011
+
44012
+
44013
+
44014
+
44015
+
44016
+
44017
+
44018
+
44019
+
44020
+ return jax.tree_map(lambda x: x[0], tree) 4.64s/it]
44021
+ run_flax_speech_recognition_seq2seq.py:336: FutureWarning: jax.tree_map is deprecated, and will be removed in a future release. Use jax.tree_util.tree_map instead.
44022
+ return jax.tree_map(lambda x: x.astype(jnp.float32) if x.dtype == jnp.bfloat16 else x, t)
44023
+ Step... (40000/50000 | Eval Loss: 0.9527401328086853 | Eval wer: 0.04187346053453917 | Eval cer: 0.02872188479352137 |): 75% 9/12 [53:25:09<17:31:17, 21025.84s/it]
44024
+
44025
+
44026
+
44027
+
44028
+
44029
+
44030
+
44031
+
44032
+
44033
+
44034
+
44035
+
44036
+
44037
+
44038
+
44039
+
44040
+
44041
+
44042
+
44043
+
44044
+
44045
+
44046
+
44047
+
44048
+
44049
+
44050
+
44051
+
44052
+
44053
+
44054
+
44055
+
44056
+
44057
+
44058
+
44059
+
44060
+
44061
+
44062
+
44063
+
44064
+
44065
+
44066
+
44067
+
44068
+
44069
+
44070
+
44071
+
44072
+
44073
+
44074
+
44075
+
44076
+
44077
+
44078
+
44079
+
44080
+
44081
+
44082
+
44083
+
44084
+
44085
+
44086
+
44087
+
44088
+
44089
+
44090
+
44091
+
44092
+
44093
+
44094
+
44095
+
44096
+
44097
+
44098
+
44099
+
44100
+
44101
+
44102
+
44103
+
44104
+
44105
+
44106
+
44107
+
44108
+
44109
+
44110
+
44111
+
44112
+
44113
+
44114
+
44115
+
44116
+
44117
+
44118
+
44119
+
44120
+
44121
+
44122
+
44123
+
44124
+
44125
+
44126
+
44127
+
44128
+
44129
+
44130
+
44131
+
44132
+
44133
+
44134
+
44135
+
44136
+
44137
+
44138
+
44139
+
44140
+
44141
+
44142
+
44143
+
44144
+
44145
+
44146
+
44147
+
44148
+
44149
+
44150
+
44151
+
44152
+
44153
+
44154
+
44155
+
44156
+
44157
+
44158
+
44159
+
44160
+
44161
+
44162
+
44163
+
44164
+
44165
+
44166
+
44167
+
44168
+
44169
+
44170
+
44171
+
44172
+
44173
+
44174
+
44175
+
44176
+
44177
+
44178
+
44179
+
44180
+
44181
+
44182
+
44183
+
44184
+
44185
+
44186
+
44187
+
44188
+
44189
+
44190
+
44191
+
44192
+
44193
+
44194
+
44195
+
44196
+
44197
+
44198
+
44199
+
44200
+
44201
+
44202
+
44203
+
44204
+
44205
+
44206
+
44207
+
44208
+
44209
+
44210
+
44211
+
44212
+
44213
+
44214
+
44215
+
44216
+
44217
+
44218
+
44219
+
44220
+
44221
+
44222
+
44223
+
44224
+
44225
+
44226
+
44227
+
44228
+
44229
+
44230
+
44231
+
44232
+
44233
+
44234
+
44235
+
44236
+
44237
+
44238
+
44239
+
44240
+
44241
+
44242
+
44243
+
44244
+
44245
+
44246
+
44247
+
44248
+
44249
+
44250
+
44251
+
44252
+
44253
+
44254
+
44255
+
44256
+
44257
+
44258
+
44259
+
44260
+
44261
+
44262
+
44263
+
44264
+
44265
+
44266
+
44267
+
44268
+
44269
+
44270
+
44271
+
44272
+
44273
+
44274
+
44275
+
44276
+
44277
+
44278
+
44279
+
44280
+
44281
+
44282
+
44283
+
44284
+
44285
+
44286
+
44287
+
44288
+
44289
+
44290
+
44291
+
44292
+
44293
+
44294
+
44295
+
44296
+
44297
+
44298
+
44299
+
44300
+
44301
+
44302
+
44303
+
44304
+
44305
+
44306
+
44307
+
44308
+
44309
+
44310
+
44311
+
44312
+
44313
+
44314
+
44315
+
44316
+
44317
+
44318
+
44319
+
44320
+
44321
+
44322
+
44323
+
44324
+
44325
+
44326
+
44327
+
44328
+
44329
+
44330
+
44331
+
44332
+
44333
+
44334
+
44335
+
44336
+
44337
+
44338
+
44339
+
44340
+
44341
+
44342
+
44343
+
44344
+
44345
+
44346
+
44347
+
44348
+
44349
+
44350
+
44351
+
44352
+
44353
+
44354
+
44355
+
44356
+
44357
+
44358
+
44359
+
44360
+
44361
+
44362
+
44363
+
44364
+
44365
+
44366
+
44367
+
44368
+
44369
+
44370
+
44371
+
44372
+
44373
+
44374
+
44375
+
44376
+
44377
+
44378
+
44379
+
44380
+
44381
+
44382
+
44383
+
44384
+
44385
+
44386
+
44387
+
44388
+
44389
+
44390
+
44391
+
44392
+
44393
+
44394
+
44395
+
44396
+
44397
+
44398
+
44399
+
44400
+
44401
+
44402
+
44403
+
44404
+
44405
+
44406
+
44407
+
44408
+
44409
+
44410
+
44411
+
44412
+
44413
+
44414
+
44415
+
44416
+
44417
+
44418
+
44419
+
44420
+
44421
+
44422
+
44423
+
44424
+
44425
+
44426
+
44427
+
44428
+
44429
+
44430
+
44431
+
44432
+
44433
+
44434
+
44435
+
44436
+
44437
+
44438
+
44439
+
44440
+
44441
+
44442
+
44443
+
44444
+
44445
+
44446
+
44447
+
44448
+
44449
+
44450
+
44451
+
44452
+
44453
+
44454
+
44455
+
44456
+
44457
+
44458
+
44459
+
44460
+
44461
+
44462
+
44463
+
44464
+
44465
+
44466
+
44467
+
44468
+
44469
+
44470
+
44471
+
44472
+
44473
+
44474
+
44475
+
44476
+
44477
+
44478
+
44479
+
44480
+
44481
+
44482
+
44483
+
44484
+
44485
+
44486
+
44487
+
44488
+
44489
+
44490
+
44491
+
44492
+
44493
+
44494
+
44495
+
44496
+
44497
+
44498
+
44499
+
44500
+
44501
+
44502
+
44503
+
44504
+
44505
+
44506
+
44507
+
44508
+
44509
+
44510
+
44511
+
44512
+
44513
+
44514
+
44515
+
44516
+
44517
+
44518
+
44519
+
44520
+
44521
+
44522
+
44523
+
44524
+
44525
+
44526
+
44527
+
44528
+
44529
+
44530
+
44531
+
44532
+
44533
+
44534
+
44535
+
44536
+
44537
+
44538
+
44539
+
44540
+
44541
+
44542
+
44543
+
44544
+
44545
+
44546
+
44547
+
44548
+
44549
+
44550
+
44551
+
44552
+
44553
+
44554
+
44555
+
44556
+
44557
+
44558
+
44559
+
44560
+
44561
+
44562
+
44563
+
44564
+
44565
+
44566
+
44567
+
44568
+
44569
+
44570
+
44571
+
44572
+
44573
+
44574
+
44575
+
44576
+
44577
+
44578
+
44579
+
44580
+
44581
+
44582
+
44583
+
44584
+
44585
+
44586
+
44587
+
44588
+
44589
+
44590
+
44591
+
44592
+
44593
+
44594
+
44595
+
44596
+
44597
+
44598
+
44599
+
44600
+
44601
+
44602
+
44603
+
44604
+
44605
+
44606
+
44607
+
44608
+
44609
+
44610
+
44611
+
44612
+
44613
+
44614
+
44615
+
44616
+
44617
+
44618
+
44619
+
44620
+
44621
+
44622
+
44623
+
44624
+
44625
+
44626
+
44627
+
44628
+
44629
+
44630
+
44631
+
44632
+
44633
+
44634
+
44635
+
44636
+
44637
+
44638
+
44639
+
44640
+
44641
+
44642
+
44643
+
44644
+
44645
+
44646
+
44647
+
44648
+
44649
+
44650
+
44651
+
44652
+
44653
+
44654
+
44655
+
44656
+
44657
+
44658
+
44659
+
44660
+
44661
+
44662
+
44663
+
44664
+
44665
+
44666
+
44667
+
44668
+
44669
+
44670
+
44671
+
44672
+
44673
+
44674
+
44675
+
44676
+
44677
+
44678
+
44679
+
44680
+
44681
+
44682
+
44683
+
44684
+
44685
+
44686
+
44687
+
44688
+
44689
+
44690
+
44691
+
44692
+
44693
+
44694
+
44695
+
44696
+
44697
+
44698
+
44699
+
44700
+
44701
+
44702
+
44703
+
44704
+
44705
+
44706
+
44707
+
44708
+
44709
+
44710
+
44711
+
44712
+
44713
+
44714
+
44715
+
44716
+
44717
+
44718
+
44719
+
44720
+
44721
+
44722
+
44723
+
44724
+
44725
+
44726
+
44727
+
44728
+
44729
+
44730
+
44731
+
44732
+
44733
+
44734
+
44735
+
44736
+
44737
+
44738
+
44739
+
44740
+
44741
+
44742
+
44743
+
44744
+
44745
+
44746
+
44747
+
44748
+
44749
+
44750
+
44751
+
44752
+
44753
+
44754
+
44755
+
44756
+
44757
+
44758
+
44759
+
44760
+
44761
+
44762
+
44763
+
44764
+
44765
+
44766
+
44767
+
44768
+
44769
+
44770
+
44771
+
44772
+
44773
+
44774
+
44775
+
44776
+
44777
+
44778
+
44779
+
44780
+
44781
+
44782
+
44783
+
44784
+
44785
+
44786
+
44787
+
44788
+
44789
+
44790
+
44791
+
44792
+
44793
+
44794
+
44795
+
44796
+
44797
+
44798
+
44799
+
44800
+
44801
+
44802
+
44803
+
44804
+
44805
+
44806
+
44807
+
44808
+
44809
+
44810
+
44811
+
44812
+
44813
+
44814
+
44815
+
44816
+
44817
+
44818
+
44819
+
44820
+
44821
+
44822
+
44823
+
44824
+
44825
+
44826
+
44827
+
44828
+
44829
+
44830
+
44831
+
44832
+
44833
+
44834
+
44835
+
44836
+
44837
+
44838
+
44839
+
44840
+
44841
+
44842
+
44843
+
44844
+
44845
+
44846
+
44847
+
44848
+
44849
+
44850
+
44851
+
44852
+
44853
+
44854
+
44855
+
44856
+
44857
+
44858
+
44859
+
44860
+
44861
+
44862
+
44863
+
44864
+
44865
+
44866
+
44867
+
44868
+
44869
+
44870
+
44871
+
44872
+
44873
+
44874
+
44875
+
44876
+
44877
+
44878
+
44879
+
44880
+
44881
+
44882
+
44883
+
44884
+
44885
+
44886
+
44887
+
44888
+
44889
+
44890
+
44891
+
44892
+
44893
+
44894
+
44895
+
44896
+
44897
+
44898
+
44899
+
44900
+
44901
+
44902
+
44903
+
44904
+
44905
+
44906
+
44907
+
44908
+
44909
+
44910
+
44911
+
44912
+
44913
+
44914
+
44915
+
44916
+
44917
+
44918
+
44919
+
44920
+
44921
+
44922
+
44923
+
44924
+
44925
+
44926
+
44927
+
44928
+
44929
+
44930
+
44931
+
44932
+
44933
+
44934
+
44935
+
44936
+
44937
+
44938
+
44939
+
44940
+
44941
+
44942
+
44943
+
44944
+
44945
+
44946
+
44947
+
44948
+
44949
+
44950
+
44951
+
44952
+
44953
+
44954
+
44955
+
44956
+
44957
+
44958
+
44959
+
44960
+
44961
+
44962
+
44963
+
44964
+
44965
+
44966
+
44967
+
44968
+
44969
+
44970
+
44971
+
44972
+
44973
+
44974
+
44975
+
44976
+
44977
+
44978
+
44979
+
44980
+
44981
+
44982
+
44983
+
44984
+
44985
+
44986
+
44987
+
44988
+
44989
+
44990
+
44991
+
44992
+
44993
+
44994
+
44995
+
44996
+
44997
+
44998
+
44999
+
45000
+
45001
+
45002
+
45003
+
45004
+
45005
+
45006
+
45007
+
45008
+
45009
+
45010
+
45011
+
45012
+
45013
+
45014
+
45015
+
45016
+
45017
+
45018
+
45019
+
45020
+
45021
+
45022
+
45023
+
45024
+
45025
+
45026
+
45027
+
45028
+
45029
+
45030
+
45031
+
45032
+
45033
+
45034
+
45035
+
45036
+
45037
+
45038
+
45039
+
45040
+
45041
+
45042
+
45043
+
45044
+
45045
+
45046
+
45047
+
45048
+
45049
+
45050
+
45051
+
45052
+
45053
+
45054
+
45055
+
45056
+
45057
+
45058
+
45059
+
45060
+
45061
+
45062
+
45063
+
45064
+
45065
+
45066
+
45067
+
45068
+
45069
+
45070
+
45071
+
45072
+
45073
+
45074
+
45075
+
45076
+
45077
+
45078
+
45079
+
45080
+
45081
+
45082
+
45083
+
45084
+
45085
+
45086
+
45087
+
45088
+
45089
+
45090
+
45091
+
45092
+
45093
+
45094
+
45095
+
45096
+
45097
+
45098
+
45099
+
45100
+
45101
+
45102
+
45103
+
45104
+
45105
+
45106
+
45107
+
45108
+
45109
+
45110
+
45111
+
45112
+
45113
+
45114
+
45115
+
45116
+
45117
+
45118
+
45119
+
45120
+
45121
+
45122
+
45123
+
45124
+
45125
+
45126
+
45127
+
45128
+
45129
+
45130
+
45131
+
45132
+
45133
+
45134
+
45135
+
45136
+
45137
+
45138
+
45139
+
45140
+
45141
+
45142
+
45143
+
45144
+
45145
+
45146
+
45147
+
45148
+
45149
+
45150
+
45151
+
45152
+
45153
+
45154
+
45155
+
45156
+
45157
+
45158
+
45159
+
45160
+
45161
+
45162
+
45163
+
45164
+
45165
+
45166
+
45167
+
45168
+
45169
+
45170
+
45171
+
45172
+
45173
+
45174
+
45175
+
45176
+
45177
+
45178
+
45179
+
45180
+
45181
+
45182
+
45183
+
45184
+
45185
+
45186
+
45187
+
45188
+
45189
+
45190
+
45191
+
45192
+
45193
+
45194
+
45195
+
45196
+
45197
+
45198
+
45199
+
45200
+
45201
+
45202
+
45203
+
45204
+
45205
+
45206
+
45207
+
45208
+
45209
+
45210
+
45211
+
45212
+
45213
+
45214
+
45215
+
45216
+
45217
+
45218
+
45219
+
45220
+
45221
+
45222
+
45223
+
45224
+
45225
+
45226
+
45227
+
45228
+
45229
+
45230
+
45231
+
45232
+
45233
+
45234
+
45235
+
45236
+
45237
+
45238
+
45239
+
45240
+
45241
+
45242
+
45243
+
45244
+
45245
+
45246
+
45247
+
45248
+
45249
+
45250
+
45251
+
45252
+
45253
+
45254
+
45255
+
45256
+
45257
+
45258
+
45259
+
45260
+
45261
+
45262
+
45263
+
45264
+
45265
+
45266
+
45267
+
45268
+
45269
+
45270
+
45271
+
45272
+
45273
+
45274
+
45275
+
45276
+
45277
+
45278
+
45279
+
45280
+
45281
+
45282
+
45283
+
45284
+
45285
+
45286
+
45287
+
45288
+
45289
+
45290
+
45291
+
45292
+
45293
+
45294
+
45295
+
45296
+
45297
+
45298
+
45299
+
45300
+
45301
+
45302
+
45303
+
45304
+
45305
+
45306
+
45307
+
45308
+
45309
+
45310
+
45311
+
45312
+
45313
+
45314
+
45315
+
45316
+
45317
+
45318
+
45319
+
45320
+
45321
+
45322
+
45323
+
45324
+
45325
+
45326
+
45327
+
45328
+
45329
+
45330
+
45331
+
45332
+
45333
+
45334
+
45335
+
45336
+
45337
+
45338
+
45339
+
45340
+
45341
+
45342
+
45343
+
45344
+
45345
+
45346
+
45347
+
45348
+
45349
+
45350
+
45351
+
45352
+
45353
+
45354
+
45355
+
45356
+
45357
+
45358
+
45359
+
45360
+
45361
+
45362
+
45363
+
45364
+
45365
+
45366
+
45367
+
45368
+
45369
+
45370
+
45371
+
45372
+
45373
+
45374
+
45375
+
45376
+
45377
+
45378
+
45379
+
45380
+
45381
+
45382
+
45383
+
45384
+
45385
+
45386
+
45387
+
45388
+
45389
+
45390
+
45391
+
45392
+
45393
+
45394
+
45395
+
45396
+
45397
+
45398
+
45399
+
45400
+
45401
+
45402
+
45403
+
45404
+
45405
+
45406
+
45407
+
45408
+
45409
+
45410
+
45411
+
45412
+
45413
+
45414
+
45415
+
45416
+
45417
+
45418
+
45419
+
45420
+
45421
+
45422
+
45423
+
45424
+
45425
+
45426
+
45427
+
45428
+
45429
+
45430
+
45431
+
45432
+
45433
+
45434
+
45435
+
45436
+
45437
+
45438
+
45439
+
45440
+
45441
+
45442
+
45443
+
45444
+
45445
+
45446
+
45447
+
45448
+
45449
+
45450
+
45451
+
45452
+
45453
+
45454
+
45455
+
45456
+
45457
+
45458
+
45459
+
45460
+
45461
+
45462
+
45463
+
45464
+
45465
+
45466
+
45467
+
45468
+
45469
+
45470
+
45471
+
45472
+
45473
+
45474
+
45475
+
45476
+
45477
+
45478
+
45479
+
45480
+
45481
+
45482
+
45483
+
45484
+
45485
+
45486
+
45487
+
45488
+
45489
+
45490
+
45491
+
45492
+
45493
+
45494
+
45495
+
45496
+
45497
+
45498
+
45499
+
45500
+
45501
+
45502
+
45503
+
45504
+
45505
+
45506
+
45507
+
45508
+
45509
+
45510
+
45511
+
45512
+
45513
+
45514
+
45515
+
45516
+
45517
+
45518
+
45519
+
45520
+
45521
+
45522
+
45523
+
45524
+
45525
+
45526
+
45527
+
45528
+
45529
+
45530
+
45531
+
45532
+
45533
+
45534
+
45535
+
45536
+
45537
+
45538
+
45539
+
45540
+
45541
+
45542
+
45543
+
45544
+
45545
+
45546
+
45547
+
45548
+
45549
+
45550
+
45551
+
45552
+
45553
+
45554
+
45555
+
45556
+
45557
+
45558
+
45559
+
45560
+
45561
+
45562
+
45563
+
45564
+
45565
+
45566
+
45567
+
45568
+
45569
+
45570
+
45571
+
45572
+
45573
+
45574
+
45575
+
45576
+
45577
+
45578
+
45579
+
45580
+
45581
+
45582
+
45583
+
45584
+
45585
+
45586
+
45587
+
45588
+
45589
+
45590
+
45591
+
45592
+
45593
+
45594
+
45595
+
45596
+
45597
+
45598
+
45599
+
45600
+
45601
+
45602
+
45603
+
45604
+
45605
+
45606
+
45607
+
45608
+
45609
+
45610
+
45611
+
45612
+
45613
+
45614
+
45615
+
45616
+
45617
+
45618
+
45619
+
45620
+
45621
+
45622
+
45623
+
45624
+
45625
+
45626
+
45627
+
45628
+
45629
+
45630
+
45631
+
45632
+
45633
+
45634
+
45635
+
45636
+
45637
+
45638
+
45639
+
45640
+
45641
+
45642
+
45643
+
45644
+
45645
+
45646
+
45647
+
45648
+
45649
+
45650
+
45651
+
45652
+
45653
+
45654
+
45655
+ Training...: 49% 2162/4393 [3:00:16<3:33:52, 5.75s/it]
45656
+ Step... (40000/50000 | Eval Loss: 0.9527401328086853 | Eval wer: 0.04187346053453917 | Eval cer: 0.02872188479352137 |)
45657
+ Step... (40025 | Loss: 0.008939584717154503, Learning Rate: 2.0153533114353195e-05, Gradient Norm: 0.29531925916671753)
45658
+ Step... (40050 | Loss: 0.013731228187680244, Learning Rate: 2.0103030692553148e-05, Gradient Norm: 0.28097018599510193)
45659
+ Step... (40075 | Loss: 0.005788063630461693, Learning Rate: 2.00525282707531e-05, Gradient Norm: 0.2137639969587326)
45660
+ Step... (40100 | Loss: 0.006226960103958845, Learning Rate: 2.000201857299544e-05, Gradient Norm: 0.1285054236650467)
45661
+ Step... (40125 | Loss: 0.007349521853029728, Learning Rate: 1.9951516151195392e-05, Gradient Norm: 0.233669713139534)
45662
+ Step... (40150 | Loss: 0.010777842253446579, Learning Rate: 1.9901013729395345e-05, Gradient Norm: 0.24161839485168457)
45663
+ Step... (40175 | Loss: 0.0042556338012218475, Learning Rate: 1.9850504031637684e-05, Gradient Norm: 0.1810857057571411)
45664
+ Step... (40200 | Loss: 0.013205939903855324, Learning Rate: 1.9800001609837636e-05, Gradient Norm: 0.2645622491836548)
45665
+ Step... (40225 | Loss: 0.010763286612927914, Learning Rate: 1.9749491912079975e-05, Gradient Norm: 0.2870539128780365)
45666
+ Step... (40250 | Loss: 0.012334151193499565, Learning Rate: 1.9698989490279928e-05, Gradient Norm: 0.2389720231294632)
45667
+ Step... (40275 | Loss: 0.0026933024637401104, Learning Rate: 1.9648485249490477e-05, Gradient Norm: 0.09100257605314255)
45668
+ Step... (40300 | Loss: 0.007666177581995726, Learning Rate: 1.959797737072222e-05, Gradient Norm: 0.14644892513751984)
45669
+ Step... (40325 | Loss: 0.006105078849941492, Learning Rate: 1.954747312993277e-05, Gradient Norm: 0.1783716231584549)
45670
+ Step... (40350 | Loss: 0.007519055623561144, Learning Rate: 1.949697070813272e-05, Gradient Norm: 0.196736142039299)
45671
+ Step... (40375 | Loss: 0.008998542092740536, Learning Rate: 1.944646101037506e-05, Gradient Norm: 0.26219305396080017)
45672
+ Step... (40400 | Loss: 0.006001105532050133, Learning Rate: 1.9395958588575013e-05, Gradient Norm: 0.17653429508209229)
45673
+ Step... (40425 | Loss: 0.003485818626359105, Learning Rate: 1.9345456166774966e-05, Gradient Norm: 0.1488671600818634)
45674
+ Step... (40450 | Loss: 0.011663154698908329, Learning Rate: 1.9294946469017304e-05, Gradient Norm: 0.21771906316280365)
45675
+ Step... (40475 | Loss: 0.01133037731051445, Learning Rate: 1.9244444047217257e-05, Gradient Norm: 0.3833330273628235)
45676
+ Step... (40500 | Loss: 0.004920892883092165, Learning Rate: 1.919394162541721e-05, Gradient Norm: 0.1484016329050064)
45677
+ Step... (40525 | Loss: 0.016179556027054787, Learning Rate: 1.914343192765955e-05, Gradient Norm: 0.27268052101135254)
45678
+ Step... (40550 | Loss: 0.007338401861488819, Learning Rate: 1.90929295058595e-05, Gradient Norm: 0.18785510957241058)
45679
+ Step... (40575 | Loss: 0.003818551544100046, Learning Rate: 1.904242526507005e-05, Gradient Norm: 0.1895657181739807)
45680
+ Step... (40600 | Loss: 0.008862589485943317, Learning Rate: 1.8991917386301793e-05, Gradient Norm: 0.2292691171169281)
45681
+ Step... (40625 | Loss: 0.0069699776358902454, Learning Rate: 1.8941413145512342e-05, Gradient Norm: 0.1485799252986908)
45682
+ Step... (40650 | Loss: 0.010429407469928265, Learning Rate: 1.8890910723712295e-05, Gradient Norm: 0.23329854011535645)
45683
+ Step... (40675 | Loss: 0.005073365289717913, Learning Rate: 1.8840401025954634e-05, Gradient Norm: 0.1328592151403427)
45684
+ Step... (40700 | Loss: 0.0052896239794790745, Learning Rate: 1.8789898604154587e-05, Gradient Norm: 0.1811048984527588)
45685
+ Step... (40725 | Loss: 0.0049026296474039555, Learning Rate: 1.873939618235454e-05, Gradient Norm: 0.1749919205904007)
45686
+ Step... (40750 | Loss: 0.004890045616775751, Learning Rate: 1.8688886484596878e-05, Gradient Norm: 0.15074510872364044)
45687
+ Step... (40775 | Loss: 0.003988181706517935, Learning Rate: 1.863838406279683e-05, Gradient Norm: 0.19411082565784454)
45688
+ Step... (40800 | Loss: 0.00978496391326189, Learning Rate: 1.8587881640996784e-05, Gradient Norm: 0.20905667543411255)
45689
+ Step... (40825 | Loss: 0.006088871043175459, Learning Rate: 1.8537371943239123e-05, Gradient Norm: 0.3005000054836273)
45690
+ Step... (40850 | Loss: 0.010514942929148674, Learning Rate: 1.8486869521439075e-05, Gradient Norm: 0.2221343070268631)
45691
+ Step... (40875 | Loss: 0.006962936837226152, Learning Rate: 1.8436365280649625e-05, Gradient Norm: 0.19761787354946136)
45692
+ Step... (40900 | Loss: 0.011204604059457779, Learning Rate: 1.8385857401881367e-05, Gradient Norm: 0.25883570313453674)
45693
+ Step... (40925 | Loss: 0.005226748064160347, Learning Rate: 1.8335353161091916e-05, Gradient Norm: 0.2500029504299164)
45694
+ Step... (40950 | Loss: 0.010302098467946053, Learning Rate: 1.828485073929187e-05, Gradient Norm: 0.18769465386867523)
45695
+ Step... (40975 | Loss: 0.005125986412167549, Learning Rate: 1.8234341041534208e-05, Gradient Norm: 0.2729404866695404)
45696
+ Step... (41000 | Loss: 0.010441966354846954, Learning Rate: 1.818383861973416e-05, Gradient Norm: 0.3461083769798279)
45697
+ Step... (41025 | Loss: 0.0025685643777251244, Learning Rate: 1.8133336197934113e-05, Gradient Norm: 0.12441041320562363)
45698
+ Step... (41050 | Loss: 0.005768472328782082, Learning Rate: 1.8082826500176452e-05, Gradient Norm: 0.16698002815246582)
45699
+ Step... (41075 | Loss: 0.00754266232252121, Learning Rate: 1.8032324078376405e-05, Gradient Norm: 0.2963465750217438)
45700
+ Step... (41100 | Loss: 0.010014562867581844, Learning Rate: 1.7981821656576358e-05, Gradient Norm: 0.16458502411842346)
45701
+ Step... (41125 | Loss: 0.0019032791024073958, Learning Rate: 1.7931311958818696e-05, Gradient Norm: 0.06771261245012283)
45702
+ Step... (41150 | Loss: 0.007408216595649719, Learning Rate: 1.788080953701865e-05, Gradient Norm: 0.1833212673664093)
45703
+ Step... (41175 | Loss: 0.006798009388148785, Learning Rate: 1.78303052962292e-05, Gradient Norm: 0.28412139415740967)
45704
+ Step... (41200 | Loss: 0.010547134093940258, Learning Rate: 1.777979741746094e-05, Gradient Norm: 0.6392953991889954)
45705
+ Step... (41225 | Loss: 0.002890239469707012, Learning Rate: 1.772929317667149e-05, Gradient Norm: 0.14131464064121246)
45706
+ Step... (41250 | Loss: 0.012237129732966423, Learning Rate: 1.7678790754871443e-05, Gradient Norm: 0.24268372356891632)
45707
+ Step... (41275 | Loss: 0.003835662268102169, Learning Rate: 1.762828105711378e-05, Gradient Norm: 0.20543915033340454)
45708
+ Step... (41300 | Loss: 0.009825979359447956, Learning Rate: 1.7577778635313734e-05, Gradient Norm: 0.2134653478860855)
45709
+ Step... (41325 | Loss: 0.0023147186730057, Learning Rate: 1.7527270756545477e-05, Gradient Norm: 0.10707230120897293)
45710
+ Step... (41350 | Loss: 0.0070909978821873665, Learning Rate: 1.7476766515756026e-05, Gradient Norm: 0.17582404613494873)
45711
+ Step... (41375 | Loss: 0.0033097451087087393, Learning Rate: 1.742626409395598e-05, Gradient Norm: 0.13646838068962097)
45712
+ Step... (41400 | Loss: 0.008395623415708542, Learning Rate: 1.7375754396198317e-05, Gradient Norm: 0.1624133586883545)
45713
+ Step... (41425 | Loss: 0.004059635568410158, Learning Rate: 1.732525197439827e-05, Gradient Norm: 0.21894600987434387)
45714
+ Step... (41450 | Loss: 0.00695464015007019, Learning Rate: 1.7274749552598223e-05, Gradient Norm: 0.1792638897895813)
45715
+ Step... (41475 | Loss: 0.009820655919611454, Learning Rate: 1.7224239854840562e-05, Gradient Norm: 0.4551573097705841)
45716
+ Step... (41500 | Loss: 0.006216324865818024, Learning Rate: 1.7173737433040515e-05, Gradient Norm: 0.17348088324069977)
45717
+ Step... (41525 | Loss: 0.008928379975259304, Learning Rate: 1.7123233192251064e-05, Gradient Norm: 0.32115671038627625)
45718
+ Step... (41550 | Loss: 0.012027319520711899, Learning Rate: 1.7072725313482806e-05, Gradient Norm: 0.2573080360889435)
45719
+ Step... (41575 | Loss: 0.007203887682408094, Learning Rate: 1.7022221072693355e-05, Gradient Norm: 0.24148206412792206)
45720
+ Step... (41600 | Loss: 0.01733251102268696, Learning Rate: 1.6971718650893308e-05, Gradient Norm: 0.28181859850883484)
45721
+ Step... (41625 | Loss: 0.00894920527935028, Learning Rate: 1.6921208953135647e-05, Gradient Norm: 0.282490611076355)
45722
+ Step... (41650 | Loss: 0.009486709721386433, Learning Rate: 1.68707065313356e-05, Gradient Norm: 0.2058715522289276)
45723
+
45724
+
45725
+
45726
+
45727
+
45728
+
45729
+
45730
+
45731
+
45732
+
45733
+
45734
+
45735
+
45736
+
45737
+
45738
+
45739
+
45740
+
45741
+
45742
+
45743
+
45744
+
45745
+
45746
+
45747
+
45748
+
45749
+
45750
+
45751
+
45752
+
45753
+
45754
+
45755
+
45756
+
45757
+
45758
+
45759
+
45760
+
45761
+
45762
+
45763
+
45764
+
45765
+
45766
+
45767
+
45768
+
45769
+
45770
+
45771
+
45772
+
45773
+
45774
+
45775
+
45776
+
45777
+
45778
+
45779
+
45780
+
45781
+
45782
+
45783
+
45784
+
45785
+
45786
+
45787
+
45788
+
45789
+
45790
+
45791
+
45792
+
45793
+
45794
+
45795
+
45796
+
45797
+
45798
+
45799
+
45800
+
45801
+
45802
+
45803
+
45804
+
45805
+
45806
+
45807
+
45808
+
45809
+
45810
+
45811
+
45812
+
45813
+
45814
+
45815
+
45816
+
45817
+
45818
+
45819
+
45820
+
45821
+
45822
+
45823
+
45824
+
45825
+
45826
+
45827
+
45828
+
45829
+
45830
+
45831
+
45832
+
45833
+
45834
+
45835
+
45836
+
45837
+
45838
+
45839
+
45840
+
45841
+
45842
+
45843
+
45844
+
45845
+
45846
+
45847
+
45848
+
45849
+
45850
+
45851
+
45852
+
45853
+
45854
+
45855
+
45856
+
45857
+
45858
+
45859
+
45860
+
45861
+
45862
+
45863
+
45864
+
45865
+
45866
+
45867
+
45868
+
45869
+
45870
+
45871
+
45872
+
45873
+
45874
+
45875
+
45876
+
45877
+
45878
+
45879
+
45880
+
45881
+
45882
+
45883
+
45884
+
45885
+
45886
+
45887
+
45888
+
45889
+
45890
+
45891
+
45892
+
45893
+
45894
+
45895
+
45896
+
45897
+
45898
+
45899
+
45900
+
45901
+
45902
+
45903
+
45904
+
45905
+
45906
+
45907
+
45908
+
45909
+
45910
+
45911
+
45912
+
45913
+
45914
+
45915
+
45916
+
45917
+
45918
+
45919
+
45920
+
45921
+
45922
+
45923
+
45924
+
45925
+
45926
+
45927
+
45928
+
45929
+
45930
+
45931
+
45932
+
45933
+
45934
+
45935
+
45936
+
45937
+
45938
+
45939
+
45940
+
45941
+
45942
+
45943
+
45944
+
45945
+
45946
+
45947
+
45948
+
45949
+
45950
+
45951
+
45952
+
45953
+
45954
+
45955
+
45956
+
45957
+
45958
+
45959
+
45960
+
45961
+
45962
+
45963
+
45964
+
45965
+
45966
+
45967
+
45968
+
45969
+
45970
+
45971
+
45972
+
45973
+
45974
+
45975
+
45976
+
45977
+
45978
+
45979
+
45980
+
45981
+
45982
+
45983
+
45984
+
45985
+
45986
+
45987
+
45988
+
45989
+
45990
+
45991
+
45992
+
45993
+
45994
+
45995
+
45996
+
45997
+
45998
+
45999
+
46000
+
46001
+
46002
+
46003
+
46004
+
46005
+
46006
+
46007
+
46008
+
46009
+
46010
+
46011
+
46012
+
46013
+
46014
+
46015
+
46016
+
46017
+
46018
+
46019
+
46020
+
46021
+
46022
+
46023
+
46024
+
46025
+
46026
+
46027
+
46028
+
46029
+
46030
+
46031
+
46032
+
46033
+
46034
+
46035
+
46036
+
46037
+
46038
+
46039
+
46040
+
46041
+
46042
+
46043
+
46044
+
46045
+
46046
+
46047
+
46048
+
46049
+
46050
+
46051
+
46052
+
46053
+
46054
+
46055
+
46056
+
46057
+
46058
+
46059
+
46060
+
46061
+
46062
+
46063
+
46064
+
46065
+
46066
+
46067
+
46068
+
46069
+
46070
+
46071
+
46072
+
46073
+
46074
+
46075
+
46076
+
46077
+
46078
+
46079
+
46080
+
46081
+
46082
+
46083
+
46084
+
46085
+
46086
+
46087
+
46088
+
46089
+
46090
+
46091
+
46092
+
46093
+
46094
+
46095
+
46096
+
46097
+
46098
+
46099
+
46100
+
46101
+
46102
+
46103
+
46104
+
46105
+
46106
+
46107
+
46108
+
46109
+
46110
+
46111
+
46112
+
46113
+
46114
+
46115
+
46116
+
46117
+
46118
+
46119
+
46120
+
46121
+
46122
+
46123
+
46124
+
46125
+
46126
+
46127
+
46128
+
46129
+
46130
+
46131
+
46132
+
46133
+
46134
+
46135
+
46136
+
46137
+
46138
+
46139
+
46140
+
46141
+
46142
+
46143
+
46144
+
46145
+
46146
+
46147
+
46148
+
46149
+
46150
+
46151
+
46152
+
46153
+
46154
+
46155
+
46156
+
46157
+
46158
+
46159
+
46160
+
46161
+
46162
+
46163
+
46164
+
46165
+
46166
+
46167
+
46168
+
46169
+
46170
+
46171
+
46172
+
46173
+
46174
+
46175
+
46176
+
46177
+
46178
+
46179
+
46180
+
46181
+
46182
+
46183
+
46184
+
46185
+
46186
+
46187
+
46188
+
46189
+
46190
+
46191
+
46192
+
46193
+
46194
+
46195
+
46196
+
46197
+
46198
+
46199
+
46200
+
46201
+
46202
+
46203
+
46204
+
46205
+
46206
+
46207
+
46208
+
46209
+
46210
+
46211
+
46212
+
46213
+
46214
+
46215
+
46216
+
46217
+
46218
+
46219
+
46220
+
46221
+
46222
+
46223
+
46224
+
46225
+
46226
+
46227
+
46228
+
46229
+
46230
+
46231
+
46232
+
46233
+
46234
+
46235
+
46236
+
46237
+
46238
+
46239
+
46240
+
46241
+
46242
+
46243
+
46244
+
46245
+
46246
+
46247
+
46248
+
46249
+
46250
+
46251
+
46252
+
46253
+
46254
+
46255
+
46256
+
46257
+
46258
+
46259
+
46260
+
46261
+
46262
+
46263
+
46264
+
46265
+
46266
+
46267
+
46268
+
46269
+
46270
+
46271
+
46272
+
46273
+
46274
+
46275
+
46276
+
46277
+
46278
+
46279
+
46280
+
46281
+
46282
+
46283
+
46284
+
46285
+
46286
+
46287
+
46288
+
46289
+
46290
+
46291
+
46292
+
46293
+
46294
+
46295
+
46296
+
46297
+
46298
+
46299
+
46300
+
46301
+
46302
+
46303
+
46304
+
46305
+
46306
+
46307
+
46308
+
46309
+
46310
+
46311
+
46312
+
46313
+
46314
+
46315
+
46316
+
46317
+
46318
+
46319
+
46320
+
46321
+
46322
+
46323
+
46324
+
46325
+
46326
+
46327
+
46328
+
46329
+
46330
+
46331
+
46332
+
46333
+
46334
+
46335
+
46336
+
46337
+
46338
+
46339
+
46340
+
46341
+
46342
+
46343
+
46344
+
46345
+
46346
+
46347
+
46348
+
46349
+
46350
+
46351
+
46352
+
46353
+
46354
+
46355
+
46356
+
46357
+
46358
+
46359
+
46360
+
46361
+
46362
+
46363
+
46364
+
46365
+
46366
+
46367
+
46368
+
46369
+
46370
+
46371
+
46372
+
46373
+
46374
+
46375
+
46376
+
46377
+
46378
+
46379
+
46380
+
46381
+
46382
+
46383
+
46384
+
46385
+
46386
+
46387
+
46388
+
46389
+
46390
+
46391
+
46392
+
46393
+
46394
+
46395
+
46396
+
46397
+
46398
+
46399
+
46400
+
46401
+
46402
+
46403
+
46404
+
46405
+
46406
+
46407
+
46408
+
46409
+
46410
+
46411
+
46412
+
46413
+
46414
+
46415
+
46416
+
46417
+
46418
+
46419
+
46420
+
46421
+
46422
+
46423
+
46424
+
46425
+
46426
+
46427
+
46428
+
46429
+
46430
+
46431
+
46432
+
46433
+
46434
+
46435
+
46436
+
46437
+
46438
+
46439
+
46440
+
46441
+
46442
+
46443
+
46444
+
46445
+
46446
+
46447
+
46448
+
46449
+
46450
+
46451
+
46452
+
46453
+
46454
+
46455
+
46456
+
46457
+
46458
+
46459
+
46460
+
46461
+
46462
+
46463
+
46464
+
46465
+
46466
+
46467
+
46468
+
46469
+
46470
+
46471
+
46472
+
46473
+
46474
+
46475
+
46476
+
46477
+
46478
+
46479
+
46480
+
46481
+
46482
+
46483
+
46484
+
46485
+
46486
+
46487
+
46488
+
46489
+
46490
+
46491
+
46492
+
46493
+
46494
+
46495
+
46496
+
46497
+
46498
+
46499
+
46500
+
46501
+
46502
+
46503
+
46504
+
46505
+
46506
+
46507
+
46508
+
46509
+
46510
+
46511
+
46512
+
46513
+
46514
+
46515
+
46516
+
46517
+
46518
+
46519
+
46520
+
46521
+
46522
+
46523
+
46524
+
46525
+
46526
+
46527
+
46528
+
46529
+
46530
+
46531
+
46532
+
46533
+
46534
+
46535
+
46536
+
46537
+
46538
+
46539
+
46540
+
46541
+
46542
+
46543
+
46544
+
46545
+
46546
+
46547
+
46548
+
46549
+
46550
+
46551
+
46552
+
46553
+
46554
+
46555
+
46556
+
46557
+
46558
+
46559
+
46560
+
46561
+
46562
+
46563
+
46564
+
46565
+
46566
+
46567
+
46568
+
46569
+
46570
+
46571
+
46572
+
46573
+
46574
+
46575
+
46576
+
46577
+
46578
+
46579
+
46580
+
46581
+
46582
+
46583
+
46584
+
46585
+
46586
+
46587
+
46588
+
46589
+
46590
+
46591
+
46592
+
46593
+
46594
+
46595
+
46596
+
46597
+
46598
+
46599
+
46600
+
46601
+
46602
+
46603
+
46604
+
46605
+
46606
+
46607
+
46608
+
46609
+
46610
+
46611
+
46612
+
46613
+
46614
+
46615
+
46616
+
46617
+
46618
+
46619
+
46620
+
46621
+
46622
+
46623
+
46624
+
46625
+
46626
+
46627
+
46628
+
46629
+
46630
+
46631
+
46632
+
46633
+
46634
+
46635
+
46636
+
46637
+
46638
+
46639
+
46640
+
46641
+
46642
+
46643
+
46644
+
46645
+
46646
+
46647
+
46648
+
46649
+
46650
+
46651
+
46652
+
46653
+
46654
+
46655
+
46656
+
46657
+
46658
+
46659
+
46660
+
46661
+
46662
+
46663
+
46664
+
46665
+
46666
+
46667
+
46668
+
46669
+
46670
+
46671
+
46672
+
46673
+
46674
+
46675
+
46676
+
46677
+
46678
+
46679
+
46680
+
46681
+
46682
+
46683
+
46684
+
46685
+
46686
+
46687
+
46688
+
46689
+
46690
+
46691
+
46692
+
46693
+
46694
+
46695
+
46696
+
46697
+
46698
+
46699
+
46700
+
46701
+
46702
+
46703
+
46704
+
46705
+
46706
+
46707
+
46708
+
46709
+
46710
+
46711
+
46712
+
46713
+
46714
+
46715
+
46716
+
46717
+
46718
+
46719
+
46720
+
46721
+
46722
+
46723
+
46724
+
46725
+
46726
+
46727
+
46728
+
46729
+
46730
+
46731
+
46732
+
46733
+
46734
+
46735
+
46736
+
46737
+
46738
+
46739
+
46740
+
46741
+
46742
+
46743
+
46744
+
46745
+
46746
+
46747
+
46748
+
46749
+
46750
+
46751
+
46752
+
46753
+
46754
+
46755
+
46756
+
46757
+
46758
+
46759
+
46760
+
46761
+
46762
+
46763
+
46764
+
46765
+
46766
+
46767
+
46768
+
46769
+
46770
+
46771
+
46772
+
46773
+
46774
+
46775
+
46776
+
46777
+
46778
+
46779
+
46780
+
46781
+
46782
+
46783
+
46784
+
46785
+
46786
+
46787
+
46788
+
46789
+
46790
+
46791
+
46792
+
46793
+
46794
+
46795
+
46796
+
46797
+
46798
+
46799
+
46800
+
46801
+
46802
+
46803
+
46804
+
46805
+
46806
+
46807
+
46808
+
46809
+
46810
+
46811
+
46812
+
46813
+
46814
+
46815
+
46816
+
46817
+
46818
+
46819
+
46820
+
46821
+
46822
+
46823
+
46824
+
46825
+
46826
+
46827
+
46828
+
46829
+
46830
+
46831
+
46832
+
46833
+
46834
+
46835
+
46836
+
46837
+
46838
+
46839
+
46840
+
46841
+
46842
+
46843
+
46844
+
46845
+
46846
+
46847
+
46848
+
46849
+
46850
+
46851
+
46852
+
46853
+
46854
+
46855
+
46856
+
46857
+
46858
+
46859
+
46860
+
46861
+
46862
+
46863
+
46864
+
46865
+
46866
+
46867
+
46868
+
46869
+
46870
+
46871
+
46872
+
46873
+
46874
+
46875
+
46876
+
46877
+
46878
+
46879
+
46880
+
46881
+
46882
+
46883
+
46884
+
46885
+
46886
+
46887
+
46888
+
46889
+
46890
+
46891
+
46892
+
46893
+
46894
+
46895
+
46896
+
46897
+
46898
+
46899
+
46900
+
46901
+
46902
+
46903
+
46904
+
46905
+
46906
+
46907
+
46908
+
46909
+
46910
+
46911
+
46912
+
46913
+
46914
+
46915
+
46916
+
46917
+
46918
+
46919
+
46920
+
46921
+
46922
+
46923
+
46924
+
46925
+
46926
+
46927
+
46928
+
46929
+
46930
+
46931
+
46932
+
46933
+
46934
+
46935
+
46936
+
46937
+
46938
+
46939
+
46940
+
46941
+
46942
+
46943
+
46944
+
46945
+
46946
+
46947
+
46948
+
46949
+
46950
+
46951
+
46952
+
46953
+
46954
+
46955
+
46956
+
46957
+
46958
+
46959
+
46960
+
46961
+
46962
+
46963
+
46964
+
46965
+
46966
+
46967
+
46968
+
46969
+
46970
+
46971
+
46972
+
46973
+
46974
+
46975
+
46976
+
46977
+
46978
+
46979
+
46980
+
46981
+
46982
+
46983
+
46984
+
46985
+
46986
+
46987
+
46988
+
46989
+
46990
+
46991
+
46992
+
46993
+
46994
+
46995
+
46996
+
46997
+
46998
+
46999
+
47000
+
47001
+
47002
+
47003
+
47004
+
47005
+
47006
+
47007
+
47008
+
47009
+
47010
+
47011
+
47012
+
47013
+
47014
+
47015
+
47016
+
47017
+
47018
+
47019
+
47020
+
47021
+
47022
+
47023
+
47024
+
47025
+
47026
+
47027
+
47028
+
47029
+
47030
+
47031
+
47032
+
47033
+
47034
+
47035
+
47036
+
47037
+
47038
+
47039
+
47040
+
47041
+
47042
+
47043
+
47044
+
47045
+
47046
+
47047
+
47048
+
47049
+
47050
+
47051
+
47052
+
47053
+
47054
+
47055
+
47056
+
47057
+
47058
+
47059
+
47060
+
47061
+
47062
+
47063
+
47064
+
47065
+
47066
+
47067
+
47068
+
47069
+
47070
+
47071
+
47072
+
47073
+
47074
+
47075
+
47076
+
47077
+
47078
+
47079
+
47080
+
47081
+
47082
+
47083
+
47084
+
47085
+
47086
+
47087
+
47088
+
47089
+
47090
+
47091
+
47092
+
47093
+
47094
+
47095
+
47096
+
47097
+
47098
+
47099
+
47100
+
47101
+
47102
+
47103
+
47104
+
47105
+
47106
+
47107
+
47108
+
47109
+
47110
+
47111
+
47112
+
47113
+
47114
+
47115
+
47116
+
47117
+
47118
+
47119
+
47120
+
47121
+
47122
+
47123
+
47124
+
47125
+
47126
+
47127
+
47128
+
47129
+
47130
+
47131
+
47132
+
47133
+
47134
+
47135
+
47136
+
47137
+
47138
+
47139
+
47140
+
47141
+
47142
+
47143
+
47144
+
47145
+
47146
+
47147
+
47148
+
47149
+
47150
+
47151
+
47152
+
47153
+
47154
+
47155
+
47156
+
47157
+
47158
+
47159
+
47160
+
47161
+
47162
+
47163
+
47164
+
47165
+
47166
+
47167
+
47168
+
47169
+
47170
+
47171
+
47172
+
47173
+
47174
+
47175
+
47176
+
47177
+
47178
+
47179
+
47180
+
47181
+
47182
+
47183
+
47184
+
47185
+
47186
+
47187
+
47188
+
47189
+
47190
+
47191
+
47192
+
47193
+
47194
+
47195
+
47196
+
47197
+
47198
+
47199
+
47200
+
47201
+
47202
+
47203
+
47204
+
47205
+
47206
+
47207
+
47208
+
47209
+
47210
+
47211
+
47212
+
47213
+
47214
+
47215
+
47216
+
47217
+
47218
+
47219
+
47220
+
47221
+
47222
+
47223
+
47224
+
47225
+
47226
+
47227
+
47228
+
47229
+
47230
+
47231
+
47232
+
47233
+
47234
+
47235
+
47236
+
47237
+
47238
+
47239
+
47240
+
47241
+
47242
+
47243
+
47244
+
47245
+
47246
+
47247
+
47248
+
47249
+
47250
+
47251
+
47252
+
47253
+
47254
+
47255
+
47256
+
47257
+
47258
+
47259
+
47260
+
47261
+
47262
+
47263
+
47264
+
47265
+
47266
+
47267
+
47268
+
47269
+
47270
+
47271
+
47272
+
47273
+
47274
+
47275
+
47276
+
47277
+
47278
+
47279
+
47280
+
47281
+
47282
+
47283
+
47284
+
47285
+
47286
+
47287
+
47288
+
47289
+
47290
+
47291
+
47292
+
47293
+
47294
+
47295
+
47296
+
47297
+
47298
+
47299
+
47300
+
47301
+
47302
+
47303
+
47304
+
47305
+
47306
+
47307
+
47308
+
47309
+
47310
+
47311
+
47312
+
47313
+
47314
+
47315
+
47316
+
47317
+
47318
+
47319
+
47320
+
47321
+
47322
+
47323
+
47324
+
47325
+
47326
+
47327
+
47328
+
47329
+
47330
+
47331
+
47332
+
47333
+
47334
+
47335
+
47336
+
47337
+
47338
+
47339
+
47340
+
47341
+
47342
+
47343
+
47344
+
47345
+
47346
+
47347
+
47348
+
47349
+
47350
+
47351
+
47352
+
47353
+
47354
+
47355
+
47356
+
47357
+
47358
+
47359
+
47360
+
47361
+
47362
+
47363
+
47364
+
47365
+
47366
+
47367
+
47368
+
47369
+
47370
+
47371
+
47372
+
47373
+
47374
+
47375
+
47376
+
47377
+
47378
+
47379
+
47380
+
47381
+
47382
+
47383
+
47384
+
47385
+ Training...: 88% 3862/4393 [5:14:58<51:10, 5.78s/it]
47386
+ Step... (41700 | Loss: 0.005744441412389278, Learning Rate: 1.676969441177789e-05, Gradient Norm: 0.1374588906764984)
47387
+ Step... (41725 | Loss: 0.008426920510828495, Learning Rate: 1.6719191989977844e-05, Gradient Norm: 0.8020673990249634)
47388
+ Step... (41750 | Loss: 0.0075722914189100266, Learning Rate: 1.6668689568177797e-05, Gradient Norm: 0.32124248147010803)
47389
+ Step... (41775 | Loss: 0.005043766926974058, Learning Rate: 1.6618179870420136e-05, Gradient Norm: 0.20893269777297974)
47390
+ Step... (41800 | Loss: 0.010060012340545654, Learning Rate: 1.656767744862009e-05, Gradient Norm: 0.21614883840084076)
47391
+ Step... (41825 | Loss: 0.014381218701601028, Learning Rate: 1.6517173207830638e-05, Gradient Norm: 0.35426953434944153)
47392
+ Step... (41850 | Loss: 0.01061459444463253, Learning Rate: 1.646666532906238e-05, Gradient Norm: 0.29452937841415405)
47393
+ Step... (41875 | Loss: 0.012180848978459835, Learning Rate: 1.641616108827293e-05, Gradient Norm: 0.3592362105846405)
47394
+ Step... (41900 | Loss: 0.00836265180259943, Learning Rate: 1.6365658666472882e-05, Gradient Norm: 0.22259344160556793)
47395
+ Step... (41925 | Loss: 0.008231023326516151, Learning Rate: 1.631514896871522e-05, Gradient Norm: 0.20443135499954224)
47396
+ Step... (41950 | Loss: 0.01289241760969162, Learning Rate: 1.6264646546915174e-05, Gradient Norm: 0.3498441278934479)
47397
+ Step... (41975 | Loss: 0.0034512323327362537, Learning Rate: 1.6214144125115126e-05, Gradient Norm: 0.13234050571918488)
47398
+ Step... (42000 | Loss: 0.011287801899015903, Learning Rate: 1.6163634427357465e-05, Gradient Norm: 0.4331746995449066)
47399
+ Step... (42025 | Loss: 0.003170725191012025, Learning Rate: 1.6113132005557418e-05, Gradient Norm: 0.13511165976524353)
47400
+ Step... (42050 | Loss: 0.011994440108537674, Learning Rate: 1.606262958375737e-05, Gradient Norm: 0.2843227684497833)
47401
+ Step... (42075 | Loss: 0.003315770300105214, Learning Rate: 1.601211988599971e-05, Gradient Norm: 0.18491697311401367)
47402
+ Step... (42100 | Loss: 0.011447119526565075, Learning Rate: 1.5961617464199662e-05, Gradient Norm: 0.23545017838478088)
47403
+ Step... (42125 | Loss: 0.005759582854807377, Learning Rate: 1.591111322341021e-05, Gradient Norm: 0.2273390293121338)
47404
+ Step... (42150 | Loss: 0.010285439901053905, Learning Rate: 1.5860605344641954e-05, Gradient Norm: 0.2514696419239044)
47405
+ Step... (42175 | Loss: 0.004782717674970627, Learning Rate: 1.5810101103852503e-05, Gradient Norm: 0.17328862845897675)
47406
+ Step... (42200 | Loss: 0.010704944841563702, Learning Rate: 1.5759598682052456e-05, Gradient Norm: 0.26759225130081177)
47407
+ Step... (42225 | Loss: 0.008894861675798893, Learning Rate: 1.5709088984294794e-05, Gradient Norm: 0.21136222779750824)
47408
+ Step... (42250 | Loss: 0.014378275722265244, Learning Rate: 1.5658586562494747e-05, Gradient Norm: 0.2818372845649719)
47409
+ Step... (42275 | Loss: 0.003270916873589158, Learning Rate: 1.56080841406947e-05, Gradient Norm: 0.1778145730495453)
47410
+ Step... (42300 | Loss: 0.01763617806136608, Learning Rate: 1.555757444293704e-05, Gradient Norm: 0.8628730773925781)
47411
+ Step... (42325 | Loss: 0.003942512441426516, Learning Rate: 1.550707202113699e-05, Gradient Norm: 0.13781097531318665)
47412
+ Step... (42350 | Loss: 0.0093576330691576, Learning Rate: 1.545656232337933e-05, Gradient Norm: 0.1558447927236557)
47413
+ Step... (42375 | Loss: 0.005177142098546028, Learning Rate: 1.5406059901579283e-05, Gradient Norm: 0.19731150567531586)
47414
+ Step... (42400 | Loss: 0.008837847970426083, Learning Rate: 1.5355557479779236e-05, Gradient Norm: 0.2248029261827469)
47415
+ Step... (42425 | Loss: 0.004063450265675783, Learning Rate: 1.5305047782021575e-05, Gradient Norm: 0.15096257627010345)
47416
+ Step... (42450 | Loss: 0.0156438909471035, Learning Rate: 1.5254544450726826e-05, Gradient Norm: 0.5371332168579102)
47417
+ Step... (42475 | Loss: 0.004050035495311022, Learning Rate: 1.5204042028926779e-05, Gradient Norm: 0.2783503830432892)
47418
+ Step... (42500 | Loss: 0.008893079124391079, Learning Rate: 1.5153533240663819e-05, Gradient Norm: 0.16905491054058075)
47419
+ Step... (42525 | Loss: 0.0038313837721943855, Learning Rate: 1.510302990936907e-05, Gradient Norm: 0.13977232575416565)
47420
+ Step... (42550 | Loss: 0.008676256984472275, Learning Rate: 1.5052526578074321e-05, Gradient Norm: 0.1817447394132614)
47421
+ Step... (42575 | Loss: 0.005087659694254398, Learning Rate: 1.5002017789811362e-05, Gradient Norm: 0.44567009806632996)
47422
+ Step... (42600 | Loss: 0.013834933750331402, Learning Rate: 1.4951514458516613e-05, Gradient Norm: 0.46296262741088867)
47423
+ Step... (42625 | Loss: 0.01469030138105154, Learning Rate: 1.4901012036716565e-05, Gradient Norm: 0.6844345331192017)
47424
+ Step... (42650 | Loss: 0.005226873327046633, Learning Rate: 1.4850503248453606e-05, Gradient Norm: 0.18194365501403809)
47425
+ Step... (42675 | Loss: 0.010348882526159286, Learning Rate: 1.4799999917158857e-05, Gradient Norm: 4.603023529052734)
47426
+ Step... (42700 | Loss: 0.00900035910308361, Learning Rate: 1.4749496585864108e-05, Gradient Norm: 0.22959791123867035)
47427
+ Step... (42725 | Loss: 0.004127497784793377, Learning Rate: 1.4698987797601148e-05, Gradient Norm: 0.16160161793231964)
47428
+ Step... (42750 | Loss: 0.009940936230123043, Learning Rate: 1.46484844663064e-05, Gradient Norm: 0.2361784130334854)
47429
+ Step... (42775 | Loss: 0.009671159088611603, Learning Rate: 1.4597982044506352e-05, Gradient Norm: 0.24877315759658813)
47430
+ Step... (42800 | Loss: 0.010518192313611507, Learning Rate: 1.4547473256243393e-05, Gradient Norm: 0.2851322591304779)
47431
+
47432
+
47433
+
47434
+
47435
+
47436
+
47437
+
47438
+
47439
+
47440
+
47441
+
47442
+
47443
+
47444
+
47445
+
47446
+
47447
+
47448
+
47449
+
47450
+
47451
+
47452
+
47453
+
47454
+
47455
+
47456
+
47457
+
47458
+
47459
+
47460
+
47461
+
47462
+
47463
+
47464
+
47465
+
47466
+
47467
+
47468
+
47469
+
47470
+
47471
+
47472
+
47473
+
47474
+
47475
+
47476
+
47477
+
47478
+
47479
+
47480
+
47481
+
47482
+
47483
+
47484
+
47485
+
47486
+
47487
+
47488
+
47489
+
47490
+
47491
+
47492
+
47493
+
47494
+
47495
+
47496
+
47497
+
47498
+
47499
+
47500
+
47501
+
47502
+
47503
+
47504
+
47505
+
47506
+
47507
+
47508
+
47509
+
47510
+
47511
+
47512
+
47513
+
47514
+
47515
+
47516
+
47517
+
47518
+
47519
+
47520
+
47521
+
47522
+
47523
+
47524
+
47525
+
47526
+
47527
+
47528
+
47529
+
47530
+
47531
+
47532
+
47533
+
47534
+
47535
+
47536
+
47537
+
47538
+
47539
+
47540
+
47541
+
47542
+
47543
+
47544
+
47545
+
47546
+
47547
+
47548
+
47549
+
47550
+
47551
+
47552
+
47553
+
47554
+
47555
+
47556
+
47557
+
47558
+
47559
+
47560
+
47561
+
47562
+
47563
+
47564
+
47565
+
47566
+
47567
+
47568
+
47569
+
47570
+
47571
+
47572
+
47573
+
47574
+
47575
+
47576
+
47577
+
47578
+
47579
+
47580
+
47581
+
47582
+
47583
+
47584
+
47585
+
47586
+
47587
+
47588
+
47589
+
47590
+
47591
+
47592
+
47593
+
47594
+
47595
+
47596
+
47597
+
47598
+
47599
+
47600
+
47601
+
47602
+
47603
+
47604
+
47605
+
47606
+
47607
+
47608
+
47609
+
47610
+
47611
+
47612
+
47613
+
47614
+
47615
+
47616
+
47617
+
47618
+
47619
+
47620
+
47621
+
47622
+
47623
+
47624
+
47625
+
47626
+
47627
+
47628
+
47629
+
47630
+
47631
+
47632
+
47633
+
47634
+
47635
+
47636
+
47637
+
47638
+
47639
+
47640
+
47641
+
47642
+
47643
+
47644
+
47645
+
47646
+
47647
+
47648
+
47649
+
47650
+
47651
+
47652
+
47653
+
47654
+
47655
+
47656
+
47657
+
47658
+
47659
+
47660
+
47661
+
47662
+
47663
+
47664
+
47665
+
47666
+
47667
+
47668
+
47669
+
47670
+
47671
+
47672
+
47673
+
47674
+
47675
+
47676
+
47677
+
47678
+
47679
+
47680
+
47681
+
47682
+
47683
+
47684
+
47685
+
47686
+
47687
+
47688
+
47689
+
47690
+
47691
+
47692
+
47693
+
47694
+
47695
+
47696
+
47697
+
47698
+
47699
+
47700
+
47701
+
47702
+
47703
+
47704
+
47705
+
47706
+
47707
+
47708
+
47709
+
47710
+
47711
+
47712
+
47713
+
47714
+
47715
+
47716
+
47717
+
47718
+
47719
+
47720
+
47721
+
47722
+
47723
+
47724
+
47725
+
47726
+
47727
+
47728
+
47729
+
47730
+
47731
+
47732
+
47733
+
47734
+
47735
+
47736
+
47737
+
47738
+
47739
+
47740
+
47741
+
47742
+
47743
+
47744
+
47745
+
47746
+
47747
+
47748
+
47749
+
47750
+
47751
+
47752
+
47753
+
47754
+
47755
+
47756
+
47757
+
47758
+
47759
+
47760
+
47761
+
47762
+
47763
+
47764
+
47765
+
47766
+
47767
+
47768
+
47769
+
47770
+
47771
+
47772
+
47773
+
47774
+
47775
+
47776
+
47777
+
47778
+
47779
+
47780
+
47781
+
47782
+
47783
+
47784
+
47785
+
47786
+
47787
+
47788
+
47789
+
47790
+
47791
+
47792
+
47793
+
47794
+
47795
+
47796
+
47797
+
47798
+
47799
+
47800
+
47801
+
47802
+
47803
+
47804
+
47805
+
47806
+
47807
+
47808
+
47809
+
47810
+
47811
+
47812
+
47813
+
47814
+
47815
+
47816
+
47817
+
47818
+
47819
+
47820
+
47821
+
47822
+
47823
+
47824
+
47825
+
47826
+
47827
+
47828
+
47829
+
47830
+
47831
+
47832
+
47833
+
47834
+
47835
+
47836
+
47837
+
47838
+
47839
+
47840
+
47841
+
47842
+
47843
+
47844
+
47845
+
47846
+
47847
+
47848
+
47849
+
47850
+
47851
+
47852
+
47853
+
47854
+
47855
+
47856
+
47857
+
47858
+
47859
+
47860
+
47861
+
47862
+
47863
+
47864
+
47865
+
47866
+
47867
+
47868
+
47869
+
47870
+
47871
+
47872
+
47873
+
47874
+
47875
+
47876
+
47877
+
47878
+
47879
+
47880
+
47881
+
47882
+
47883
+
47884
+
47885
+
47886
+
47887
+
47888
+
47889
+
47890
+
47891
+
47892
+
47893
+
47894
+
47895
+
47896
+
47897
+
47898
+
47899
+
47900
+
47901
+
47902
+
47903
+
47904
+
47905
+
47906
+
47907
+
47908
+
47909
+
47910
+
47911
+
47912
+
47913
+
47914
+
47915
+
47916
+
47917
+
47918
+
47919
+
47920
+
47921
+
47922
+
47923
+
47924
+
47925
+
47926
+
47927
+
47928
+
47929
+
47930
+
47931
+
47932
+
47933
+
47934
+
47935
+
47936
+
47937
+
47938
+
47939
+
47940
+
47941
+
47942
+
47943
+
47944
+
47945
+
47946
+
47947
+
47948
+
47949
+ Step... (40000/50000 | Eval Loss: 0.9527401328086853 | Eval wer: 0.04187346053453917 | Eval cer: 0.02872188479352137 |): 83% 10/12 [58:33:51<11:44:51, 21145.58s/it]
47950
+ Training...: 0% 0/4393 [00:00<?, ?it/s]
47951
+ Step... (42875 | Loss: 0.013169726356863976, Learning Rate: 1.4395957805390935e-05, Gradient Norm: 0.2936961054801941)
47952
+ Step... (42900 | Loss: 0.007032770197838545, Learning Rate: 1.4345454474096186e-05, Gradient Norm: 0.20587222278118134)
47953
+ Step... (42925 | Loss: 0.004653589334338903, Learning Rate: 1.429495205229614e-05, Gradient Norm: 0.18729034066200256)
47954
+ Step... (42950 | Loss: 0.005974963773041964, Learning Rate: 1.424444326403318e-05, Gradient Norm: 0.1742350161075592)
47955
+ Step... (42975 | Loss: 0.0030157752335071564, Learning Rate: 1.419393993273843e-05, Gradient Norm: 0.09114522486925125)
47956
+ Step... (43000 | Loss: 0.017072968184947968, Learning Rate: 1.4143436601443682e-05, Gradient Norm: 0.344232439994812)
47957
+ Step... (43025 | Loss: 0.007778532337397337, Learning Rate: 1.4092927813180722e-05, Gradient Norm: 0.24894601106643677)
47958
+ Step... (43050 | Loss: 0.005137579515576363, Learning Rate: 1.4042424481885973e-05, Gradient Norm: 0.15516820549964905)
47959
+ Step... (43075 | Loss: 0.008852087892591953, Learning Rate: 1.3991922060085926e-05, Gradient Norm: 0.2543483078479767)
47960
+ Step... (43100 | Loss: 0.005538458935916424, Learning Rate: 1.3941412362328265e-05, Gradient Norm: 0.20741888880729675)
47961
+ Step... (43125 | Loss: 0.009585311636328697, Learning Rate: 1.3890909940528218e-05, Gradient Norm: 0.32343360781669617)
47962
+ Step... (43150 | Loss: 0.009101358242332935, Learning Rate: 1.3840406609233469e-05, Gradient Norm: 0.1852715015411377)
47963
+ Step... (43175 | Loss: 0.0038933581672608852, Learning Rate: 1.378989782097051e-05, Gradient Norm: 0.200323686003685)
47964
+ Step... (43200 | Loss: 0.006398520432412624, Learning Rate: 1.373939448967576e-05, Gradient Norm: 0.3812198340892792)
47965
+ Step... (43225 | Loss: 0.00649980129674077, Learning Rate: 1.3688892067875713e-05, Gradient Norm: 0.2510472238063812)
47966
+ Step... (43250 | Loss: 0.007147450000047684, Learning Rate: 1.3638382370118052e-05, Gradient Norm: 0.2402760088443756)
47967
+ Step... (43275 | Loss: 0.005947265308350325, Learning Rate: 1.3587879948318005e-05, Gradient Norm: 0.21199539303779602)
47968
+ Step... (43300 | Loss: 0.01116273459047079, Learning Rate: 1.3537376617023256e-05, Gradient Norm: 0.1889847368001938)
47969
+ Step... (43325 | Loss: 0.009462771005928516, Learning Rate: 1.3486867828760296e-05, Gradient Norm: 0.4412449598312378)
47970
+ Step... (43350 | Loss: 0.01658274605870247, Learning Rate: 1.3436364497465547e-05, Gradient Norm: 0.24183489382266998)
47971
+ Step... (43375 | Loss: 0.005811340641230345, Learning Rate: 1.3385855709202588e-05, Gradient Norm: 0.13360491394996643)
47972
+ Step... (43400 | Loss: 0.006220472510904074, Learning Rate: 1.3335352377907839e-05, Gradient Norm: 0.15546762943267822)
47973
+ Step... (43425 | Loss: 0.006581796798855066, Learning Rate: 1.3284849956107792e-05, Gradient Norm: 0.34990987181663513)
47974
+ Step... (43450 | Loss: 0.011822747066617012, Learning Rate: 1.3234341167844832e-05, Gradient Norm: 0.2574056386947632)
47975
+ Step... (43475 | Loss: 0.00934094749391079, Learning Rate: 1.3183837836550083e-05, Gradient Norm: 0.3183192014694214)
47976
+ Step... (43500 | Loss: 0.005102855619043112, Learning Rate: 1.3133334505255334e-05, Gradient Norm: 0.14597401022911072)
47977
+ Step... (43525 | Loss: 0.008145734667778015, Learning Rate: 1.3082825716992375e-05, Gradient Norm: 0.19706687331199646)
47978
+ Step... (43550 | Loss: 0.009935390204191208, Learning Rate: 1.3032322385697626e-05, Gradient Norm: 0.26017606258392334)
47979
+ Step... (43575 | Loss: 0.007766792085021734, Learning Rate: 1.2981819963897578e-05, Gradient Norm: 0.2769607603549957)
47980
+ Step... (43600 | Loss: 0.010773919522762299, Learning Rate: 1.2931311175634619e-05, Gradient Norm: 0.23994655907154083)
47981
+ Step... (43625 | Loss: 0.0029236481059342623, Learning Rate: 1.288080784433987e-05, Gradient Norm: 0.1163998618721962)
47982
+ Step... (43650 | Loss: 0.013007597997784615, Learning Rate: 1.2830304513045121e-05, Gradient Norm: 0.2571600377559662)
47983
+ Step... (43675 | Loss: 0.006016268394887447, Learning Rate: 1.2779795724782161e-05, Gradient Norm: 0.2098936140537262)
47984
+ Step... (43700 | Loss: 0.00377640244551003, Learning Rate: 1.2729292393487412e-05, Gradient Norm: 0.0722104161977768)
47985
+ Step... (43725 | Loss: 0.005282220430672169, Learning Rate: 1.2678789971687365e-05, Gradient Norm: 0.19065354764461517)
47986
+ Step... (43750 | Loss: 0.005376732908189297, Learning Rate: 1.2628281183424406e-05, Gradient Norm: 0.16388669610023499)
47987
+ Step... (43775 | Loss: 0.008202340453863144, Learning Rate: 1.2577777852129657e-05, Gradient Norm: 0.1876869797706604)
47988
+ Step... (43800 | Loss: 0.004423276986926794, Learning Rate: 1.2527274520834908e-05, Gradient Norm: 0.14415715634822845)
47989
+ Step... (43825 | Loss: 0.0041627101600170135, Learning Rate: 1.2476765732571948e-05, Gradient Norm: 0.18802481889724731)
47990
+ Step... (43850 | Loss: 0.012115256860852242, Learning Rate: 1.24262624012772e-05, Gradient Norm: 0.35901159048080444)
47991
+ Step... (43875 | Loss: 0.005729792173951864, Learning Rate: 1.2375759979477152e-05, Gradient Norm: 0.3747013211250305)
47992
+ Step... (43900 | Loss: 0.005576059687882662, Learning Rate: 1.2325251191214193e-05, Gradient Norm: 0.14393489062786102)
47993
+
47994
+
47995
+
47996
+
47997
+
47998
+
47999
+
48000
+
48001
+
48002
+
48003
+
48004
+
48005
+
48006
+
48007
+
48008
+
48009
+
48010
+
48011
+
48012
+
48013
+
48014
+
48015
+
48016
+
48017
+
48018
+
48019
+
48020
+
48021
+
48022
+
48023
+
48024
+
48025
+
48026
+
48027
+
48028
+
48029
+
48030
+
48031
+
48032
+
48033
+
48034
+
48035
+
48036
+
48037
+
48038
+
48039
+
48040
+
48041
+
48042
+
48043
+
48044
+
48045
+
48046
+
48047
+
48048
+
48049
+
48050
+
48051
+
48052
+
48053
+
48054
+
48055
+
48056
+
48057
+
48058
+
48059
+
48060
+
48061
+
48062
+
48063
+
48064
+
48065
+
48066
+
48067
+
48068
+
48069
+
48070
+
48071
+
48072
+
48073
+
48074
+
48075
+
48076
+
48077
+
48078
+
48079
+
48080
+
48081
+
48082
+
48083
+
48084
+
48085
+
48086
+
48087
+
48088
+
48089
+
48090
+
48091
+
48092
+
48093
+
48094
+
48095
+
48096
+
48097
+
48098
+
48099
+
48100
+
48101
+
48102
+
48103
+
48104
+
48105
+
48106
+
48107
+
48108
+
48109
+
48110
+
48111
+
48112
+
48113
+
48114
+
48115
+
48116
+
48117
+
48118
+
48119
+
48120
+
48121
+
48122
+
48123
+
48124
+
48125
+
48126
+
48127
+
48128
+
48129
+
48130
+
48131
+
48132
+
48133
+
48134
+
48135
+
48136
+
48137
+
48138
+
48139
+
48140
+
48141
+
48142
+
48143
+
48144
+
48145
+
48146
+
48147
+
48148
+
48149
+
48150
+
48151
+
48152
+
48153
+
48154
+
48155
+
48156
+
48157
+
48158
+
48159
+
48160
+
48161
+
48162
+
48163
+
48164
+
48165
+
48166
+
48167
+
48168
+
48169
+
48170
+
48171
+
48172
+
48173
+
48174
+
48175
+
48176
+
48177
+
48178
+
48179
+
48180
+
48181
+
48182
+
48183
+
48184
+
48185
+
48186
+
48187
+
48188
+
48189
+
48190
+
48191
+
48192
+
48193
+
48194
+
48195
+
48196
+
48197
+
48198
+
48199
+
48200
+
48201
+
48202
+
48203
+
48204
+
48205
+
48206
+
48207
+
48208
+
48209
+
48210
+
48211
+
48212
+
48213
+
48214
+
48215
+
48216
+
48217
+
48218
+
48219
+
48220
+
48221
+
48222
+
48223
+
48224
+
48225
+
48226
+
48227
+
48228
+
48229
+
48230
+
48231
+
48232
+
48233
+
48234
+
48235
+
48236
+
48237
+
48238
+
48239
+
48240
+
48241
+
48242
+
48243
+
48244
+
48245
+
48246
+
48247
+
48248
+
48249
+
48250
+
48251
+
48252
+
48253
+
48254
+
48255
+
48256
+
48257
+
48258
+
48259
+
48260
+
48261
+
48262
+
48263
+
48264
+
48265
+
48266
+
48267
+
48268
+
48269
+
48270
+
48271
+
48272
+
48273
+
48274
+
48275
+
48276
+
48277
+
48278
+
48279
+
48280
+
48281
+
48282
+
48283
+
48284
+
48285
+
48286
+
48287
+
48288
+
48289
+
48290
+
48291
+
48292
+
48293
+
48294
+
48295
+
48296
+
48297
+
48298
+
48299
+
48300
+
48301
+
48302
+
48303
+
48304
+
48305
+
48306
+
48307
+
48308
+
48309
+
48310
+
48311
+
48312
+
48313
+
48314
+
48315
+
48316
+
48317
+
48318
+
48319
+
48320
+
48321
+
48322
+
48323
+
48324
+
48325
+
48326
+
48327
+
48328
+
48329
+
48330
+
48331
+
48332
+
48333
+
48334
+
48335
+
48336
+
48337
+
48338
+
48339
+
48340
+
48341
+
48342
+
48343
+
48344
+
48345
+
48346
+
48347
+
48348
+
48349
+
48350
+
48351
+
48352
+
48353
+
48354
+
48355
+
48356
+
48357
+
48358
+
48359
+
48360
+
48361
+
48362
+
48363
+
48364
+
48365
+
48366
+
48367
+
48368
+
48369
+
48370
+
48371
+
48372
+
48373
+
48374
+
48375
+
48376
+
48377
+
48378
+
48379
+
48380
+
48381
+
48382
+
48383
+
48384
+
48385
+
48386
+
48387
+
48388
+
48389
+
48390
+
48391
+
48392
+
48393
+
48394
+
48395
+
48396
+
48397
+
48398
+
48399
+
48400
+
48401
+
48402
+
48403
+
48404
+
48405
+
48406
+
48407
+
48408
+
48409
+
48410
+
48411
+
48412
+
48413
+
48414
+
48415
+
48416
+
48417
+
48418
+
48419
+
48420
+
48421
+
48422
+
48423
+
48424
+
48425
+
48426
+
48427
+
48428
+
48429
+
48430
+
48431
+
48432
+
48433
+
48434
+
48435
+
48436
+
48437
+
48438
+
48439
+
48440
+
48441
+
48442
+
48443
+
48444
+
48445
+
48446
+
48447
+
48448
+
48449
+
48450
+
48451
+
48452
+
48453
+
48454
+
48455
+
48456
+
48457
+
48458
+
48459
+
48460
+
48461
+
48462
+
48463
+
48464
+
48465
+
48466
+
48467
+
48468
+
48469
+
48470
+
48471
+
48472
+
48473
+
48474
+
48475
+
48476
+
48477
+
48478
+
48479
+
48480
+
48481
+
48482
+
48483
+
48484
+
48485
+
48486
+
48487
+
48488
+
48489
+
48490
+
48491
+
48492
+
48493
+
48494
+
48495
+
48496
+
48497
+
48498
+
48499
+
48500
+
48501
+
48502
+
48503
+
48504
+
48505
+
48506
+
48507
+
48508
+
48509
+
48510
+
48511
+
48512
+
48513
+
48514
+
48515
+
48516
+
48517
+
48518
+
48519
+
48520
+
48521
+
48522
+
48523
+
48524
+
48525
+
48526
+
48527
+
48528
+
48529
+
48530
+
48531
+
48532
+
48533
+
48534
+
48535
+
48536
+
48537
+
48538
+
48539
+
48540
+
48541
+
48542
+
48543
+
48544
+
48545
+
48546
+
48547
+
48548
+
48549
+
48550
+
48551
+
48552
+
48553
+
48554
+
48555
+
48556
+
48557
+
48558
+
48559
+
48560
+
48561
+
48562
+
48563
+
48564
+
48565
+
48566
+
48567
+
48568
+
48569
+
48570
+
48571
+
48572
+
48573
+
48574
+
48575
+
48576
+
48577
+
48578
+
48579
+
48580
+
48581
+
48582
+
48583
+
48584
+
48585
+
48586
+
48587
+
48588
+
48589
+
48590
+
48591
+
48592
+
48593
+
48594
+
48595
+
48596
+
48597
+
48598
+
48599
+
48600
+
48601
+
48602
+
48603
+
48604
+
48605
+
48606
+
48607
+
48608
+
48609
+
48610
+
48611
+
48612
+
48613
+
48614
+
48615
+
48616
+
48617
+
48618
+
48619
+
48620
+
48621
+
48622
+
48623
+
48624
+
48625
+
48626
+
48627
+
48628
+
48629
+
48630
+
48631
+
48632
+
48633
+
48634
+
48635
+
48636
+
48637
+
48638
+
48639
+
48640
+
48641
+
48642
+
48643
+
48644
+
48645
+
48646
+
48647
+
48648
+
48649
+
48650
+
48651
+
48652
+
48653
+
48654
+
48655
+
48656
+
48657
+
48658
+
48659
+
48660
+
48661
+
48662
+
48663
+
48664
+
48665
+
48666
+
48667
+
48668
+
48669
+
48670
+
48671
+
48672
+
48673
+
48674
+
48675
+
48676
+
48677
+
48678
+
48679
+
48680
+
48681
+
48682
+
48683
+
48684
+
48685
+
48686
+
48687
+
48688
+
48689
+
48690
+
48691
+
48692
+
48693
+
48694
+
48695
+
48696
+
48697
+
48698
+
48699
+
48700
+
48701
+
48702
+
48703
+
48704
+
48705
+
48706
+
48707
+
48708
+
48709
+
48710
+
48711
+
48712
+
48713
+
48714
+
48715
+
48716
+
48717
+
48718
+
48719
+
48720
+
48721
+
48722
+
48723
+
48724
+
48725
+
48726
+
48727
+
48728
+
48729
+
48730
+
48731
+
48732
+
48733
+
48734
+
48735
+
48736
+
48737
+
48738
+
48739
+
48740
+
48741
+
48742
+
48743
+
48744
+
48745
+
48746
+
48747
+
48748
+
48749
+
48750
+
48751
+
48752
+
48753
+
48754
+
48755
+
48756
+
48757
+
48758
+
48759
+
48760
+
48761
+
48762
+
48763
+
48764
+
48765
+
48766
+
48767
+
48768
+
48769
+
48770
+
48771
+
48772
+
48773
+
48774
+
48775
+
48776
+
48777
+
48778
+
48779
+
48780
+
48781
+
48782
+
48783
+
48784
+
48785
+
48786
+
48787
+
48788
+
48789
+
48790
+
48791
+
48792
+
48793
+
48794
+
48795
+
48796
+
48797
+
48798
+
48799
+
48800
+
48801
+
48802
+
48803
+
48804
+
48805
+
48806
+
48807
+
48808
+
48809
+
48810
+
48811
+
48812
+
48813
+
48814
+
48815
+
48816
+
48817
+
48818
+
48819
+
48820
+
48821
+
48822
+
48823
+
48824
+
48825
+
48826
+
48827
+
48828
+
48829
+
48830
+
48831
+
48832
+
48833
+
48834
+
48835
+
48836
+
48837
+
48838
+
48839
+
48840
+
48841
+
48842
+
48843
+
48844
+
48845
+
48846
+
48847
+
48848
+
48849
+
48850
+
48851
+
48852
+
48853
+
48854
+
48855
+
48856
+
48857
+
48858
+
48859
+
48860
+
48861
+
48862
+
48863
+
48864
+
48865
+
48866
+
48867
+
48868
+
48869
+
48870
+
48871
+
48872
+
48873
+
48874
+
48875
+
48876
+
48877
+
48878
+
48879
+
48880
+
48881
+
48882
+
48883
+
48884
+
48885
+
48886
+
48887
+
48888
+
48889
+
48890
+
48891
+
48892
+
48893
+
48894
+
48895
+
48896
+
48897
+
48898
+
48899
+
48900
+
48901
+
48902
+
48903
+
48904
+
48905
+
48906
+
48907
+
48908
+
48909
+
48910
+
48911
+
48912
+
48913
+
48914
+
48915
+
48916
+
48917
+
48918
+
48919
+
48920
+
48921
+
48922
+
48923
+
48924
+
48925
+
48926
+
48927
+
48928
+
48929
+
48930
+
48931
+
48932
+
48933
+
48934
+
48935
+
48936
+
48937
+
48938
+
48939
+
48940
+
48941
+
48942
+
48943
+
48944
+
48945
+
48946
+
48947
+
48948
+
48949
+
48950
+
48951
+
48952
+
48953
+
48954
+
48955
+
48956
+
48957
+
48958
+
48959
+
48960
+
48961
+
48962
+
48963
+
48964
+
48965
+
48966
+
48967
+
48968
+
48969
+
48970
+
48971
+
48972
+
48973
+
48974
+
48975
+
48976
+
48977
+
48978
+
48979
+
48980
+
48981
+
48982
+
48983
+
48984
+
48985
+
48986
+
48987
+
48988
+
48989
+
48990
+
48991
+
48992
+
48993
+
48994
+
48995
+
48996
+
48997
+
48998
+
48999
+
49000
+
49001
+
49002
+
49003
+
49004
+
49005
+
49006
+
49007
+
49008
+
49009
+
49010
+
49011
+
49012
+
49013
+
49014
+
49015
+
49016
+
49017
+
49018
+
49019
+
49020
+
49021
+
49022
+
49023
+
49024
+
49025
+
49026
+
49027
+
49028
+
49029
+
49030
+
49031
+
49032
+
49033
+
49034
+
49035
+
49036
+
49037
+
49038
+
49039
+
49040
+
49041
+
49042
+
49043
+
49044
+
49045
+
49046
+
49047
+
49048
+
49049
+
49050
+
49051
+
49052
+
49053
+
49054
+
49055
+
49056
+
49057
+
49058
+
49059
+
49060
+
49061
+
49062
+
49063
+
49064
+
49065
+
49066
+
49067
+
49068
+
49069
+
49070
+
49071
+
49072
+
49073
+
49074
+
49075
+
49076
+
49077
+
49078
+
49079
+
49080
+
49081
+
49082
+
49083
+
49084
+
49085
+
49086
+
49087
+
49088
+
49089
+
49090
+
49091
+
49092
+
49093
+
49094
+
49095
+
49096
+
49097
+
49098
+
49099
+
49100
+
49101
+
49102
+
49103
+
49104
+
49105
+
49106
+
49107
+
49108
+
49109
+
49110
+
49111
+
49112
+
49113
+
49114
+
49115
+
49116
+
49117
+
49118
+
49119
+
49120
+
49121
+
49122
+
49123
+
49124
+
49125
+
49126
+
49127
+
49128
+
49129
+
49130
+
49131
+
49132
+
49133
+
49134
+
49135
+
49136
+
49137
+
49138
+
49139
+
49140
+
49141
+
49142
+
49143
+
49144
+
49145
+
49146
+
49147
+
49148
+
49149
+
49150
+
49151
+
49152
+
49153
+
49154
+
49155
+
49156
+
49157
+
49158
+
49159
+
49160
+
49161
+
49162
+
49163
+
49164
+
49165
+
49166
+
49167
+
49168
+
49169
+
49170
+
49171
+
49172
+
49173
+
49174
+
49175
+
49176
+
49177
+
49178
+
49179
+
49180
+
49181
+
49182
+
49183
+
49184
+
49185
+
49186
+
49187
+
49188
+
49189
+
49190
+
49191
+
49192
+
49193
+
49194
+
49195
+
49196
+
49197
+
49198
+
49199
+
49200
+
49201
+
49202
+
49203
+
49204
+
49205
+
49206
+
49207
+
49208
+
49209
+
49210
+
49211
+
49212
+
49213
+
49214
+
49215
+
49216
+
49217
+
49218
+
49219
+
49220
+
49221
+
49222
+
49223
+
49224
+
49225
+
49226
+
49227
+
49228
+
49229
+
49230
+
49231
+
49232
+
49233
+
49234
+
49235
+
49236
+
49237
+
49238
+
49239
+
49240
+
49241
+
49242
+
49243
+
49244
+
49245
+
49246
+
49247
+
49248
+
49249
+
49250
+
49251
+
49252
+
49253
+
49254
+
49255
+
49256
+
49257
+
49258
+
49259
+
49260
+
49261
+
49262
+
49263
+
49264
+
49265
+
49266
+
49267
+
49268
+
49269
+
49270
+
49271
+
49272
+
49273
+
49274
+
49275
+
49276
+
49277
+
49278
+
49279
+
49280
+
49281
+
49282
+
49283
+
49284
+
49285
+
49286
+
49287
+
49288
+
49289
+
49290
+
49291
+
49292
+
49293
+
49294
+
49295
+
49296
+
49297
+
49298
+
49299
+
49300
+
49301
+
49302
+
49303
+
49304
+
49305
+
49306
+
49307
+
49308
+
49309
+
49310
+
49311
+
49312
+
49313
+
49314
+
49315
+
49316
+
49317
+
49318
+
49319
+
49320
+
49321
+
49322
+
49323
+
49324
+
49325
+
49326
+
49327
+
49328
+
49329
+
49330
+
49331
+
49332
+
49333
+
49334
+
49335
+
49336
+
49337
+
49338
+
49339
+
49340
+
49341
+
49342
+
49343
+
49344
+
49345
+
49346
+
49347
+
49348
+
49349
+
49350
+
49351
+
49352
+
49353
+
49354
+
49355
+
49356
+
49357
+
49358
+
49359
+
49360
+
49361
+
49362
+
49363
+
49364
+
49365
+
49366
+
49367
+
49368
+
49369
+
49370
+
49371
+
49372
+
49373
+
49374
+
49375
+
49376
+
49377
+
49378
+
49379
+
49380
+
49381
+
49382
+
49383
+
49384
+
49385
+
49386
+
49387
+
49388
+
49389
+
49390
+
49391
+
49392
+
49393
+
49394
+
49395
+
49396
+
49397
+
49398
+
49399
+
49400
+
49401
+
49402
+
49403
+
49404
+
49405
+
49406
+
49407
+
49408
+
49409
+
49410
+
49411
+
49412
+
49413
+
49414
+
49415
+
49416
+
49417
+
49418
+
49419
+
49420
+
49421
+
49422
+
49423
+
49424
+
49425
+
49426
+
49427
+
49428
+
49429
+
49430
+
49431
+
49432
+
49433
+
49434
+
49435
+
49436
+
49437
+
49438
+
49439
+
49440
+
49441
+
49442
+
49443
+
49444
+
49445
+
49446
+
49447
+
49448
+
49449
+
49450
+
49451
+
49452
+
49453
+
49454
+
49455
+
49456
+
49457
+
49458
+
49459
+
49460
+
49461
+
49462
+
49463
+
49464
+
49465
+
49466
+
49467
+
49468
+
49469
+
49470
+
49471
+
49472
+
49473
+
49474
+
49475
+
49476
+
49477
+
49478
+
49479
+
49480
+
49481
+
49482
+
49483
+
49484
+
49485
+
49486
+
49487
+
49488
+
49489
+
49490
+
49491
+
49492
+
49493
+
49494
+
49495
+
49496
+
49497
+
49498
+
49499
+
49500
+
49501
+
49502
+
49503
+
49504
+
49505
+
49506
+
49507
+
49508
+
49509
+
49510
+
49511
+
49512
+
49513
+
49514
+
49515
+
49516
+
49517
+
49518
+
49519
+
49520
+
49521
+
49522
+
49523
+
49524
+
49525
+
49526
+
49527
+
49528
+
49529
+
49530
+
49531
+
49532
+
49533
+
49534
+
49535
+
49536
+
49537
+
49538
+
49539
+
49540
+
49541
+
49542
+
49543
+
49544
+
49545
+
49546
+
49547
+
49548
+
49549
+
49550
+
49551
+
49552
+
49553
+
49554
+
49555
+
49556
+
49557
+
49558
+
49559
+
49560
+
49561
+
49562
+
49563
+
49564
+
49565
+
49566
+
49567
+
49568
+
49569
+
49570
+
49571
+
49572
+
49573
+
49574
+
49575
+
49576
+
49577
+
49578
+
49579
+
49580
+
49581
+
49582
+
49583
+
49584
+
49585
+
49586
+
49587
+
49588
+
49589
+
49590
+
49591
+
49592
+
49593
+
49594
+
49595
+
49596
+
49597
+
49598
+
49599
+
49600
+
49601
+
49602
+
49603
+
49604
+
49605
+
49606
+
49607
+
49608
+
49609
+
49610
+
49611
+
49612
+
49613
+
49614
+
49615
+
49616
+
49617
+
49618
+
49619
+
49620
+
49621
+
49622
+
49623
+
49624
+
49625
+
49626
+
49627
+
49628
+
49629
+
49630
+
49631
+
49632
+
49633
+
49634
+
49635
+
49636
+
49637
+
49638
+
49639
+
49640
+
49641
+
49642
+
49643
+
49644
+
49645
+
49646
+
49647
+
49648
+
49649
+
49650
+
49651
+
49652
+
49653
+
49654
+
49655
+
49656
+
49657
+
49658
+
49659
+
49660
+
49661
+
49662
+
49663
+
49664
+
49665
+
49666
+
49667
+
49668
+
49669
+
49670
+
49671
+
49672
+
49673
+ Training...: 39% 1719/4393 [2:16:28<4:08:58, 5.59s/it]
49674
+ Step... (43950 | Loss: 0.001941631082445383, Learning Rate: 1.2224244528624695e-05, Gradient Norm: 0.07129886001348495)
49675
+ Step... (43975 | Loss: 0.0024623877834528685, Learning Rate: 1.2173735740361735e-05, Gradient Norm: 0.16456308960914612)
49676
+ Step... (44000 | Loss: 0.005050975829362869, Learning Rate: 1.2123232409066986e-05, Gradient Norm: 0.16223661601543427)
49677
+ Step... (44025 | Loss: 0.003012409433722496, Learning Rate: 1.2072729987266939e-05, Gradient Norm: 0.2435944676399231)
49678
+ Step... (44050 | Loss: 0.002762197982519865, Learning Rate: 1.202222119900398e-05, Gradient Norm: 0.0841917023062706)
49679
+ Step... (44075 | Loss: 0.009030467830598354, Learning Rate: 1.197171786770923e-05, Gradient Norm: 0.832758903503418)
49680
+ Step... (44100 | Loss: 0.005016005132347345, Learning Rate: 1.1921214536414482e-05, Gradient Norm: 0.1769050508737564)
49681
+ Step... (44125 | Loss: 0.0022536602336913347, Learning Rate: 1.1870705748151522e-05, Gradient Norm: 0.17979957163333893)
49682
+ Step... (44150 | Loss: 0.004251351114362478, Learning Rate: 1.1820202416856773e-05, Gradient Norm: 0.2023152858018875)
49683
+ Step... (44175 | Loss: 0.001228940673172474, Learning Rate: 1.1769699995056726e-05, Gradient Norm: 0.08239232003688812)
49684
+ Step... (44200 | Loss: 0.0023271413519978523, Learning Rate: 1.1719191206793766e-05, Gradient Norm: 0.13417169451713562)
49685
+ Step... (44225 | Loss: 0.0019650713074952364, Learning Rate: 1.1668687875499018e-05, Gradient Norm: 0.14780059456825256)
49686
+ Step... (44250 | Loss: 0.00524902855977416, Learning Rate: 1.1618184544204269e-05, Gradient Norm: 0.23804089426994324)
49687
+ Step... (44275 | Loss: 0.0012860909337177873, Learning Rate: 1.1567675755941309e-05, Gradient Norm: 0.12285351753234863)
49688
+ Step... (44300 | Loss: 0.007503165397793055, Learning Rate: 1.151717242464656e-05, Gradient Norm: 0.2330530434846878)
49689
+ Step... (44325 | Loss: 0.0006487671053037047, Learning Rate: 1.1466670002846513e-05, Gradient Norm: 0.03001929260790348)
49690
+ Step... (44350 | Loss: 0.004298559855669737, Learning Rate: 1.1416161214583553e-05, Gradient Norm: 0.13478074967861176)
49691
+ Step... (44375 | Loss: 0.002003872999921441, Learning Rate: 1.1365657883288804e-05, Gradient Norm: 0.19664216041564941)
49692
+ Step... (44400 | Loss: 0.0063357907347381115, Learning Rate: 1.1315149095025845e-05, Gradient Norm: 0.33638206124305725)
49693
+ Step... (44425 | Loss: 0.0043112244457006454, Learning Rate: 1.1264645763731096e-05, Gradient Norm: 0.4819294214248657)
49694
+ Step... (44450 | Loss: 0.005926346872001886, Learning Rate: 1.1214142432436347e-05, Gradient Norm: 0.20366427302360535)
49695
+ Step... (44475 | Loss: 0.0005286787636578083, Learning Rate: 1.1163633644173387e-05, Gradient Norm: 0.023813458159565926)
49696
+ Step... (44500 | Loss: 0.002619663020595908, Learning Rate: 1.1113130312878639e-05, Gradient Norm: 0.14585494995117188)
49697
+ Step... (44525 | Loss: 0.0008802820811979473, Learning Rate: 1.1062627891078591e-05, Gradient Norm: 0.05259334295988083)
49698
+ Step... (44550 | Loss: 0.007913529872894287, Learning Rate: 1.1012119102815632e-05, Gradient Norm: 0.2658507525920868)
49699
+ Step... (44575 | Loss: 0.0005179516738280654, Learning Rate: 1.0961615771520883e-05, Gradient Norm: 0.0357116162776947)
49700
+ Step... (44600 | Loss: 0.006706691812723875, Learning Rate: 1.0911112440226134e-05, Gradient Norm: 0.19824419915676117)
49701
+ Step... (44625 | Loss: 0.000933133764192462, Learning Rate: 1.0860603651963174e-05, Gradient Norm: 0.06666836887598038)
49702
+ Step... (44650 | Loss: 0.0057255276478827, Learning Rate: 1.0810100320668425e-05, Gradient Norm: 0.14937716722488403)
49703
+ Step... (44675 | Loss: 0.0018604991491883993, Learning Rate: 1.0759597898868378e-05, Gradient Norm: 0.1628139764070511)
49704
+ Step... (44700 | Loss: 0.00716938404366374, Learning Rate: 1.0709089110605419e-05, Gradient Norm: 0.1885882467031479)
49705
+ Step... (44725 | Loss: 0.001772589748725295, Learning Rate: 1.065858577931067e-05, Gradient Norm: 0.1799585074186325)
49706
+ Step... (44750 | Loss: 0.0021098533179610968, Learning Rate: 1.060808244801592e-05, Gradient Norm: 0.07331665605306625)
49707
+ Step... (44775 | Loss: 0.0028699678368866444, Learning Rate: 1.0557573659752961e-05, Gradient Norm: 0.3392123579978943)
49708
+ Step... (44800 | Loss: 0.008694657124578953, Learning Rate: 1.0507070328458212e-05, Gradient Norm: 0.3680751323699951)
49709
+ Step... (44825 | Loss: 0.0008265204378403723, Learning Rate: 1.0456567906658165e-05, Gradient Norm: 0.055144477635622025)
49710
+ Step... (44850 | Loss: 0.002472145017236471, Learning Rate: 1.0406059118395206e-05, Gradient Norm: 0.12441207468509674)
49711
+ Step... (44875 | Loss: 0.0018344148993492126, Learning Rate: 1.0355555787100457e-05, Gradient Norm: 0.10230448096990585)
49712
+ Step... (44900 | Loss: 0.005570901557803154, Learning Rate: 1.0305052455805708e-05, Gradient Norm: 0.11906146258115768)
49713
+ Step... (44925 | Loss: 0.000813481688965112, Learning Rate: 1.0254543667542748e-05, Gradient Norm: 0.06892543286085129)
49714
+ Step... (44950 | Loss: 0.004630820360034704, Learning Rate: 1.0204040336248e-05, Gradient Norm: 0.18374110758304596)
49715
+ Step... (44975 | Loss: 0.000776787637732923, Learning Rate: 1.0153537914447952e-05, Gradient Norm: 0.04277883097529411)
49716
+ Step... (45000 | Loss: 0.00541204446926713, Learning Rate: 1.0103029126184992e-05, Gradient Norm: 0.1890871673822403)
49717
+ Step... (45025 | Loss: 0.002748969243839383, Learning Rate: 1.0052525794890244e-05, Gradient Norm: 0.21585719287395477)
49718
+ Step... (45050 | Loss: 0.004655120428651571, Learning Rate: 1.0002022463595495e-05, Gradient Norm: 0.15042269229888916)
49719
+ Step... (45075 | Loss: 0.001320728100836277, Learning Rate: 9.951513675332535e-06, Gradient Norm: 0.12991033494472504)
49720
+ Step... (45100 | Loss: 0.00846528448164463, Learning Rate: 9.901010344037786e-06, Gradient Norm: 0.2679833173751831)
49721
+ Step... (45125 | Loss: 0.0013782067690044641, Learning Rate: 9.850507922237739e-06, Gradient Norm: 0.13327546417713165)
49722
+ Step... (45150 | Loss: 0.00527406157925725, Learning Rate: 9.79999913397478e-06, Gradient Norm: 0.14178180694580078)
49723
+ Step... (45175 | Loss: 0.0010317324195057154, Learning Rate: 9.74949580268003e-06, Gradient Norm: 0.06164395809173584)
49724
+ Step... (45200 | Loss: 0.008524654433131218, Learning Rate: 9.698992471385282e-06, Gradient Norm: 0.26919326186180115)
49725
+ Step... (45225 | Loss: 0.0011031378526240587, Learning Rate: 9.648483683122322e-06, Gradient Norm: 0.0625748485326767)
49726
+ Step... (45250 | Loss: 0.011103019118309021, Learning Rate: 9.597980351827573e-06, Gradient Norm: 0.256208598613739)
49727
+ Step... (45275 | Loss: 0.0013706317404285073, Learning Rate: 9.547477930027526e-06, Gradient Norm: 0.09869988262653351)
49728
+ Step... (45300 | Loss: 0.0035194139927625656, Learning Rate: 9.496969141764566e-06, Gradient Norm: 0.1313745528459549)
49729
+ Step... (45325 | Loss: 0.0014361287467181683, Learning Rate: 9.446465810469817e-06, Gradient Norm: 0.07393680512905121)
49730
+ Step... (45350 | Loss: 0.0033916528336703777, Learning Rate: 9.395962479175068e-06, Gradient Norm: 0.11351648718118668)
49731
+ Step... (45375 | Loss: 0.005707416217774153, Learning Rate: 9.345453690912109e-06, Gradient Norm: 0.35709279775619507)
49732
+ Step... (45400 | Loss: 0.0034715337678790092, Learning Rate: 9.29495035961736e-06, Gradient Norm: 0.16799037158489227)
49733
+ Step... (45425 | Loss: 0.0010701533174142241, Learning Rate: 9.2444415713544e-06, Gradient Norm: 0.09919247031211853)
49734
+ Step... (45450 | Loss: 0.006887105293571949, Learning Rate: 9.193939149554353e-06, Gradient Norm: 0.12867707014083862)
49735
+ Step... (45475 | Loss: 0.0009366113226860762, Learning Rate: 9.143435818259604e-06, Gradient Norm: 0.06387642025947571)
49736
+ Step... (45500 | Loss: 0.0053701805882155895, Learning Rate: 9.092927029996645e-06, Gradient Norm: 0.1894822120666504)
49737
+ Step... (45525 | Loss: 0.0031545665115118027, Learning Rate: 9.042423698701896e-06, Gradient Norm: 0.20755034685134888)
49738
+ Step... (45550 | Loss: 0.005287761799991131, Learning Rate: 8.991920367407147e-06, Gradient Norm: 0.15759794414043427)
49739
+ Step... (45575 | Loss: 0.0004927063127979636, Learning Rate: 8.941411579144187e-06, Gradient Norm: 0.03520292416214943)
49740
+ Step... (45600 | Loss: 0.0040636854246258736, Learning Rate: 8.89090915734414e-06, Gradient Norm: 0.12357061356306076)
49741
+
49742
+
49743
+
49744
+
49745
+
49746
+
49747
+
49748
+
49749
+
49750
+
49751
+
49752
+
49753
+
49754
+
49755
+
49756
+
49757
+
49758
+
49759
+
49760
+
49761
+
49762
+
49763
+
49764
+
49765
+
49766
+
49767
+
49768
+
49769
+
49770
+
49771
+
49772
+
49773
+
49774
+
49775
+
49776
+
49777
+
49778
+
49779
+
49780
+
49781
+
49782
+
49783
+
49784
+
49785
+
49786
+
49787
+
49788
+
49789
+
49790
+
49791
+
49792
+
49793
+
49794
+
49795
+
49796
+
49797
+
49798
+
49799
+
49800
+
49801
+
49802
+
49803
+
49804
+
49805
+
49806
+
49807
+
49808
+
49809
+
49810
+
49811
+
49812
+
49813
+
49814
+
49815
+
49816
+
49817
+
49818
+
49819
+
49820
+
49821
+
49822
+
49823
+
49824
+
49825
+
49826
+
49827
+
49828
+
49829
+
49830
+
49831
+
49832
+
49833
+
49834
+
49835
+
49836
+
49837
+
49838
+
49839
+
49840
+
49841
+
49842
+
49843
+
49844
+
49845
+
49846
+
49847
+
49848
+
49849
+
49850
+
49851
+
49852
+
49853
+
49854
+
49855
+
49856
+
49857
+
49858
+
49859
+
49860
+
49861
+
49862
+
49863
+
49864
+
49865
+
49866
+
49867
+
49868
+
49869
+
49870
+
49871
+
49872
+
49873
+
49874
+
49875
+
49876
+
49877
+
49878
+
49879
+
49880
+
49881
+
49882
+
49883
+
49884
+
49885
+
49886
+
49887
+
49888
+
49889
+
49890
+
49891
+
49892
+
49893
+
49894
+
49895
+
49896
+
49897
+
49898
+
49899
+
49900
+
49901
+
49902
+
49903
+
49904
+
49905
+
49906
+
49907
+
49908
+
49909
+
49910
+
49911
+
49912
+
49913
+
49914
+
49915
+
49916
+
49917
+
49918
+
49919
+
49920
+
49921
+
49922
+
49923
+
49924
+
49925
+
49926
+
49927
+
49928
+
49929
+
49930
+
49931
+
49932
+
49933
+
49934
+
49935
+
49936
+
49937
+
49938
+
49939
+
49940
+
49941
+
49942
+
49943
+
49944
+
49945
+
49946
+
49947
+
49948
+
49949
+
49950
+
49951
+
49952
+
49953
+
49954
+
49955
+
49956
+
49957
+
49958
+
49959
+
49960
+
49961
+
49962
+
49963
+
49964
+
49965
+
49966
+
49967
+
49968
+
49969
+
49970
+
49971
+
49972
+
49973
+
49974
+
49975
+
49976
+
49977
+
49978
+
49979
+
49980
+
49981
+
49982
+
49983
+
49984
+
49985
+
49986
+
49987
+
49988
+
49989
+
49990
+
49991
+
49992
+
49993
+
49994
+
49995
+
49996
+
49997
+
49998
+
49999
+
50000
+
50001
+
50002
+
50003
+
50004
+
50005
+
50006
+
50007
+
50008
+
50009
+
50010
+
50011
+
50012
+
50013
+
50014
+
50015
+
50016
+
50017
+
50018
+
50019
+
50020
+
50021
+
50022
+
50023
+
50024
+
50025
+
50026
+
50027
+
50028
+
50029
+
50030
+
50031
+
50032
+
50033
+
50034
+
50035
+
50036
+
50037
+
50038
+
50039
+
50040
+
50041
+
50042
+
50043
+
50044
+
50045
+
50046
+
50047
+
50048
+
50049
+
50050
+
50051
+
50052
+
50053
+
50054
+
50055
+
50056
+
50057
+
50058
+
50059
+
50060
+
50061
+
50062
+
50063
+
50064
+
50065
+
50066
+
50067
+
50068
+
50069
+
50070
+
50071
+
50072
+
50073
+
50074
+
50075
+
50076
+
50077
+
50078
+
50079
+
50080
+
50081
+
50082
+
50083
+
50084
+
50085
+
50086
+
50087
+
50088
+
50089
+
50090
+
50091
+
50092
+
50093
+
50094
+
50095
+
50096
+
50097
+
50098
+
50099
+
50100
+
50101
+
50102
+
50103
+
50104
+
50105
+
50106
+
50107
+
50108
+
50109
+
50110
+
50111
+
50112
+
50113
+
50114
+
50115
+
50116
+
50117
+
50118
+
50119
+
50120
+
50121
+
50122
+
50123
+
50124
+
50125
+
50126
+
50127
+
50128
+
50129
+
50130
+
50131
+
50132
+
50133
+
50134
+
50135
+
50136
+
50137
+
50138
+
50139
+
50140
+
50141
+
50142
+
50143
+
50144
+
50145
+
50146
+
50147
+
50148
+
50149
+
50150
+
50151
+
50152
+
50153
+
50154
+
50155
+
50156
+
50157
+
50158
+
50159
+
50160
+
50161
+
50162
+
50163
+
50164
+
50165
+
50166
+
50167
+
50168
+
50169
+
50170
+
50171
+
50172
+
50173
+
50174
+
50175
+
50176
+
50177
+
50178
+
50179
+
50180
+
50181
+
50182
+
50183
+
50184
+
50185
+
50186
+
50187
+
50188
+
50189
+
50190
+
50191
+
50192
+
50193
+
50194
+
50195
+
50196
+
50197
+
50198
+
50199
+
50200
+
50201
+
50202
+
50203
+
50204
+
50205
+
50206
+
50207
+
50208
+
50209
+
50210
+
50211
+
50212
+
50213
+
50214
+
50215
+
50216
+
50217
+
50218
+
50219
+
50220
+
50221
+
50222
+
50223
+
50224
+
50225
+
50226
+
50227
+
50228
+
50229
+
50230
+
50231
+
50232
+
50233
+
50234
+
50235
+
50236
+
50237
+
50238
+
50239
+
50240
+
50241
+
50242
+
50243
+
50244
+
50245
+
50246
+
50247
+
50248
+
50249
+
50250
+
50251
+
50252
+
50253
+
50254
+
50255
+
50256
+
50257
+
50258
+
50259
+
50260
+
50261
+
50262
+
50263
+
50264
+
50265
+
50266
+
50267
+
50268
+
50269
+
50270
+
50271
+
50272
+
50273
+
50274
+
50275
+
50276
+
50277
+
50278
+
50279
+
50280
+
50281
+
50282
+
50283
+
50284
+
50285
+
50286
+
50287
+
50288
+
50289
+
50290
+
50291
+
50292
+
50293
+
50294
+
50295
+
50296
+
50297
+
50298
+
50299
+
50300
+
50301
+
50302
+
50303
+
50304
+
50305
+
50306
+
50307
+
50308
+
50309
+
50310
+
50311
+
50312
+
50313
+
50314
+
50315
+
50316
+
50317
+
50318
+
50319
+
50320
+
50321
+
50322
+
50323
+
50324
+
50325
+
50326
+
50327
+
50328
+
50329
+
50330
+
50331
+
50332
+
50333
+
50334
+
50335
+
50336
+
50337
+
50338
+
50339
+
50340
+
50341
+
50342
+
50343
+
50344
+
50345
+
50346
+
50347
+
50348
+
50349
+
50350
+
50351
+
50352
+
50353
+
50354
+
50355
+
50356
+
50357
+
50358
+
50359
+
50360
+
50361
+
50362
+
50363
+
50364
+
50365
+
50366
+
50367
+
50368
+
50369
+
50370
+
50371
+
50372
+
50373
+
50374
+
50375
+
50376
+
50377
+
50378
+
50379
+
50380
+
50381
+
50382
+
50383
+
50384
+
50385
+
50386
+
50387
+
50388
+
50389
+
50390
+
50391
+
50392
+
50393
+
50394
+
50395
+
50396
+
50397
+
50398
+
50399
+
50400
+
50401
+
50402
+
50403
+
50404
+
50405
+
50406
+
50407
+
50408
+
50409
+
50410
+
50411
+
50412
+
50413
+
50414
+
50415
+
50416
+
50417
+
50418
+
50419
+
50420
+
50421
+
50422
+
50423
+
50424
+
50425
+
50426
+
50427
+
50428
+
50429
+
50430
+
50431
+
50432
+
50433
+
50434
+
50435
+
50436
+
50437
+
50438
+
50439
+
50440
+
50441
+
50442
+
50443
+
50444
+
50445
+
50446
+
50447
+
50448
+
50449
+
50450
+
50451
+
50452
+
50453
+
50454
+
50455
+
50456
+
50457
+
50458
+
50459
+
50460
+
50461
+
50462
+
50463
+
50464
+
50465
+
50466
+
50467
+
50468
+
50469
+
50470
+
50471
+
50472
+
50473
+
50474
+
50475
+
50476
+
50477
+
50478
+
50479
+
50480
+
50481
+
50482
+
50483
+
50484
+
50485
+
50486
+
50487
+
50488
+
50489
+
50490
+
50491
+
50492
+
50493
+
50494
+
50495
+
50496
+
50497
+
50498
+
50499
+
50500
+
50501
+
50502
+
50503
+
50504
+
50505
+
50506
+
50507
+
50508
+
50509
+
50510
+
50511
+
50512
+
50513
+
50514
+
50515
+
50516
+
50517
+
50518
+
50519
+
50520
+
50521
+
50522
+
50523
+
50524
+
50525
+
50526
+
50527
+
50528
+
50529
+
50530
+
50531
+
50532
+
50533
+
50534
+
50535
+
50536
+
50537
+
50538
+
50539
+
50540
+
50541
+
50542
+
50543
+
50544
+
50545
+
50546
+
50547
+
50548
+
50549
+
50550
+
50551
+
50552
+
50553
+
50554
+
50555
+
50556
+
50557
+
50558
+
50559
+
50560
+
50561
+
50562
+
50563
+
50564
+
50565
+
50566
+
50567
+
50568
+
50569
+
50570
+
50571
+
50572
+
50573
+
50574
+
50575
+
50576
+
50577
+
50578
+
50579
+
50580
+
50581
+
50582
+
50583
+
50584
+
50585
+
50586
+
50587
+
50588
+
50589
+
50590
+
50591
+
50592
+
50593
+
50594
+
50595
+
50596
+
50597
+
50598
+
50599
+
50600
+
50601
+
50602
+
50603
+
50604
+
50605
+
50606
+
50607
+
50608
+
50609
+
50610
+
50611
+
50612
+
50613
+
50614
+
50615
+
50616
+
50617
+
50618
+
50619
+
50620
+
50621
+
50622
+
50623
+
50624
+
50625
+
50626
+
50627
+
50628
+
50629
+
50630
+
50631
+
50632
+
50633
+
50634
+
50635
+
50636
+
50637
+
50638
+
50639
+
50640
+
50641
+
50642
+
50643
+
50644
+
50645
+
50646
+
50647
+
50648
+
50649
+
50650
+
50651
+
50652
+
50653
+
50654
+
50655
+
50656
+
50657
+
50658
+
50659
+
50660
+
50661
+
50662
+
50663
+
50664
+
50665
+
50666
+
50667
+
50668
+
50669
+
50670
+
50671
+
50672
+
50673
+
50674
+
50675
+
50676
+
50677
+
50678
+
50679
+
50680
+
50681
+
50682
+
50683
+
50684
+
50685
+
50686
+
50687
+
50688
+
50689
+
50690
+
50691
+
50692
+
50693
+
50694
+
50695
+
50696
+
50697
+
50698
+
50699
+
50700
+
50701
+
50702
+
50703
+
50704
+
50705
+
50706
+
50707
+
50708
+
50709
+
50710
+
50711
+
50712
+
50713
+
50714
+
50715
+
50716
+
50717
+
50718
+
50719
+
50720
+
50721
+
50722
+
50723
+
50724
+
50725
+
50726
+
50727
+
50728
+
50729
+
50730
+
50731
+
50732
+
50733
+
50734
+
50735
+
50736
+
50737
+
50738
+
50739
+
50740
+
50741
+
50742
+
50743
+
50744
+
50745
+
50746
+
50747
+
50748
+
50749
+
50750
+
50751
+
50752
+
50753
+
50754
+
50755
+
50756
+
50757
+
50758
+
50759
+
50760
+
50761
+
50762
+
50763
+
50764
+
50765
+
50766
+
50767
+
50768
+
50769
+
50770
+
50771
+
50772
+
50773
+
50774
+
50775
+
50776
+
50777
+
50778
+
50779
+
50780
+
50781
+
50782
+
50783
+
50784
+
50785
+
50786
+
50787
+
50788
+
50789
+
50790
+
50791
+
50792
+
50793
+
50794
+
50795
+
50796
+
50797
+
50798
+
50799
+
50800
+
50801
+
50802
+
50803
+
50804
+
50805
+
50806
+
50807
+
50808
+
50809
+
50810
+
50811
+
50812
+
50813
+
50814
+
50815
+
50816
+
50817
+
50818
+
50819
+
50820
+
50821
+
50822
+
50823
+
50824
+
50825
+
50826
+
50827
+
50828
+
50829
+
50830
+
50831
+
50832
+
50833
+
50834
+
50835
+
50836
+
50837
+
50838
+
50839
+
50840
+
50841
+
50842
+
50843
+
50844
+
50845
+
50846
+
50847
+
50848
+
50849
+
50850
+
50851
+
50852
+
50853
+
50854
+
50855
+
50856
+
50857
+
50858
+
50859
+
50860
+
50861
+
50862
+
50863
+
50864
+
50865
+
50866
+
50867
+
50868
+
50869
+
50870
+
50871
+
50872
+
50873
+
50874
+
50875
+
50876
+
50877
+
50878
+
50879
+
50880
+
50881
+
50882
+
50883
+
50884
+
50885
+
50886
+
50887
+
50888
+
50889
+
50890
+
50891
+
50892
+
50893
+
50894
+
50895
+
50896
+
50897
+
50898
+
50899
+
50900
+
50901
+
50902
+
50903
+
50904
+
50905
+
50906
+
50907
+
50908
+
50909
+
50910
+
50911
+
50912
+
50913
+
50914
+
50915
+
50916
+
50917
+
50918
+
50919
+
50920
+
50921
+
50922
+
50923
+
50924
+
50925
+
50926
+
50927
+
50928
+
50929
+
50930
+
50931
+
50932
+
50933
+
50934
+
50935
+
50936
+
50937
+
50938
+
50939
+
50940
+
50941
+
50942
+
50943
+
50944
+
50945
+
50946
+
50947
+
50948
+
50949
+
50950
+
50951
+
50952
+
50953
+
50954
+
50955
+
50956
+
50957
+
50958
+
50959
+
50960
+
50961
+
50962
+
50963
+
50964
+
50965
+
50966
+
50967
+
50968
+
50969
+
50970
+
50971
+
50972
+
50973
+
50974
+
50975
+
50976
+
50977
+
50978
+
50979
+
50980
+
50981
+
50982
+
50983
+
50984
+
50985
+
50986
+
50987
+
50988
+
50989
+
50990
+
50991
+
50992
+
50993
+
50994
+
50995
+
50996
+
50997
+
50998
+
50999
+
51000
+
51001
+
51002
+
51003
+
51004
+
51005
+
51006
+
51007
+
51008
+
51009
+
51010
+
51011
+
51012
+
51013
+
51014
+
51015
+
51016
+
51017
+
51018
+
51019
+
51020
+
51021
+
51022
+
51023
+
51024
+
51025
+
51026
+
51027
+
51028
+
51029
+
51030
+
51031
+
51032
+
51033
+
51034
+
51035
+
51036
+
51037
+
51038
+
51039
+
51040
+
51041
+
51042
+
51043
+
51044
+
51045
+
51046
+
51047
+
51048
+
51049
+
51050
+
51051
+
51052
+
51053
+
51054
+
51055
+
51056
+
51057
+
51058
+
51059
+
51060
+
51061
+
51062
+
51063
+
51064
+
51065
+
51066
+
51067
+
51068
+
51069
+
51070
+
51071
+
51072
+
51073
+
51074
+
51075
+
51076
+
51077
+
51078
+
51079
+
51080
+
51081
+
51082
+
51083
+
51084
+
51085
+
51086
+
51087
+
51088
+
51089
+
51090
+
51091
+
51092
+
51093
+
51094
+
51095
+
51096
+
51097
+
51098
+
51099
+
51100
+
51101
+
51102
+
51103
+
51104
+
51105
+
51106
+
51107
+
51108
+
51109
+
51110
+
51111
+
51112
+
51113
+
51114
+
51115
+
51116
+
51117
+
51118
+
51119
+
51120
+
51121
+
51122
+
51123
+
51124
+
51125
+
51126
+
51127
+
51128
+
51129
+
51130
+
51131
+
51132
+
51133
+
51134
+
51135
+
51136
+
51137
+
51138
+
51139
+
51140
+
51141
+
51142
+
51143
+
51144
+
51145
+
51146
+
51147
+
51148
+
51149
+
51150
+
51151
+
51152
+
51153
+
51154
+
51155
+
51156
+
51157
+
51158
+
51159
+
51160
+
51161
+
51162
+
51163
+
51164
+
51165
+
51166
+
51167
+
51168
+
51169
+
51170
+
51171
+
51172
+
51173
+
51174
+
51175
+
51176
+
51177
+
51178
+
51179
+
51180
+
51181
+
51182
+
51183
+
51184
+
51185
+
51186
+
51187
+
51188
+
51189
+
51190
+
51191
+
51192
+
51193
+
51194
+
51195
+
51196
+
51197
+
51198
+
51199
+
51200
+
51201
+
51202
+
51203
+
51204
+
51205
+
51206
+
51207
+
51208
+
51209
+
51210
+
51211
+
51212
+
51213
+
51214
+
51215
+
51216
+
51217
+
51218
+
51219
+
51220
+
51221
+
51222
+
51223
+
51224
+
51225
+
51226
+
51227
+
51228
+
51229
+
51230
+
51231
+
51232
+
51233
+
51234
+
51235
+
51236
+
51237
+
51238
+
51239
+
51240
+
51241
+
51242
+
51243
+
51244
+
51245
+
51246
+
51247
+
51248
+
51249
+
51250
+
51251
+
51252
+
51253
+
51254
+
51255
+
51256
+
51257
+
51258
+
51259
+
51260
+
51261
+
51262
+
51263
+
51264
+
51265
+
51266
+
51267
+
51268
+
51269
+
51270
+
51271
+
51272
+
51273
+
51274
+
51275
+
51276
+
51277
+
51278
+
51279
+
51280
+
51281
+
51282
+
51283
+
51284
+
51285
+
51286
+
51287
+
51288
+
51289
+
51290
+
51291
+
51292
+
51293
+
51294
+
51295
+
51296
+
51297
+
51298
+
51299
+
51300
+
51301
+
51302
+
51303
+
51304
+
51305
+
51306
+
51307
+
51308
+
51309
+
51310
+
51311
+
51312
+
51313
+
51314
+
51315
+
51316
+
51317
+
51318
+
51319
+
51320
+
51321
+
51322
+
51323
+
51324
+
51325
+
51326
+
51327
+
51328
+
51329
+
51330
+
51331
+
51332
+
51333
+
51334
+
51335
+
51336
+
51337
+
51338
+
51339
+
51340
+
51341
+
51342
+
51343
+
51344
+
51345
+
51346
+
51347
+
51348
+
51349
+
51350
+
51351
+
51352
+
51353
+
51354
+
51355
+
51356
+
51357
+
51358
+
51359
+
51360
+
51361
+
51362
+
51363
+
51364
+
51365
+
51366
+
51367
+
51368
+
51369
+
51370
+
51371
+
51372
+
51373
+
51374
+
51375
+
51376
+
51377
+
51378
+
51379
+
51380
+
51381
+
51382
+
51383
+
51384
+
51385
+
51386
+
51387
+
51388
+
51389
+
51390
+
51391
+
51392
+
51393
+
51394
+
51395
+
51396
+
51397
+
51398
+
51399
+
51400
+
51401
+ Training...: 78% 3420/4393 [4:31:14<1:34:13, 5.81s/it]
51402
+ Step... (45650 | Loss: 0.0038062885869294405, Learning Rate: 8.789897037786432e-06, Gradient Norm: 0.12565107643604279)
51403
+ Step... (45675 | Loss: 0.001629532314836979, Learning Rate: 8.739393706491683e-06, Gradient Norm: 0.1391201615333557)
51404
+ Step... (45700 | Loss: 0.002758459188044071, Learning Rate: 8.688890375196934e-06, Gradient Norm: 0.14725351333618164)
51405
+ Step... (45725 | Loss: 0.0015460265567526221, Learning Rate: 8.638381586933974e-06, Gradient Norm: 0.12334706634283066)
51406
+ Step... (45750 | Loss: 0.008942586369812489, Learning Rate: 8.587879165133927e-06, Gradient Norm: 0.19431252777576447)
51407
+ Step... (45775 | Loss: 0.0018675904721021652, Learning Rate: 8.537375833839178e-06, Gradient Norm: 0.22979864478111267)
51408
+ Step... (45800 | Loss: 0.0032472757156938314, Learning Rate: 8.486867045576219e-06, Gradient Norm: 0.16855953633785248)
51409
+ Step... (45825 | Loss: 0.0020079060923308134, Learning Rate: 8.43636371428147e-06, Gradient Norm: 0.1927141398191452)
51410
+ Step... (45850 | Loss: 0.002314024604856968, Learning Rate: 8.38586038298672e-06, Gradient Norm: 0.10700368136167526)
51411
+ Step... (45875 | Loss: 0.0023375945165753365, Learning Rate: 8.335351594723761e-06, Gradient Norm: 0.25536608695983887)
51412
+ Step... (45900 | Loss: 0.005266377702355385, Learning Rate: 8.284849172923714e-06, Gradient Norm: 0.14284679293632507)
51413
+ Step... (45925 | Loss: 0.002833940787240863, Learning Rate: 8.234345841628965e-06, Gradient Norm: 0.13957801461219788)
51414
+ Step... (45950 | Loss: 0.004558590240776539, Learning Rate: 8.183837053366005e-06, Gradient Norm: 0.1979239284992218)
51415
+ Step... (45975 | Loss: 0.0004439667100086808, Learning Rate: 8.133333722071256e-06, Gradient Norm: 0.023552939295768738)
51416
+ Step... (46000 | Loss: 0.003100600326433778, Learning Rate: 8.082830390776508e-06, Gradient Norm: 0.09397352486848831)
51417
+ Step... (46025 | Loss: 0.00844496488571167, Learning Rate: 8.032321602513548e-06, Gradient Norm: 0.34702354669570923)
51418
+ Step... (46050 | Loss: 0.00445242365822196, Learning Rate: 7.981818271218799e-06, Gradient Norm: 0.14293652772903442)
51419
+ Step... (46075 | Loss: 0.0017299428582191467, Learning Rate: 7.931315849418752e-06, Gradient Norm: 0.1346822828054428)
51420
+ Step... (46100 | Loss: 0.004057885613292456, Learning Rate: 7.880807061155792e-06, Gradient Norm: 0.12305817753076553)
51421
+ Step... (46125 | Loss: 0.001030999468639493, Learning Rate: 7.830303729861043e-06, Gradient Norm: 0.05135162174701691)
51422
+ Step... (46150 | Loss: 0.002076568314805627, Learning Rate: 7.779800398566294e-06, Gradient Norm: 0.10389827191829681)
51423
+ Step... (46175 | Loss: 0.0029374691657721996, Learning Rate: 7.729291610303335e-06, Gradient Norm: 0.2249249517917633)
51424
+ Step... (46200 | Loss: 0.003002022160217166, Learning Rate: 7.678788279008586e-06, Gradient Norm: 0.20660805702209473)
51425
+ Step... (46225 | Loss: 0.0015353269409388304, Learning Rate: 7.628285857208539e-06, Gradient Norm: 0.13045524060726166)
51426
+ Step... (46250 | Loss: 0.010492435656487942, Learning Rate: 7.577776614198228e-06, Gradient Norm: 0.2511439621448517)
51427
+ Step... (46275 | Loss: 0.001488243229687214, Learning Rate: 7.52727373765083e-06, Gradient Norm: 0.11494113504886627)
51428
+ Step... (46300 | Loss: 0.002643989399075508, Learning Rate: 7.476770861103432e-06, Gradient Norm: 0.12378372997045517)
51429
+ Step... (46325 | Loss: 0.0028585607651621103, Learning Rate: 7.426261618093122e-06, Gradient Norm: 0.17171597480773926)
51430
+ Step... (46350 | Loss: 0.005086139775812626, Learning Rate: 7.375758741545724e-06, Gradient Norm: 0.11662453413009644)
51431
+ Step... (46375 | Loss: 0.001086848322302103, Learning Rate: 7.325255864998326e-06, Gradient Norm: 0.06781523674726486)
51432
+ Step... (46400 | Loss: 0.002841377630829811, Learning Rate: 7.274746621988015e-06, Gradient Norm: 0.13784265518188477)
51433
+ Step... (46425 | Loss: 0.0030452024657279253, Learning Rate: 7.224243745440617e-06, Gradient Norm: 0.27625468373298645)
51434
+ Step... (46450 | Loss: 0.010637513361871243, Learning Rate: 7.173734502430307e-06, Gradient Norm: 0.2823644280433655)
51435
+ Step... (46475 | Loss: 0.004557167179882526, Learning Rate: 7.123231625882909e-06, Gradient Norm: 0.536411702632904)
51436
+ Step... (46500 | Loss: 0.00306858797557652, Learning Rate: 7.072728749335511e-06, Gradient Norm: 0.09498882293701172)
51437
+ Step... (46525 | Loss: 0.0018582177581265569, Learning Rate: 7.0222195063252e-06, Gradient Norm: 0.15445643663406372)
51438
+ Step... (46550 | Loss: 0.0055601089261472225, Learning Rate: 6.971716629777802e-06, Gradient Norm: 0.2561977505683899)
51439
+ Step... (46575 | Loss: 0.0004325220361351967, Learning Rate: 6.921213753230404e-06, Gradient Norm: 0.02027466520667076)
51440
+ Step... (46600 | Loss: 0.004028357565402985, Learning Rate: 6.870704510220094e-06, Gradient Norm: 0.14455808699131012)
51441
+ Step... (46625 | Loss: 0.0032141306437551975, Learning Rate: 6.820201633672696e-06, Gradient Norm: 0.19990961253643036)
51442
+ Step... (46650 | Loss: 0.003672192106023431, Learning Rate: 6.7696987571252976e-06, Gradient Norm: 0.0882977843284607)
51443
+ Step... (46675 | Loss: 0.002671650843694806, Learning Rate: 6.719189514114987e-06, Gradient Norm: 0.29029080271720886)
51444
+ Step... (46700 | Loss: 0.0021792224142700434, Learning Rate: 6.668686637567589e-06, Gradient Norm: 0.0928763747215271)
51445
+ Step... (46725 | Loss: 0.0018916006665676832, Learning Rate: 6.618183761020191e-06, Gradient Norm: 0.18946678936481476)
51446
+ Step... (46750 | Loss: 0.002835277933627367, Learning Rate: 6.5676745180098806e-06, Gradient Norm: 0.1169157475233078)
51447
+ Step... (46775 | Loss: 0.0008665637578815222, Learning Rate: 6.5171716414624825e-06, Gradient Norm: 0.06307638436555862)
51448
+ Step... (46800 | Loss: 0.005216824356466532, Learning Rate: 6.4666687649150845e-06, Gradient Norm: 0.17191486060619354)
51449
+ Step... (46825 | Loss: 0.0019698829855769873, Learning Rate: 6.416159521904774e-06, Gradient Norm: 0.12737050652503967)
51450
+ Step... (46850 | Loss: 0.002469806233420968, Learning Rate: 6.365656645357376e-06, Gradient Norm: 0.09056347608566284)
51451
+ Step... (46875 | Loss: 0.0011732580605894327, Learning Rate: 6.315153768809978e-06, Gradient Norm: 0.07834524661302567)
51452
+ Step... (46900 | Loss: 0.0027421254198998213, Learning Rate: 6.2646445257996675e-06, Gradient Norm: 0.12933117151260376)
51453
+ Step... (46925 | Loss: 0.002762808231636882, Learning Rate: 6.214141649252269e-06, Gradient Norm: 0.2802877128124237)
51454
+ Step... (46950 | Loss: 0.003183376742526889, Learning Rate: 6.163638772704871e-06, Gradient Norm: 0.09256728738546371)
51455
+ Step... (46975 | Loss: 0.0011139512062072754, Learning Rate: 6.113129529694561e-06, Gradient Norm: 0.11546455323696136)
51456
+ Step... (47000 | Loss: 0.003426450304687023, Learning Rate: 6.062626653147163e-06, Gradient Norm: 0.13084343075752258)
51457
+ Step... (47025 | Loss: 0.0024451303761452436, Learning Rate: 6.012123776599765e-06, Gradient Norm: 0.28388726711273193)
51458
+ Step... (47050 | Loss: 0.0022857924923300743, Learning Rate: 5.961614533589454e-06, Gradient Norm: 0.14490585029125214)
51459
+ Step... (47075 | Loss: 0.0015037512639537454, Learning Rate: 5.911111657042056e-06, Gradient Norm: 0.20273494720458984)
51460
+ Step... (47100 | Loss: 0.004381109494715929, Learning Rate: 5.860608780494658e-06, Gradient Norm: 0.1728397160768509)
51461
+ Step... (47125 | Loss: 0.0018952905666083097, Learning Rate: 5.810099537484348e-06, Gradient Norm: 0.1924336552619934)
51462
+ Step... (47150 | Loss: 0.003915037028491497, Learning Rate: 5.75959666093695e-06, Gradient Norm: 0.13231444358825684)
51463
+ Step... (47175 | Loss: 0.0013379824813455343, Learning Rate: 5.709093784389552e-06, Gradient Norm: 0.09984851628541946)
51464
+ Step... (47200 | Loss: 0.0042300745844841, Learning Rate: 5.658584541379241e-06, Gradient Norm: 0.15374016761779785)
51465
+ Step... (47225 | Loss: 0.0012262854725122452, Learning Rate: 5.608081664831843e-06, Gradient Norm: 0.10231379419565201)
51466
+ Step... (47250 | Loss: 0.0038886629045009613, Learning Rate: 5.557578788284445e-06, Gradient Norm: 0.15827754139900208)
51467
+ Step... (47275 | Loss: 0.0015737712383270264, Learning Rate: 5.507069545274135e-06, Gradient Norm: 0.12450241297483444)
51468
+ Step... (47300 | Loss: 0.0057818274945020676, Learning Rate: 5.456566668726737e-06, Gradient Norm: 0.14007054269313812)
51469
+
51470
+
51471
+
51472
+
51473
+
51474
+
51475
+
51476
+
51477
+
51478
+
51479
+
51480
+
51481
+
51482
+
51483
+
51484
+
51485
+
51486
+
51487
+
51488
+
51489
+
51490
+
51491
+
51492
+
51493
+
51494
+
51495
+
51496
+
51497
+
51498
+
51499
+
51500
+
51501
+
51502
+
51503
+
51504
+
51505
+
51506
+
51507
+
51508
+
51509
+
51510
+
51511
+
51512
+
51513
+
51514
+
51515
+
51516
+
51517
+
51518
+
51519
+
51520
+
51521
+
51522
+
51523
+
51524
+
51525
+
51526
+
51527
+
51528
+
51529
+
51530
+
51531
+
51532
+
51533
+
51534
+
51535
+
51536
+
51537
+
51538
+
51539
+
51540
+
51541
+
51542
+
51543
+
51544
+
51545
+
51546
+
51547
+
51548
+
51549
+
51550
+
51551
+
51552
+
51553
+
51554
+
51555
+
51556
+
51557
+
51558
+
51559
+
51560
+
51561
+
51562
+
51563
+
51564
+
51565
+
51566
+
51567
+
51568
+
51569
+
51570
+
51571
+
51572
+
51573
+
51574
+
51575
+
51576
+
51577
+
51578
+
51579
+
51580
+
51581
+
51582
+
51583
+
51584
+
51585
+
51586
+
51587
+
51588
+
51589
+
51590
+
51591
+
51592
+
51593
+
51594
+
51595
+
51596
+
51597
+
51598
+
51599
+
51600
+
51601
+
51602
+
51603
+
51604
+
51605
+
51606
+
51607
+
51608
+
51609
+
51610
+
51611
+
51612
+
51613
+
51614
+
51615
+
51616
+
51617
+
51618
+
51619
+
51620
+
51621
+
51622
+
51623
+
51624
+
51625
+
51626
+
51627
+
51628
+
51629
+
51630
+
51631
+
51632
+
51633
+
51634
+
51635
+
51636
+
51637
+
51638
+
51639
+
51640
+
51641
+
51642
+
51643
+
51644
+
51645
+
51646
+
51647
+
51648
+
51649
+
51650
+
51651
+
51652
+
51653
+
51654
+
51655
+
51656
+
51657
+
51658
+
51659
+
51660
+
51661
+
51662
+
51663
+
51664
+
51665
+
51666
+
51667
+
51668
+
51669
+
51670
+
51671
+
51672
+
51673
+
51674
+
51675
+
51676
+
51677
+
51678
+
51679
+
51680
+
51681
+
51682
+
51683
+
51684
+
51685
+
51686
+
51687
+
51688
+
51689
+
51690
+
51691
+
51692
+
51693
+
51694
+
51695
+
51696
+
51697
+
51698
+
51699
+
51700
+
51701
+
51702
+
51703
+
51704
+
51705
+
51706
+
51707
+
51708
+
51709
+
51710
+
51711
+
51712
+
51713
+
51714
+
51715
+
51716
+
51717
+
51718
+
51719
+
51720
+
51721
+
51722
+
51723
+
51724
+
51725
+
51726
+
51727
+
51728
+
51729
+
51730
+
51731
+
51732
+
51733
+
51734
+
51735
+
51736
+
51737
+
51738
+
51739
+
51740
+
51741
+
51742
+
51743
+
51744
+
51745
+
51746
+
51747
+
51748
+
51749
+
51750
+
51751
+
51752
+
51753
+
51754
+
51755
+
51756
+
51757
+
51758
+
51759
+
51760
+
51761
+
51762
+
51763
+
51764
+
51765
+
51766
+
51767
+
51768
+
51769
+
51770
+
51771
+
51772
+
51773
+
51774
+
51775
+
51776
+
51777
+
51778
+
51779
+
51780
+
51781
+
51782
+
51783
+
51784
+
51785
+
51786
+
51787
+
51788
+
51789
+
51790
+
51791
+
51792
+
51793
+
51794
+
51795
+
51796
+
51797
+
51798
+
51799
+
51800
+
51801
+
51802
+
51803
+
51804
+
51805
+
51806
+
51807
+
51808
+
51809
+
51810
+
51811
+
51812
+
51813
+
51814
+
51815
+
51816
+
51817
+
51818
+
51819
+
51820
+
51821
+
51822
+
51823
+
51824
+
51825
+
51826
+
51827
+
51828
+
51829
+
51830
+
51831
+
51832
+
51833
+
51834
+
51835
+
51836
+
51837
+
51838
+
51839
+
51840
+
51841
+
51842
+
51843
+
51844
+
51845
+
51846
+
51847
+
51848
+
51849
+
51850
+
51851
+
51852
+
51853
+
51854
+
51855
+
51856
+
51857
+
51858
+
51859
+
51860
+
51861
+
51862
+
51863
+
51864
+
51865
+
51866
+
51867
+
51868
+
51869
+
51870
+
51871
+
51872
+
51873
+
51874
+
51875
+
51876
+
51877
+
51878
+
51879
+
51880
+
51881
+
51882
+
51883
+
51884
+
51885
+
51886
+
51887
+
51888
+
51889
+
51890
+
51891
+
51892
+
51893
+
51894
+
51895
+
51896
+
51897
+
51898
+
51899
+
51900
+
51901
+
51902
+
51903
+
51904
+
51905
+
51906
+
51907
+
51908
+
51909
+
51910
+
51911
+
51912
+
51913
+
51914
+
51915
+
51916
+
51917
+
51918
+
51919
+
51920
+
51921
+
51922
+
51923
+
51924
+
51925
+
51926
+
51927
+
51928
+
51929
+
51930
+
51931
+
51932
+
51933
+
51934
+
51935
+
51936
+
51937
+
51938
+
51939
+
51940
+
51941
+
51942
+
51943
+
51944
+
51945
+
51946
+
51947
+
51948
+
51949
+
51950
+
51951
+
51952
+
51953
+
51954
+
51955
+
51956
+
51957
+
51958
+
51959
+
51960
+
51961
+
51962
+
51963
+
51964
+
51965
+
51966
+
51967
+
51968
+
51969
+
51970
+
51971
+
51972
+
51973
+
51974
+
51975
+
51976
+
51977
+
51978
+
51979
+
51980
+
51981
+
51982
+
51983
+
51984
+
51985
+
51986
+
51987
+
51988
+
51989
+
51990
+
51991
+
51992
+
51993
+
51994
+
51995
+
51996
+
51997
+
51998
+
51999
+
52000
+
52001
+
52002
+
52003
+
52004
+
52005
+
52006
+
52007
+
52008
+
52009
+
52010
+
52011
+
52012
+
52013
+
52014
+
52015
+
52016
+
52017
+
52018
+
52019
+
52020
+
52021
+
52022
+
52023
+
52024
+
52025
+
52026
+
52027
+
52028
+
52029
+
52030
+
52031
+
52032
+
52033
+
52034
+
52035
+
52036
+
52037
+
52038
+
52039
+
52040
+
52041
+
52042
+
52043
+
52044
+
52045
+
52046
+
52047
+
52048
+
52049
+
52050
+
52051
+
52052
+
52053
+
52054
+
52055
+
52056
+
52057
+
52058
+
52059
+
52060
+
52061
+
52062
+
52063
+
52064
+
52065
+
52066
+
52067
+
52068
+
52069
+
52070
+
52071
+
52072
+
52073
+
52074
+
52075
+
52076
+
52077
+
52078
+
52079
+
52080
+
52081
+
52082
+
52083
+
52084
+
52085
+
52086
+
52087
+
52088
+
52089
+
52090
+
52091
+
52092
+
52093
+
52094
+
52095
+
52096
+
52097
+
52098
+
52099
+
52100
+
52101
+
52102
+
52103
+
52104
+
52105
+
52106
+
52107
+
52108
+
52109
+
52110
+
52111
+
52112
+
52113
+
52114
+
52115
+
52116
+
52117
+
52118
+
52119
+
52120
+
52121
+
52122
+
52123
+
52124
+
52125
+
52126
+
52127
+
52128
+
52129
+
52130
+
52131
+
52132
+
52133
+
52134
+
52135
+
52136
+
52137
+
52138
+
52139
+
52140
+
52141
+
52142
+
52143
+
52144
+
52145
+
52146
+
52147
+
52148
+
52149
+
52150
+
52151
+
52152
+
52153
+
52154
+
52155
+
52156
+
52157
+
52158
+
52159
+
52160
+
52161
+
52162
+
52163
+
52164
+
52165
+
52166
+
52167
+
52168
+
52169
+
52170
+
52171
+
52172
+
52173
+
52174
+
52175
+
52176
+
52177
+
52178
+
52179
+
52180
+
52181
+
52182
+
52183
+
52184
+
52185
+
52186
+
52187
+
52188
+
52189
+
52190
+
52191
+
52192
+
52193
+
52194
+
52195
+
52196
+
52197
+
52198
+
52199
+
52200
+
52201
+
52202
+
52203
+
52204
+
52205
+
52206
+
52207
+
52208
+
52209
+
52210
+
52211
+
52212
+
52213
+
52214
+
52215
+
52216
+
52217
+
52218
+
52219
+
52220
+
52221
+
52222
+
52223
+
52224
+
52225
+
52226
+
52227
+
52228
+
52229
+
52230
+
52231
+
52232
+
52233
+
52234
+
52235
+
52236
+
52237
+
52238
+
52239
+
52240
+
52241
+
52242
+
52243
+
52244
+
52245
+
52246
+
52247
+
52248
+
52249
+
52250
+
52251
+
52252
+
52253
+
52254
+
52255
+
52256
+
52257
+
52258
+
52259
+
52260
+
52261
+
52262
+
52263
+
52264
+
52265
+
52266
+
52267
+
52268
+
52269
+
52270
+
52271
+
52272
+
52273
+
52274
+
52275
+
52276
+
52277
+
52278
+
52279
+
52280
+
52281
+
52282
+
52283
+
52284
+
52285
+
52286
+
52287
+
52288
+
52289
+
52290
+
52291
+
52292
+
52293
+
52294
+
52295
+
52296
+
52297
+
52298
+
52299
+
52300
+
52301
+
52302
+
52303
+
52304
+
52305
+
52306
+
52307
+
52308
+
52309
+
52310
+
52311
+
52312
+
52313
+
52314
+
52315
+
52316
+
52317
+
52318
+
52319
+
52320
+
52321
+
52322
+
52323
+
52324
+
52325
+
52326
+
52327
+
52328
+
52329
+
52330
+
52331
+
52332
+
52333
+
52334
+
52335
+
52336
+
52337
+
52338
+
52339
+
52340
+
52341
+
52342
+
52343
+
52344
+
52345
+
52346
+
52347
+
52348
+
52349
+
52350
+
52351
+
52352
+
52353
+
52354
+
52355
+
52356
+
52357
+
52358
+
52359
+
52360
+
52361
+
52362
+
52363
+
52364
+
52365
+
52366
+
52367
+
52368
+
52369
+
52370
+
52371
+
52372
+
52373
+
52374
+
52375
+
52376
+
52377
+
52378
+
52379
+
52380
+
52381
+
52382
+
52383
+
52384
+
52385
+
52386
+
52387
+
52388
+
52389
+
52390
+
52391
+
52392
+
52393
+
52394
+
52395
+
52396
+
52397
+
52398
+
52399
+
52400
+
52401
+
52402
+
52403
+
52404
+
52405
+
52406
+
52407
+
52408
+
52409
+
52410
+
52411
+
52412
+
52413
+
52414
+
52415
+
52416
+
52417
+
52418
+
52419
+
52420
+
52421
+ Step... (40000/50000 | Eval Loss: 0.9527401328086853 | Eval wer: 0.04187346053453917 | Eval cer: 0.02872188479352137 |): 92% 11/12 [64:22:01<5:51:07, 21067.42s/it]
52422
+ Step... (47350 | Loss: 0.003852912224829197, Learning Rate: 5.355554549169028e-06, Gradient Norm: 0.1632377952337265)
52423
+ Step... (47375 | Loss: 0.0011032108450308442, Learning Rate: 5.30505167262163e-06, Gradient Norm: 0.07574035227298737)
52424
+ Step... (47400 | Loss: 0.0037336035165935755, Learning Rate: 5.254548796074232e-06, Gradient Norm: 0.1369144171476364)
52425
+ Step... (47425 | Loss: 0.0005997814005240798, Learning Rate: 5.204039553063922e-06, Gradient Norm: 0.06363093107938766)
52426
+ Step... (47450 | Loss: 0.0032057648058980703, Learning Rate: 5.153536676516524e-06, Gradient Norm: 0.14400610327720642)
52427
+ Step... (47475 | Loss: 0.006660703103989363, Learning Rate: 5.1030337999691255e-06, Gradient Norm: 0.6412332057952881)
52428
+ Step... (47500 | Loss: 0.005310621112585068, Learning Rate: 5.052524556958815e-06, Gradient Norm: 0.1291334629058838)
52429
+ Step... (47525 | Loss: 0.0011368023697286844, Learning Rate: 5.002021680411417e-06, Gradient Norm: 0.10793142765760422)
52430
+ Step... (47550 | Loss: 0.007397759705781937, Learning Rate: 4.9515128921484575e-06, Gradient Norm: 0.17516382038593292)
52431
+ Step... (47575 | Loss: 0.007730554323643446, Learning Rate: 4.9010095608537085e-06, Gradient Norm: 0.5665212273597717)
52432
+ Step... (47600 | Loss: 0.010844968259334564, Learning Rate: 4.8505066843063105e-06, Gradient Norm: 0.20311814546585083)
52433
+ Step... (47625 | Loss: 0.002866756869480014, Learning Rate: 4.799997896043351e-06, Gradient Norm: 0.19541442394256592)
52434
+ Step... (47650 | Loss: 0.004438326694071293, Learning Rate: 4.749494564748602e-06, Gradient Norm: 0.16690881550312042)
52435
+ Step... (47675 | Loss: 0.0008846035343594849, Learning Rate: 4.698991688201204e-06, Gradient Norm: 0.08671644330024719)
52436
+ Step... (47700 | Loss: 0.003090446349233389, Learning Rate: 4.648482899938244e-06, Gradient Norm: 0.30508506298065186)
52437
+ Step... (47725 | Loss: 0.0011106578167527914, Learning Rate: 4.5979795686434954e-06, Gradient Norm: 0.1418560892343521)
52438
+ Step... (47750 | Loss: 0.004936729092150927, Learning Rate: 4.547476692096097e-06, Gradient Norm: 0.16554218530654907)
52439
+ Step... (47775 | Loss: 0.002189367078244686, Learning Rate: 4.496967903833138e-06, Gradient Norm: 0.7181901335716248)
52440
+ Step... (47800 | Loss: 0.005074391607195139, Learning Rate: 4.446464572538389e-06, Gradient Norm: 0.15904028713703156)
52441
+ Step... (47825 | Loss: 0.0004910172428935766, Learning Rate: 4.395961695990991e-06, Gradient Norm: 0.026046399027109146)
52442
+ Step... (47850 | Loss: 0.0034627090208232403, Learning Rate: 4.34545245298068e-06, Gradient Norm: 0.12519370019435883)
52443
+ Step... (47875 | Loss: 0.00038485752884298563, Learning Rate: 4.294949576433282e-06, Gradient Norm: 0.021067677065730095)
52444
+ Step... (47900 | Loss: 0.0040473295375704765, Learning Rate: 4.244446699885884e-06, Gradient Norm: 0.14678457379341125)
52445
+ Step... (47925 | Loss: 0.0014786692336201668, Learning Rate: 4.193937456875574e-06, Gradient Norm: 0.1168798878788948)
52446
+ Step... (47950 | Loss: 0.0030647104140371084, Learning Rate: 4.143434580328176e-06, Gradient Norm: 0.1271444410085678)
52447
+ Step... (47975 | Loss: 0.0010495752794668078, Learning Rate: 4.092931703780778e-06, Gradient Norm: 0.09029767662286758)
52448
+ Step... (48000 | Loss: 0.0059296865947544575, Learning Rate: 4.042422460770467e-06, Gradient Norm: 0.14342369139194489)
52449
+ Step... (48025 | Loss: 0.0016038187313824892, Learning Rate: 3.991919584223069e-06, Gradient Norm: 0.10601940006017685)
52450
+ Step... (48050 | Loss: 0.0042427415028214455, Learning Rate: 3.941416707675671e-06, Gradient Norm: 0.14081290364265442)
52451
+ Step... (48075 | Loss: 0.0007257165852934122, Learning Rate: 3.890907464665361e-06, Gradient Norm: 0.04569496214389801)
52452
+ Step... (48100 | Loss: 0.0033516022376716137, Learning Rate: 3.840404588117963e-06, Gradient Norm: 0.13442592322826385)
52453
+ Step... (48125 | Loss: 0.002494408516213298, Learning Rate: 3.7899017115705647e-06, Gradient Norm: 0.1595575213432312)
52454
+ Step... (48150 | Loss: 0.006877778097987175, Learning Rate: 3.7393926959339296e-06, Gradient Norm: 0.22835040092468262)
52455
+ Step... (48175 | Loss: 0.0006659498903900385, Learning Rate: 3.688889592012856e-06, Gradient Norm: 0.059616561979055405)
52456
+ Step... (48200 | Loss: 0.005937668960541487, Learning Rate: 3.638386715465458e-06, Gradient Norm: 0.1822139024734497)
52457
+ Step... (48225 | Loss: 0.00040371695649810135, Learning Rate: 3.587877699828823e-06, Gradient Norm: 0.022684333845973015)
52458
+ Step... (48250 | Loss: 0.007745903916656971, Learning Rate: 3.5373745959077496e-06, Gradient Norm: 0.24524857103824615)
52459
+ Step... (48275 | Loss: 0.0005150886136107147, Learning Rate: 3.4868717193603516e-06, Gradient Norm: 0.06518520414829254)
52460
+ Step... (48300 | Loss: 0.004047275520861149, Learning Rate: 3.4363627037237165e-06, Gradient Norm: 0.21056640148162842)
52461
+
52462
+
52463
+
52464
+
52465
+
52466
+
52467
+
52468
+
52469
+
52470
+
52471
+
52472
+
52473
+
52474
+
52475
+
52476
+
52477
+
52478
+
52479
+
52480
+
52481
+
52482
+
52483
+
52484
+
52485
+
52486
+
52487
+
52488
+
52489
+
52490
+
52491
+
52492
+
52493
+
52494
+
52495
+
52496
+
52497
+
52498
+
52499
+
52500
+
52501
+
52502
+
52503
+
52504
+
52505
+
52506
+
52507
+
52508
+
52509
+
52510
+
52511
+
52512
+
52513
+
52514
+
52515
+
52516
+
52517
+
52518
+
52519
+
52520
+
52521
+
52522
+
52523
+
52524
+
52525
+
52526
+
52527
+
52528
+
52529
+
52530
+
52531
+
52532
+
52533
+
52534
+
52535
+
52536
+
52537
+
52538
+
52539
+
52540
+
52541
+
52542
+
52543
+
52544
+
52545
+
52546
+
52547
+
52548
+
52549
+
52550
+
52551
+
52552
+
52553
+
52554
+
52555
+
52556
+
52557
+
52558
+
52559
+
52560
+
52561
+
52562
+
52563
+
52564
+
52565
+
52566
+
52567
+
52568
+
52569
+
52570
+
52571
+
52572
+
52573
+
52574
+
52575
+
52576
+
52577
+
52578
+
52579
+
52580
+
52581
+
52582
+
52583
+
52584
+
52585
+
52586
+
52587
+
52588
+
52589
+
52590
+
52591
+
52592
+
52593
+
52594
+
52595
+
52596
+
52597
+
52598
+
52599
+
52600
+
52601
+
52602
+
52603
+
52604
+
52605
+
52606
+
52607
+
52608
+
52609
+
52610
+
52611
+
52612
+
52613
+
52614
+
52615
+
52616
+
52617
+
52618
+
52619
+
52620
+
52621
+
52622
+
52623
+
52624
+
52625
+
52626
+
52627
+
52628
+
52629
+
52630
+
52631
+
52632
+
52633
+
52634
+
52635
+
52636
+
52637
+
52638
+
52639
+
52640
+
52641
+
52642
+
52643
+
52644
+
52645
+
52646
+
52647
+
52648
+
52649
+
52650
+
52651
+
52652
+
52653
+
52654
+
52655
+
52656
+
52657
+
52658
+
52659
+
52660
+
52661
+
52662
+
52663
+
52664
+
52665
+
52666
+
52667
+
52668
+
52669
+
52670
+
52671
+
52672
+
52673
+
52674
+
52675
+
52676
+
52677
+
52678
+
52679
+
52680
+
52681
+
52682
+
52683
+
52684
+
52685
+
52686
+
52687
+
52688
+
52689
+
52690
+
52691
+
52692
+
52693
+
52694
+
52695
+
52696
+
52697
+
52698
+
52699
+
52700
+
52701
+
52702
+
52703
+
52704
+
52705
+
52706
+
52707
+
52708
+
52709
+
52710
+
52711
+
52712
+
52713
+
52714
+
52715
+
52716
+
52717
+
52718
+
52719
+
52720
+
52721
+
52722
+
52723
+
52724
+
52725
+
52726
+
52727
+
52728
+
52729
+
52730
+
52731
+
52732
+
52733
+
52734
+
52735
+
52736
+
52737
+
52738
+
52739
+
52740
+
52741
+
52742
+
52743
+
52744
+
52745
+
52746
+
52747
+
52748
+
52749
+
52750
+
52751
+
52752
+
52753
+
52754
+
52755
+
52756
+
52757
+
52758
+
52759
+
52760
+
52761
+
52762
+
52763
+
52764
+
52765
+
52766
+
52767
+
52768
+
52769
+
52770
+
52771
+
52772
+
52773
+
52774
+
52775
+
52776
+
52777
+
52778
+
52779
+
52780
+
52781
+
52782
+
52783
+
52784
+
52785
+
52786
+
52787
+
52788
+
52789
+
52790
+
52791
+
52792
+
52793
+
52794
+
52795
+
52796
+
52797
+
52798
+
52799
+
52800
+
52801
+
52802
+
52803
+
52804
+
52805
+
52806
+
52807
+
52808
+
52809
+
52810
+
52811
+
52812
+
52813
+
52814
+
52815
+
52816
+
52817
+
52818
+
52819
+
52820
+
52821
+
52822
+
52823
+
52824
+
52825
+
52826
+
52827
+
52828
+
52829
+
52830
+
52831
+
52832
+
52833
+
52834
+
52835
+
52836
+
52837
+
52838
+
52839
+
52840
+
52841
+
52842
+
52843
+
52844
+
52845
+
52846
+
52847
+
52848
+
52849
+
52850
+
52851
+
52852
+
52853
+
52854
+
52855
+
52856
+
52857
+
52858
+
52859
+
52860
+
52861
+
52862
+
52863
+
52864
+
52865
+
52866
+
52867
+
52868
+
52869
+
52870
+
52871
+
52872
+
52873
+
52874
+
52875
+
52876
+
52877
+
52878
+
52879
+
52880
+
52881
+
52882
+
52883
+
52884
+
52885
+
52886
+
52887
+
52888
+
52889
+
52890
+
52891
+
52892
+
52893
+
52894
+
52895
+
52896
+
52897
+
52898
+
52899
+
52900
+
52901
+
52902
+
52903
+
52904
+
52905
+
52906
+
52907
+
52908
+
52909
+
52910
+
52911
+
52912
+
52913
+
52914
+
52915
+
52916
+
52917
+
52918
+
52919
+
52920
+
52921
+
52922
+
52923
+
52924
+
52925
+
52926
+
52927
+
52928
+
52929
+
52930
+
52931
+
52932
+
52933
+
52934
+
52935
+
52936
+
52937
+
52938
+
52939
+
52940
+
52941
+
52942
+
52943
+
52944
+
52945
+
52946
+
52947
+
52948
+
52949
+
52950
+
52951
+
52952
+
52953
+
52954
+
52955
+
52956
+
52957
+
52958
+
52959
+
52960
+
52961
+
52962
+
52963
+
52964
+
52965
+
52966
+
52967
+
52968
+
52969
+
52970
+
52971
+
52972
+
52973
+
52974
+
52975
+
52976
+
52977
+
52978
+
52979
+
52980
+
52981
+
52982
+
52983
+
52984
+
52985
+
52986
+
52987
+
52988
+
52989
+
52990
+
52991
+
52992
+
52993
+
52994
+
52995
+
52996
+
52997
+
52998
+
52999
+
53000
+
53001
+
53002
+
53003
+
53004
+
53005
+
53006
+
53007
+
53008
+
53009
+
53010
+
53011
+
53012
+
53013
+
53014
+
53015
+
53016
+
53017
+
53018
+
53019
+
53020
+
53021
+
53022
+
53023
+
53024
+
53025
+
53026
+
53027
+
53028
+
53029
+
53030
+
53031
+
53032
+
53033
+
53034
+
53035
+
53036
+
53037
+
53038
+
53039
+
53040
+
53041
+
53042
+
53043
+
53044
+
53045
+
53046
+
53047
+
53048
+
53049
+
53050
+
53051
+
53052
+
53053
+
53054
+
53055
+
53056
+
53057
+
53058
+
53059
+
53060
+
53061
+
53062
+
53063
+
53064
+
53065
+
53066
+
53067
+
53068
+
53069
+
53070
+
53071
+
53072
+
53073
+
53074
+
53075
+
53076
+
53077
+
53078
+
53079
+
53080
+
53081
+
53082
+
53083
+
53084
+
53085
+
53086
+
53087
+
53088
+
53089
+
53090
+
53091
+
53092
+
53093
+
53094
+
53095
+
53096
+
53097
+
53098
+
53099
+
53100
+
53101
+
53102
+
53103
+
53104
+
53105
+
53106
+
53107
+
53108
+
53109
+
53110
+
53111
+
53112
+
53113
+
53114
+
53115
+
53116
+
53117
+
53118
+
53119
+
53120
+
53121
+
53122
+
53123
+
53124
+
53125
+
53126
+
53127
+
53128
+
53129
+
53130
+
53131
+
53132
+
53133
+
53134
+
53135
+
53136
+
53137
+
53138
+
53139
+
53140
+
53141
+
53142
+
53143
+
53144
+
53145
+
53146
+
53147
+
53148
+
53149
+
53150
+
53151
+
53152
+
53153
+
53154
+
53155
+
53156
+
53157
+
53158
+
53159
+
53160
+
53161
+
53162
+
53163
+
53164
+
53165
+
53166
+
53167
+
53168
+
53169
+
53170
+
53171
+
53172
+
53173
+
53174
+
53175
+
53176
+
53177
+
53178
+
53179
+
53180
+
53181
+
53182
+
53183
+
53184
+
53185
+
53186
+
53187
+
53188
+
53189
+
53190
+
53191
+
53192
+
53193
+
53194
+
53195
+
53196
+
53197
+
53198
+
53199
+
53200
+
53201
+
53202
+
53203
+
53204
+
53205
+
53206
+
53207
+
53208
+
53209
+
53210
+
53211
+
53212
+
53213
+
53214
+
53215
+
53216
+
53217
+
53218
+
53219
+
53220
+
53221
+
53222
+
53223
+
53224
+
53225
+
53226
+
53227
+
53228
+
53229
+
53230
+
53231
+
53232
+
53233
+
53234
+
53235
+
53236
+
53237
+
53238
+
53239
+
53240
+
53241
+
53242
+
53243
+
53244
+
53245
+
53246
+
53247
+
53248
+
53249
+
53250
+
53251
+
53252
+
53253
+
53254
+
53255
+
53256
+
53257
+
53258
+
53259
+
53260
+
53261
+
53262
+
53263
+
53264
+
53265
+
53266
+
53267
+
53268
+
53269
+
53270
+
53271
+
53272
+
53273
+
53274
+
53275
+
53276
+
53277
+
53278
+
53279
+
53280
+
53281
+
53282
+
53283
+
53284
+
53285
+
53286
+
53287
+
53288
+
53289
+
53290
+
53291
+
53292
+
53293
+
53294
+
53295
+
53296
+
53297
+
53298
+
53299
+
53300
+
53301
+
53302
+
53303
+
53304
+
53305
+
53306
+
53307
+
53308
+
53309
+
53310
+
53311
+
53312
+
53313
+
53314
+
53315
+
53316
+
53317
+
53318
+
53319
+
53320
+
53321
+
53322
+
53323
+
53324
+
53325
+
53326
+
53327
+
53328
+
53329
+
53330
+
53331
+
53332
+
53333
+
53334
+
53335
+
53336
+
53337
+
53338
+
53339
+
53340
+
53341
+
53342
+
53343
+
53344
+
53345
+
53346
+
53347
+
53348
+
53349
+
53350
+
53351
+
53352
+
53353
+
53354
+
53355
+
53356
+
53357
+
53358
+
53359
+
53360
+
53361
+
53362
+
53363
+
53364
+
53365
+
53366
+
53367
+
53368
+
53369
+
53370
+
53371
+
53372
+
53373
+
53374
+
53375
+
53376
+
53377
+
53378
+
53379
+
53380
+
53381
+
53382
+
53383
+
53384
+
53385
+
53386
+
53387
+
53388
+
53389
+
53390
+
53391
+
53392
+
53393
+
53394
+
53395
+
53396
+
53397
+
53398
+
53399
+
53400
+
53401
+
53402
+
53403
+
53404
+
53405
+
53406
+
53407
+
53408
+
53409
+
53410
+
53411
+
53412
+
53413
+
53414
+
53415
+
53416
+
53417
+
53418
+
53419
+
53420
+
53421
+
53422
+
53423
+
53424
+
53425
+
53426
+
53427
+
53428
+
53429
+
53430
+
53431
+
53432
+
53433
+
53434
+
53435
+
53436
+
53437
+
53438
+
53439
+
53440
+
53441
+
53442
+
53443
+
53444
+
53445
+
53446
+
53447
+
53448
+
53449
+
53450
+
53451
+
53452
+
53453
+
53454
+
53455
+
53456
+
53457
+
53458
+
53459
+
53460
+
53461
+
53462
+
53463
+
53464
+
53465
+
53466
+
53467
+
53468
+
53469
+
53470
+
53471
+
53472
+
53473
+
53474
+
53475
+
53476
+
53477
+
53478
+
53479
+
53480
+
53481
+
53482
+
53483
+
53484
+
53485
+
53486
+
53487
+
53488
+
53489
+
53490
+
53491
+
53492
+
53493
+
53494
+
53495
+
53496
+
53497
+
53498
+
53499
+
53500
+
53501
+
53502
+
53503
+
53504
+
53505
+
53506
+
53507
+
53508
+
53509
+
53510
+
53511
+
53512
+
53513
+
53514
+
53515
+
53516
+
53517
+
53518
+
53519
+
53520
+
53521
+
53522
+
53523
+
53524
+
53525
+
53526
+
53527
+
53528
+
53529
+
53530
+
53531
+
53532
+
53533
+
53534
+
53535
+
53536
+
53537
+
53538
+
53539
+
53540
+
53541
+
53542
+
53543
+
53544
+
53545
+
53546
+
53547
+
53548
+
53549
+
53550
+
53551
+
53552
+
53553
+
53554
+
53555
+
53556
+
53557
+
53558
+
53559
+
53560
+
53561
+
53562
+
53563
+
53564
+
53565
+
53566
+
53567
+
53568
+
53569
+
53570
+
53571
+
53572
+
53573
+
53574
+
53575
+
53576
+
53577
+
53578
+
53579
+
53580
+
53581
+
53582
+
53583
+
53584
+
53585
+
53586
+
53587
+
53588
+
53589
+
53590
+
53591
+
53592
+
53593
+
53594
+
53595
+
53596
+
53597
+
53598
+
53599
+
53600
+
53601
+
53602
+
53603
+
53604
+
53605
+
53606
+
53607
+
53608
+
53609
+
53610
+
53611
+
53612
+
53613
+
53614
+
53615
+
53616
+
53617
+
53618
+
53619
+
53620
+
53621
+
53622
+
53623
+
53624
+
53625
+
53626
+
53627
+
53628
+
53629
+
53630
+
53631
+
53632
+
53633
+
53634
+
53635
+
53636
+
53637
+
53638
+
53639
+
53640
+
53641
+
53642
+
53643
+
53644
+
53645
+
53646
+
53647
+
53648
+
53649
+
53650
+
53651
+
53652
+
53653
+
53654
+
53655
+
53656
+
53657
+
53658
+
53659
+
53660
+
53661
+
53662
+
53663
+
53664
+
53665
+
53666
+
53667
+
53668
+
53669
+
53670
+
53671
+
53672
+
53673
+
53674
+
53675
+
53676
+
53677
+
53678
+
53679
+
53680
+
53681
+
53682
+
53683
+
53684
+
53685
+
53686
+
53687
+
53688
+
53689
+
53690
+
53691
+
53692
+
53693
+
53694
+
53695
+
53696
+
53697
+
53698
+
53699
+
53700
+
53701
+
53702
+
53703
+
53704
+
53705
+
53706
+
53707
+
53708
+
53709
+
53710
+
53711
+
53712
+
53713
+
53714
+
53715
+
53716
+
53717
+
53718
+
53719
+
53720
+
53721
+
53722
+
53723
+
53724
+
53725
+
53726
+
53727
+
53728
+
53729
+
53730
+
53731
+
53732
+
53733
+
53734
+
53735
+
53736
+
53737
+
53738
+
53739
+
53740
+
53741
+
53742
+
53743
+
53744
+
53745
+
53746
+
53747
+
53748
+
53749
+
53750
+
53751
+
53752
+
53753
+
53754
+
53755
+
53756
+
53757
+
53758
+
53759
+
53760
+
53761
+
53762
+
53763
+
53764
+
53765
+
53766
+
53767
+
53768
+
53769
+
53770
+
53771
+
53772
+
53773
+
53774
+
53775
+
53776
+
53777
+
53778
+
53779
+
53780
+
53781
+
53782
+
53783
+
53784
+
53785
+
53786
+
53787
+
53788
+
53789
+
53790
+
53791
+
53792
+
53793
+
53794
+
53795
+
53796
+
53797
+
53798
+
53799
+
53800
+
53801
+
53802
+
53803
+
53804
+
53805
+
53806
+
53807
+
53808
+
53809
+
53810
+
53811
+
53812
+
53813
+
53814
+
53815
+
53816
+
53817
+
53818
+
53819
+
53820
+
53821
+
53822
+
53823
+
53824
+
53825
+
53826
+
53827
+
53828
+
53829
+
53830
+
53831
+
53832
+
53833
+
53834
+
53835
+
53836
+
53837
+
53838
+
53839
+
53840
+
53841
+
53842
+
53843
+
53844
+
53845
+
53846
+
53847
+
53848
+
53849
+
53850
+
53851
+
53852
+
53853
+
53854
+
53855
+
53856
+
53857
+
53858
+
53859
+
53860
+
53861
+
53862
+
53863
+
53864
+
53865
+
53866
+
53867
+
53868
+
53869
+
53870
+
53871
+
53872
+
53873
+
53874
+
53875
+
53876
+
53877
+
53878
+
53879
+
53880
+
53881
+
53882
+
53883
+
53884
+
53885
+
53886
+
53887
+
53888
+
53889
+
53890
+
53891
+
53892
+
53893
+
53894
+
53895
+
53896
+
53897
+
53898
+
53899
+
53900
+
53901
+
53902
+
53903
+
53904
+
53905
+
53906
+
53907
+
53908
+
53909
+
53910
+
53911
+
53912
+
53913
+
53914
+
53915
+
53916
+
53917
+
53918
+
53919
+
53920
+
53921
+
53922
+
53923
+
53924
+
53925
+
53926
+
53927
+
53928
+
53929
+
53930
+
53931
+
53932
+
53933
+
53934
+
53935
+
53936
+
53937
+
53938
+
53939
+
53940
+
53941
+
53942
+
53943
+
53944
+
53945
+
53946
+
53947
+
53948
+
53949
+
53950
+
53951
+
53952
+
53953
+
53954
+
53955
+
53956
+
53957
+
53958
+
53959
+
53960
+
53961
+
53962
+
53963
+
53964
+
53965
+
53966
+
53967
+
53968
+
53969
+
53970
+
53971
+
53972
+
53973
+
53974
+
53975
+
53976
+
53977
+
53978
+
53979
+
53980
+
53981
+
53982
+
53983
+
53984
+
53985
+
53986
+
53987
+
53988
+
53989
+
53990
+
53991
+
53992
+
53993
+
53994
+
53995
+
53996
+
53997
+
53998
+
53999
+
54000
+
54001
+
54002
+
54003
+
54004
+
54005
+
54006
+
54007
+
54008
+
54009
+
54010
+
54011
+
54012
+
54013
+
54014
+
54015
+
54016
+
54017
+
54018
+
54019
+
54020
+
54021
+
54022
+
54023
+
54024
+
54025
+
54026
+
54027
+
54028
+
54029
+
54030
+
54031
+
54032
+
54033
+
54034
+
54035
+
54036
+
54037
+
54038
+
54039
+
54040
+
54041
+
54042
+
54043
+
54044
+
54045
+
54046
+
54047
+
54048
+
54049
+
54050
+
54051
+
54052
+
54053
+
54054
+
54055
+
54056
+
54057
+
54058
+
54059
+
54060
+
54061
+
54062
+
54063
+
54064
+
54065
+
54066
+
54067
+
54068
+
54069
+
54070
+
54071
+
54072
+
54073
+
54074
+
54075
+
54076
+
54077
+
54078
+
54079
+
54080
+
54081
+
54082
+
54083
+
54084
+
54085
+
54086
+
54087
+
54088
+
54089
+
54090
+
54091
+
54092
+
54093
+
54094
+
54095
+
54096
+
54097
+
54098
+
54099
+
54100
+
54101
+
54102
+ Step... (40000/50000 | Eval Loss: 0.9527401328086853 | Eval wer: 0.04187346053453917 | Eval cer: 0.02872188479352137 |): 100% 12/12 [66:36:10<00:00, 19980.90s/it]
54103
+ run_flax_speech_recognition_seq2seq.py:1425: FutureWarning: jax.tree_map is deprecated, and will be removed in a future release. Use jax.tree_util.tree_map instead.
54104
+ params = jax.device_get(jax.tree_map(lambda x: x[0], state.params))
54105
+ Configuration saved in /home/sanchitgandhi/flax-wav2vec2-2-bart-large-ls-960h-black-box/config.json
54106
+ Model weights saved in /home/sanchitgandhi/flax-wav2vec2-2-bart-large-ls-960h-black-box/flax_model.msgpack
54107
+ tokenizer config file saved in ./tokenizer_config.json
54108
+ Special tokens file saved in ./special_tokens_map.json
54109
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
54110
+ To disable this warning, you can either:
54111
+ - Avoid using `tokenizers` before the fork if possible
54112
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
54113
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
54114
+ To disable this warning, you can either:
54115
+ - Avoid using `tokenizers` before the fork if possible
54116
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
54117
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
54118
+ To disable this warning, you can either:
54119
+ - Avoid using `tokenizers` before the fork if possible
54120
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
54121
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
54122
+ To disable this warning, you can either:
54123
+ - Avoid using `tokenizers` before the fork if possible
54124
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
54125
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
54126
+ To disable this warning, you can either:
54127
+ - Avoid using `tokenizers` before the fork if possible
54128
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
54129
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
54130
+ To disable this warning, you can either:
54131
+ - Avoid using `tokenizers` before the fork if possible
54132
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
54133
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
54134
  To disable this warning, you can either:
54135
  - Avoid using `tokenizers` before the fork if possible
wandb/run-20220828_085247-2hx8pk65/files/wandb-summary.json CHANGED
@@ -1 +1 @@
1
- {"train/decoder_grad_norm": 0.18293321132659912, "train/decoder_param_norm": 1063.0654296875, "train/encoder_grad_norm": 0.148331880569458, "train/encoder_param_norm": 2323.336669921875, "train/grad_norm": 0.23551413416862488, "layer_grad_norm/": {"decoder": {"model": {"decoder": {"embed_positions": {"embedding": 0.008046639151871204}, "embed_tokens": {"embedding": 0.060666970908641815}, "layernorm_embedding": {"bias": 0.003096886444836855, "scale": 0.0024350089952349663}, "layers": {"FlaxBartDecoderLayers": {"encoder_attn": {"k_proj": {"bias": 5.256703389022732e-06, "kernel": 0.011271456256508827}, "out_proj": {"bias": 0.007880721241235733, "kernel": 0.03872065246105194}, "q_proj": {"bias": 0.0004976371419616044, "kernel": 0.01105893962085247}, "v_proj": {"bias": 0.015226359479129314, "kernel": 0.030586158856749535}}, "encoder_attn_layer_norm": {"bias": 0.01159473042935133, "scale": 0.012393548153340816}, "fc1": {"bias": 0.004439335782080889, "kernel": 0.10113218426704407}, "fc2": {"bias": 0.011387723498046398, "kernel": 0.09879624843597412}, "final_layer_norm": {"bias": 0.026725012809038162, "scale": 0.026244478300213814}, "self_attn": {"k_proj": {"bias": 1.8891130366682773e-06, "kernel": 0.010783703066408634}, "out_proj": {"bias": 0.016153542324900627, "kernel": 0.039619628340005875}, "q_proj": {"bias": 0.0008692306582815945, "kernel": 0.01054653525352478}, "v_proj": {"bias": 0.01716933213174343, "kernel": 0.04919762164354324}}, "self_attn_layer_norm": {"bias": 0.007215098012238741, "scale": 0.008393414318561554}}}}}}, "encoder": {"adapter": {"layers": {"0": {"conv": {"bias": 0.022141050547361374, "kernel": 0.049536462873220444}}, "1": {"conv": {"bias": 0.017486272379755974, "kernel": 0.03386249393224716}}, "2": {"conv": {"bias": 0.020221581682562828, "kernel": 0.05212334543466568}}}}, "encoder": {"layer_norm": {"bias": 0.07752241939306259, "scale": 0.03339512273669243}, "layers": {"FlaxWav2Vec2EncoderLayers": {"attention": {"k_proj": {"bias": 1.787670839803468e-06, "kernel": 0.016130739822983742}, "out_proj": {"bias": 0.001517767203040421, "kernel": 0.0303343553096056}, "q_proj": {"bias": 0.002082030288875103, "kernel": 0.01582256518304348}, "v_proj": {"bias": 0.00666964752599597, "kernel": 0.027823949232697487}}, "feed_forward": {"intermediate_dense": {"bias": 0.002947209170088172, "kernel": 0.03786110132932663}, "output_dense": {"bias": 0.0014471356989815831, "kernel": 0.03480111435055733}}, "final_layer_norm": {"bias": 0.014765393920242786, "scale": 0.01634989120066166}, "layer_norm": {"bias": 0.03281298279762268, "scale": 0.024696579203009605}}}, "pos_conv_embed": {"conv": {"bias": 0.0003536914009600878, "weight_g": 0.0017076540971174836, "weight_v": 0.0059150331653654575}}}, "feature_extractor": {"conv_layers": {"0": {"conv": {"bias": 0.0, "kernel": 0.0}, "layer_norm": {"bias": 0.0, "scale": 0.0}}, "1": {"conv": {"bias": 0.0, "kernel": 0.0}, "layer_norm": {"bias": 0.0, "scale": 0.0}}, "2": {"conv": {"bias": 0.0, "kernel": 0.0}, "layer_norm": {"bias": 0.0, "scale": 0.0}}, "3": {"conv": {"bias": 0.0, "kernel": 0.0}, "layer_norm": {"bias": 0.0, "scale": 0.0}}, "4": {"conv": {"bias": 0.0, "kernel": 0.0}, "layer_norm": {"bias": 0.0, "scale": 0.0}}, "5": {"conv": {"bias": 0.0, "kernel": 0.0}, "layer_norm": {"bias": 0.0, "scale": 0.0}}, "6": {"conv": {"bias": 0.0, "kernel": 0.0}, "layer_norm": {"bias": 0.0, "scale": 0.0}}}}, "feature_projection": {"layer_norm": {"bias": 0.00199363986030221, "scale": 0.0019663891289383173}, "projection": {"bias": 0.000548368610907346, "kernel": 0.015020000748336315}}, "masked_spec_embed": 0.0}}, "layer_param_norm/": {"decoder": {"model": {"decoder": {"embed_positions": {"embedding": 58.655296325683594}, "embed_tokens": {"embedding": 628.4320678710938}, "layernorm_embedding": {"bias": 2.423579692840576, "scale": 13.864307403564453}, "layers": {"FlaxBartDecoderLayers": {"encoder_attn": {"k_proj": {"bias": 47.95212173461914, "kernel": 331.52020263671875}, "out_proj": {"bias": 6.132379531860352, "kernel": 227.88941955566406}, "q_proj": {"bias": 20.880966186523438, "kernel": 338.4452819824219}, "v_proj": {"bias": 3.570829391479492, "kernel": 232.0838623046875}}, "encoder_attn_layer_norm": {"bias": 11.024534225463867, "scale": 57.68368148803711}, "fc1": {"bias": 26.054283142089844, "kernel": 346.79296875}, "fc2": {"bias": 7.841187953948975, "kernel": 248.94561767578125}, "final_layer_norm": {"bias": 3.914222002029419, "scale": 63.57343673706055}, "self_attn": {"k_proj": {"bias": 59.552547454833984, "kernel": 280.0233459472656}, "out_proj": {"bias": 3.7681872844696045, "kernel": 132.86544799804688}, "q_proj": {"bias": 32.200714111328125, "kernel": 283.1105651855469}, "v_proj": {"bias": 2.561887741088867, "kernel": 141.2297821044922}}, "self_attn_layer_norm": {"bias": 8.89633846282959, "scale": 84.71704864501953}}}}}}, "encoder": {"adapter": {"layers": {"0": {"conv": {"bias": 1.1991273164749146, "kernel": 62.47041320800781}}, "1": {"conv": {"bias": 1.3806315660476685, "kernel": 60.54581069946289}}, "2": {"conv": {"bias": 1.5864932537078857, "kernel": 60.361602783203125}}}}, "encoder": {"layer_norm": {"bias": 0.2923617959022522, "scale": 4.219600677490234}, "layers": {"FlaxWav2Vec2EncoderLayers": {"attention": {"k_proj": {"bias": 19.385311126708984, "kernel": 553.5303955078125}, "out_proj": {"bias": 16.84536361694336, "kernel": 705.300537109375}, "q_proj": {"bias": 40.93048858642578, "kernel": 546.0474243164062}, "v_proj": {"bias": 15.559945106506348, "kernel": 696.8070068359375}}, "feed_forward": {"intermediate_dense": {"bias": 24.44249725341797, "kernel": 1377.5845947265625}, "output_dense": {"bias": 20.795345306396484, "kernel": 1303.997314453125}}, "final_layer_norm": {"bias": 32.52366256713867, "scale": 142.01145935058594}, "layer_norm": {"bias": 7.272017478942871, "scale": 45.770084381103516}}}, "pos_conv_embed": {"conv": {"bias": 15.211658477783203, "weight_g": 21.0489501953125, "weight_v": 214.11346435546875}}}, "feature_extractor": {"conv_layers": {"0": {"conv": {"bias": 0.5982058644294739, "kernel": 8.08896541595459}, "layer_norm": {"bias": 10.069783210754395, "scale": 10.451257705688477}}, "1": {"conv": {"bias": 4.74075174331665, "kernel": 90.8435287475586}, "layer_norm": {"bias": 6.922820091247559, "scale": 19.5467586517334}}, "2": {"conv": {"bias": 6.7732415199279785, "kernel": 146.13897705078125}, "layer_norm": {"bias": 9.044225692749023, "scale": 19.424888610839844}}, "3": {"conv": {"bias": 5.224758148193359, "kernel": 159.10508728027344}, "layer_norm": {"bias": 8.319666862487793, "scale": 17.64743423461914}}, "4": {"conv": {"bias": 4.434978008270264, "kernel": 157.35813903808594}, "layer_norm": {"bias": 9.193974494934082, "scale": 15.562357902526855}}, "5": {"conv": {"bias": 5.297643661499023, "kernel": 131.1835174560547}, "layer_norm": {"bias": 10.735219955444336, "scale": 13.812533378601074}}, "6": {"conv": {"bias": 5.615579128265381, "kernel": 136.41822814941406}, "layer_norm": {"bias": 12.515308380126953, "scale": 11.152680397033691}}}}, "feature_projection": {"layer_norm": {"bias": 9.234997749328613, "scale": 27.593135833740234}, "projection": {"bias": 4.323590278625488, "kernel": 88.11954498291016}}, "masked_spec_embed": 26.247730255126953}}, "train/learning_rate": 2.0254545233910903e-05, "train/loss": 0.008378183469176292, "train/param_norm": 2554.995361328125, "_timestamp": 1661871981, "_runtime": 195214, "_step": 39975, "eval/loss": 0.9876799583435059, "eval/wer": 0.04759016212639241, "eval/cer": 0.035079873533571844, "eval/step_10k": {"_type": "table-file", "sha256": "8b44e8a00a036a18ffdf81b4d076c8bf849ea6649001c69e94fa439b14f110ee", "size": 26434, "artifact_path": "wandb-client-artifact://18m0dj4hts3yiat04x5pvmncavkjapd5wb8bznb37vw8c0lqna3m2yjd1wtdrfstuoo7ejt2sphvjo0zuw1e5ne5d3qbkd7c1fylclfggig6us5tsmsj2uum5pchx48n:latest/eval/step_10k.table.json", "_latest_artifact_path": "wandb-client-artifact://18m0dj4hts3yiat04x5pvmncavkjapd5wb8bznb37vw8c0lqna3m2yjd1wtdrfstuoo7ejt2sphvjo0zuw1e5ne5d3qbkd7c1fylclfggig6us5tsmsj2uum5pchx48n:latest/eval/step_10k.table.json", "path": "media/table/eval/step_10k_10000_8b44e8a00a036a18ffdf.table.json", "ncols": 7, "nrows": 50}, "eval/step_20k": {"_type": "table-file", "sha256": "a0a50c5d8793ca99e4646f70c3624f8742c5285825bc1c59ab4083ac4de9d6e3", "size": 26657, "artifact_path": "wandb-client-artifact://13ri9hnxp93kf7dsdol2hs1j0v7bpkwwvujpi27awdck0fjm6vfog0dun9k9toif5xrt3cijlotddakikiw0bnbo3go679b4d2spq9c0w865vq0k9auiszkkbvev62fc:latest/eval/step_20k.table.json", "_latest_artifact_path": "wandb-client-artifact://13ri9hnxp93kf7dsdol2hs1j0v7bpkwwvujpi27awdck0fjm6vfog0dun9k9toif5xrt3cijlotddakikiw0bnbo3go679b4d2spq9c0w865vq0k9auiszkkbvev62fc:latest/eval/step_20k.table.json", "path": "media/table/eval/step_20k_20000_a0a50c5d8793ca99e464.table.json", "ncols": 7, "nrows": 50}, "eval/step_30k": {"_type": "table-file", "sha256": "509ad8614e16ae2800f100e82fd40356d3559e0425cf50e8e19c3155dc6c604e", "size": 26610, "artifact_path": "wandb-client-artifact://1v7mr89p2ldfoyyvrj6x6g0s8qpvgnioclm4puk5i0wphipeujkyna0xf7y63nt4a5qivkdzytt4lzhwuraoe53a5683ln1yzntyw9kh7vhcxz7oci70rpw62o1tz6ho:latest/eval/step_30k.table.json", "_latest_artifact_path": "wandb-client-artifact://1v7mr89p2ldfoyyvrj6x6g0s8qpvgnioclm4puk5i0wphipeujkyna0xf7y63nt4a5qivkdzytt4lzhwuraoe53a5683ln1yzntyw9kh7vhcxz7oci70rpw62o1tz6ho:latest/eval/step_30k.table.json", "path": "media/table/eval/step_30k_30000_509ad8614e16ae2800f1.table.json", "ncols": 7, "nrows": 50}}
 
1
+ {"train/decoder_grad_norm": 0.07715553045272827, "train/decoder_param_norm": 1063.1578369140625, "train/encoder_grad_norm": 0.07033974677324295, "train/encoder_param_norm": 2323.486572265625, "train/grad_norm": 0.1044062077999115, "layer_grad_norm/": {"decoder": {"model": {"decoder": {"embed_positions": {"embedding": 0.005401859525591135}, "embed_tokens": {"embedding": 0.03052273392677307}, "layernorm_embedding": {"bias": 0.0020871469751000404, "scale": 0.0012841655407100916}, "layers": {"FlaxBartDecoderLayers": {"encoder_attn": {"k_proj": {"bias": 1.8833308104149182e-06, "kernel": 0.004420427139848471}, "out_proj": {"bias": 0.00347402086481452, "kernel": 0.015828590840101242}, "q_proj": {"bias": 0.00020214646065142006, "kernel": 0.004523565527051687}, "v_proj": {"bias": 0.006682044826447964, "kernel": 0.012984688393771648}}, "encoder_attn_layer_norm": {"bias": 0.004988634027540684, "scale": 0.005556626245379448}, "fc1": {"bias": 0.0017050342867150903, "kernel": 0.04050837457180023}, "fc2": {"bias": 0.005049724597483873, "kernel": 0.03927548602223396}, "final_layer_norm": {"bias": 0.013566207140684128, "scale": 0.008338828571140766}, "self_attn": {"k_proj": {"bias": 5.705806529476831e-07, "kernel": 0.003607015358284116}, "out_proj": {"bias": 0.008329739794135094, "kernel": 0.015504402108490467}, "q_proj": {"bias": 0.0002979960117954761, "kernel": 0.003639827948063612}, "v_proj": {"bias": 0.00870799645781517, "kernel": 0.022596830502152443}}, "self_attn_layer_norm": {"bias": 0.003212399547919631, "scale": 0.00401513883844018}}}}}}, "encoder": {"adapter": {"layers": {"0": {"conv": {"bias": 0.009786142967641354, "kernel": 0.02071578986942768}}, "1": {"conv": {"bias": 0.007721441797912121, "kernel": 0.01437926385551691}}, "2": {"conv": {"bias": 0.009065583348274231, "kernel": 0.021983565762639046}}}}, "encoder": {"layer_norm": {"bias": 0.03353462368249893, "scale": 0.01526513509452343}, "layers": {"FlaxWav2Vec2EncoderLayers": {"attention": {"k_proj": {"bias": 1.1060902807003004e-06, "kernel": 0.00823200587183237}, "out_proj": {"bias": 0.0008450962486676872, "kernel": 0.01538914442062378}, "q_proj": {"bias": 0.0011472117621451616, "kernel": 0.008041738532483578}, "v_proj": {"bias": 0.0036684211809188128, "kernel": 0.014704424887895584}}, "feed_forward": {"intermediate_dense": {"bias": 0.002029425697401166, "kernel": 0.019275842234492302}, "output_dense": {"bias": 0.0007828964735381305, "kernel": 0.01704106293618679}}, "final_layer_norm": {"bias": 0.010386798530817032, "scale": 0.011868822388350964}, "layer_norm": {"bias": 0.01724540814757347, "scale": 0.015423807315528393}}}, "pos_conv_embed": {"conv": {"bias": 0.0002918594400398433, "weight_g": 0.0027034797240048647, "weight_v": 0.005031736101955175}}}, "feature_extractor": {"conv_layers": {"0": {"conv": {"bias": 0.0, "kernel": 0.0}, "layer_norm": {"bias": 0.0, "scale": 0.0}}, "1": {"conv": {"bias": 0.0, "kernel": 0.0}, "layer_norm": {"bias": 0.0, "scale": 0.0}}, "2": {"conv": {"bias": 0.0, "kernel": 0.0}, "layer_norm": {"bias": 0.0, "scale": 0.0}}, "3": {"conv": {"bias": 0.0, "kernel": 0.0}, "layer_norm": {"bias": 0.0, "scale": 0.0}}, "4": {"conv": {"bias": 0.0, "kernel": 0.0}, "layer_norm": {"bias": 0.0, "scale": 0.0}}, "5": {"conv": {"bias": 0.0, "kernel": 0.0}, "layer_norm": {"bias": 0.0, "scale": 0.0}}, "6": {"conv": {"bias": 0.0, "kernel": 0.0}, "layer_norm": {"bias": 0.0, "scale": 0.0}}}}, "feature_projection": {"layer_norm": {"bias": 0.0014273464912548661, "scale": 0.001675860257819295}, "projection": {"bias": 0.0004162571858614683, "kernel": 0.01199527457356453}}, "masked_spec_embed": 0.0}}, "layer_param_norm/": {"decoder": {"model": {"decoder": {"embed_positions": {"embedding": 58.65697479248047}, "embed_tokens": {"embedding": 628.4302978515625}, "layernorm_embedding": {"bias": 2.422621965408325, "scale": 13.860967636108398}, "layers": {"FlaxBartDecoderLayers": {"encoder_attn": {"k_proj": {"bias": 47.94939041137695, "kernel": 331.5506591796875}, "out_proj": {"bias": 6.131547927856445, "kernel": 227.9162139892578}, "q_proj": {"bias": 20.88374137878418, "kernel": 338.4700622558594}, "v_proj": {"bias": 3.56612491607666, "kernel": 232.11032104492188}}, "encoder_attn_layer_norm": {"bias": 11.035151481628418, "scale": 57.715087890625}, "fc1": {"bias": 26.05365562438965, "kernel": 346.8977355957031}, "fc2": {"bias": 7.839730739593506, "kernel": 249.0161590576172}, "final_layer_norm": {"bias": 3.914802074432373, "scale": 63.62209701538086}, "self_attn": {"k_proj": {"bias": 59.55521011352539, "kernel": 280.0371398925781}, "out_proj": {"bias": 3.7659459114074707, "kernel": 132.872802734375}, "q_proj": {"bias": 32.20103073120117, "kernel": 283.1228942871094}, "v_proj": {"bias": 2.560319423675537, "kernel": 141.23605346679688}}, "self_attn_layer_norm": {"bias": 8.895964622497559, "scale": 84.7233657836914}}}}}}, "encoder": {"adapter": {"layers": {"0": {"conv": {"bias": 1.2016971111297607, "kernel": 62.545352935791016}}, "1": {"conv": {"bias": 1.3880459070205688, "kernel": 60.6671257019043}}, "2": {"conv": {"bias": 1.5934075117111206, "kernel": 60.4710807800293}}}}, "encoder": {"layer_norm": {"bias": 0.2923099100589752, "scale": 4.224204063415527}, "layers": {"FlaxWav2Vec2EncoderLayers": {"attention": {"k_proj": {"bias": 19.386531829833984, "kernel": 553.5894165039062}, "out_proj": {"bias": 16.841169357299805, "kernel": 705.33203125}, "q_proj": {"bias": 40.93061828613281, "kernel": 546.1094970703125}, "v_proj": {"bias": 15.554323196411133, "kernel": 696.839111328125}}, "feed_forward": {"intermediate_dense": {"bias": 24.43609619140625, "kernel": 1377.650146484375}, "output_dense": {"bias": 20.79059600830078, "kernel": 1304.09033203125}}, "final_layer_norm": {"bias": 32.522762298583984, "scale": 142.02593994140625}, "layer_norm": {"bias": 7.270603656768799, "scale": 45.793148040771484}}}, "pos_conv_embed": {"conv": {"bias": 15.207948684692383, "weight_g": 21.04806900024414, "weight_v": 214.14303588867188}}}, "feature_extractor": {"conv_layers": {"0": {"conv": {"bias": 0.5982058644294739, "kernel": 8.08896541595459}, "layer_norm": {"bias": 10.069783210754395, "scale": 10.451257705688477}}, "1": {"conv": {"bias": 4.74075174331665, "kernel": 90.8435287475586}, "layer_norm": {"bias": 6.922820091247559, "scale": 19.5467586517334}}, "2": {"conv": {"bias": 6.7732415199279785, "kernel": 146.13897705078125}, "layer_norm": {"bias": 9.044225692749023, "scale": 19.424888610839844}}, "3": {"conv": {"bias": 5.224758148193359, "kernel": 159.10508728027344}, "layer_norm": {"bias": 8.319666862487793, "scale": 17.64743423461914}}, "4": {"conv": {"bias": 4.434978008270264, "kernel": 157.35813903808594}, "layer_norm": {"bias": 9.193974494934082, "scale": 15.562357902526855}}, "5": {"conv": {"bias": 5.297643661499023, "kernel": 131.1835174560547}, "layer_norm": {"bias": 10.735219955444336, "scale": 13.812533378601074}}, "6": {"conv": {"bias": 5.615579128265381, "kernel": 136.41822814941406}, "layer_norm": {"bias": 12.515308380126953, "scale": 11.152680397033691}}}}, "feature_projection": {"layer_norm": {"bias": 9.220093727111816, "scale": 27.58247947692871}, "projection": {"bias": 4.322972774505615, "kernel": 88.09733581542969}}, "masked_spec_embed": 26.247730255126953}}, "train/learning_rate": 5.252361034990827e-08, "train/loss": 0.0029563589487224817, "train/param_norm": 2555.17041015625, "_timestamp": 1661920077, "_runtime": 243310, "_step": 49975, "eval/loss": 0.9527401328086853, "eval/wer": 0.04187346053453917, "eval/cer": 0.02872188479352137, "eval/step_10k": {"_type": "table-file", "sha256": "8b44e8a00a036a18ffdf81b4d076c8bf849ea6649001c69e94fa439b14f110ee", "size": 26434, "artifact_path": "wandb-client-artifact://18m0dj4hts3yiat04x5pvmncavkjapd5wb8bznb37vw8c0lqna3m2yjd1wtdrfstuoo7ejt2sphvjo0zuw1e5ne5d3qbkd7c1fylclfggig6us5tsmsj2uum5pchx48n:latest/eval/step_10k.table.json", "_latest_artifact_path": "wandb-client-artifact://18m0dj4hts3yiat04x5pvmncavkjapd5wb8bznb37vw8c0lqna3m2yjd1wtdrfstuoo7ejt2sphvjo0zuw1e5ne5d3qbkd7c1fylclfggig6us5tsmsj2uum5pchx48n:latest/eval/step_10k.table.json", "path": "media/table/eval/step_10k_10000_8b44e8a00a036a18ffdf.table.json", "ncols": 7, "nrows": 50}, "eval/step_20k": {"_type": "table-file", "sha256": "a0a50c5d8793ca99e4646f70c3624f8742c5285825bc1c59ab4083ac4de9d6e3", "size": 26657, "artifact_path": "wandb-client-artifact://13ri9hnxp93kf7dsdol2hs1j0v7bpkwwvujpi27awdck0fjm6vfog0dun9k9toif5xrt3cijlotddakikiw0bnbo3go679b4d2spq9c0w865vq0k9auiszkkbvev62fc:latest/eval/step_20k.table.json", "_latest_artifact_path": "wandb-client-artifact://13ri9hnxp93kf7dsdol2hs1j0v7bpkwwvujpi27awdck0fjm6vfog0dun9k9toif5xrt3cijlotddakikiw0bnbo3go679b4d2spq9c0w865vq0k9auiszkkbvev62fc:latest/eval/step_20k.table.json", "path": "media/table/eval/step_20k_20000_a0a50c5d8793ca99e464.table.json", "ncols": 7, "nrows": 50}, "eval/step_30k": {"_type": "table-file", "sha256": "509ad8614e16ae2800f100e82fd40356d3559e0425cf50e8e19c3155dc6c604e", "size": 26610, "artifact_path": "wandb-client-artifact://1v7mr89p2ldfoyyvrj6x6g0s8qpvgnioclm4puk5i0wphipeujkyna0xf7y63nt4a5qivkdzytt4lzhwuraoe53a5683ln1yzntyw9kh7vhcxz7oci70rpw62o1tz6ho:latest/eval/step_30k.table.json", "_latest_artifact_path": "wandb-client-artifact://1v7mr89p2ldfoyyvrj6x6g0s8qpvgnioclm4puk5i0wphipeujkyna0xf7y63nt4a5qivkdzytt4lzhwuraoe53a5683ln1yzntyw9kh7vhcxz7oci70rpw62o1tz6ho:latest/eval/step_30k.table.json", "path": "media/table/eval/step_30k_30000_509ad8614e16ae2800f1.table.json", "ncols": 7, "nrows": 50}, "eval/step_40k": {"_type": "table-file", "sha256": "1c187b6d789cca3b4ec0da0aa76f40085364204904c4e4e7f9a7ae021659a9ab", "size": 26557, "artifact_path": "wandb-client-artifact://tbhmdixx2c1ta253dre6w8pjaxb8fbzac6u2cugh4vsvvgiykum9l788fn7omfbr8f3ir5gq1fo2cb7bwlha8musehb13eb9eb7hppcmvnyc3gg5i8eprg7w7z0jqrgu:latest/eval/step_40k.table.json", "_latest_artifact_path": "wandb-client-artifact://tbhmdixx2c1ta253dre6w8pjaxb8fbzac6u2cugh4vsvvgiykum9l788fn7omfbr8f3ir5gq1fo2cb7bwlha8musehb13eb9eb7hppcmvnyc3gg5i8eprg7w7z0jqrgu:latest/eval/step_40k.table.json", "path": "media/table/eval/step_40k_40000_1c187b6d789cca3b4ec0.table.json", "ncols": 7, "nrows": 50}}
wandb/run-20220828_085247-2hx8pk65/logs/debug-internal.log CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d093f3f44fdf5811d85d2d82d6e9265c3c9bd01c68e29987453a71c23bf7fee1
3
- size 13709864
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1f80a39e659e1401a6be03c1a16ddc1e34a04ddabd64c6c956b346c7ad49ffd6
3
+ size 17019871
wandb/run-20220828_085247-2hx8pk65/run-2hx8pk65.wandb CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:704bf667df7d300df731d8d0bb3301cfbf4b05488ba632c3bc0bcf3bd69bea8c
3
- size 17153865
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4091a31807c620f1680b2cbc9c2340da78bef2667936aaabb64b632931b7b707
3
+ size 21371807