File size: 11,359 Bytes
1c4abd7
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5200b6e
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
2022-01-29 21:54:51,708 INFO    MainThread:7453 [wandb_setup.py:_flush():71] setting env: {}
2022-01-29 21:54:51,708 INFO    MainThread:7453 [wandb_setup.py:_flush():71] setting login settings: {}
2022-01-29 21:54:51,708 INFO    MainThread:7453 [wandb_init.py:_log_setup():371] Logging user logs to /workspace/xls-r-300m-fr/wandb/run-20220129_215451-1vipdbow/logs/debug.log
2022-01-29 21:54:51,709 INFO    MainThread:7453 [wandb_init.py:_log_setup():372] Logging internal logs to /workspace/xls-r-300m-fr/wandb/run-20220129_215451-1vipdbow/logs/debug-internal.log
2022-01-29 21:54:51,709 INFO    MainThread:7453 [wandb_init.py:init():404] calling init triggers
2022-01-29 21:54:51,709 INFO    MainThread:7453 [wandb_init.py:init():409] wandb.init called with sweep_config: {}
config: {}
2022-01-29 21:54:51,709 INFO    MainThread:7453 [wandb_init.py:init():460] starting backend
2022-01-29 21:54:51,709 INFO    MainThread:7453 [backend.py:_multiprocessing_setup():99] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
2022-01-29 21:54:51,806 INFO    MainThread:7453 [backend.py:ensure_launched():216] starting backend process...
2022-01-29 21:54:51,893 INFO    MainThread:7453 [backend.py:ensure_launched():221] started backend process with pid: 37713
2022-01-29 21:54:51,897 INFO    MainThread:7453 [wandb_init.py:init():469] backend started and connected
2022-01-29 21:54:51,906 INFO    MainThread:7453 [wandb_init.py:init():533] updated telemetry
2022-01-29 21:54:52,074 INFO    MainThread:7453 [wandb_init.py:init():563] communicating current version
2022-01-29 21:54:52,818 INFO    MainThread:7453 [wandb_init.py:init():568] got version response 
2022-01-29 21:54:52,818 INFO    MainThread:7453 [wandb_init.py:init():578] communicating run to backend with 30 second timeout
2022-01-29 21:54:53,017 INFO    MainThread:7453 [wandb_init.py:init():606] starting run threads in backend
2022-01-29 21:54:53,621 INFO    MainThread:7453 [wandb_run.py:_console_start():1810] atexit reg
2022-01-29 21:54:53,622 INFO    MainThread:7453 [wandb_run.py:_redirect():1684] redirect: SettingsConsole.REDIRECT
2022-01-29 21:54:53,623 INFO    MainThread:7453 [wandb_run.py:_redirect():1689] Redirecting console.
2022-01-29 21:54:53,628 INFO    MainThread:7453 [wandb_run.py:_redirect():1745] Redirects installed.
2022-01-29 21:54:53,628 INFO    MainThread:7453 [wandb_init.py:init():633] run started, returning control to user process
2022-01-29 21:54:53,630 INFO    MainThread:7453 [wandb_run.py:_config_callback():956] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'float32', 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['Wav2Vec2ForPreTraining'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 216, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'facebook/wav2vec2-xls-r-300m', 'transformers_version': '4.17.0.dev0', 'feat_extract_dropout': 0.0, 'model_type': 'wav2vec2', 'num_feat_extract_layers': 7, 'hidden_size': 1024, 'feat_extract_norm': 'layer', 'feat_extract_activation': 'gelu', 'conv_dim': [512, 512, 512, 512, 512, 512, 512], 'conv_stride': [5, 2, 2, 2, 2, 2, 2], 'conv_kernel': [10, 3, 3, 3, 3, 2, 2], 'conv_bias': True, 'num_conv_pos_embeddings': 128, 'num_conv_pos_embedding_groups': 16, 'num_hidden_layers': 24, 'intermediate_size': 4096, 'hidden_act': 'gelu', 'num_attention_heads': 16, 'hidden_dropout': 0.0, 'attention_dropout': 0.0, 'activation_dropout': 0.05, 'feat_proj_dropout': 0.0, 'final_dropout': 0.0, 'layerdrop': 0.0, 'layer_norm_eps': 1e-05, 'initializer_range': 0.02, 'vocab_size': 218, 'do_stable_layer_norm': True, 'use_weighted_layer_sum': False, 'apply_spec_augment': True, 'mask_time_prob': 0.75, 'mask_time_length': 10, 'mask_time_min_masks': 2, 'mask_feature_prob': 0.4, 'mask_feature_length': 10, 'mask_feature_min_masks': 0, 'num_codevectors_per_group': 320, 'num_codevector_groups': 2, 'contrastive_logits_temperature': 0.1, 'feat_quantizer_dropout': 0.0, 'num_negatives': 100, 'codevector_dim': 768, 'proj_codevector_dim': 768, 'diversity_loss_weight': 0.1, 'ctc_loss_reduction': 'mean', 'ctc_zero_infinity': False, 'add_adapter': False, 'adapter_kernel_size': 3, 'adapter_stride': 2, 'num_adapter_layers': 3, 'output_hidden_size': 1024, 'classifier_proj_size': 256, 'tdnn_dim': [512, 512, 512, 512, 1500], 'tdnn_kernel': [5, 3, 3, 1, 1], 'tdnn_dilation': [1, 2, 3, 1, 1], 'xvector_output_dim': 512, 'output_dir': './', 'overwrite_output_dir': True, 'do_train': True, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 64, 'per_device_eval_batch_size': 64, 'per_gpu_train_batch_size': 'None', 'per_gpu_eval_batch_size': 'None', 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': 'None', 'learning_rate': 0.0001, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 2.0, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'warmup_ratio': 0.0, 'warmup_steps': 1500, 'log_level': -1, 'log_level_replica': -1, 'log_on_each_node': True, 'logging_dir': './runs/Jan29_20-37-14_job-1abccd0a-3293-4ffe-8274-9e8f841f653f', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 100, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 2, 'save_on_each_node': False, 'no_cuda': False, 'seed': 42, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'amp', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': 'None', 'local_rank': -1, 'xpu_backend': 'None', 'tpu_num_cores': 'None', 'tpu_metrics_debug': False, 'debug': '[]', 'dataloader_drop_last': False, 'eval_steps': 500, 'dataloader_num_workers': 0, 'past_index': -1, 'run_name': 'xls-r-300m-fr', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': 'None', 'load_best_model_at_end': True, 'metric_for_best_model': 'loss', 'greater_is_better': False, 'ignore_data_skip': False, 'sharded_ddp': '[]', 'deepspeed': 'None', 'label_smoothing_factor': 0.0, 'optim': 'adamw_hf', 'adafactor': False, 'group_by_length': True, 'length_column_name': 'input_length', 'report_to': "['wandb']", 'ddp_find_unused_parameters': 'None', 'ddp_bucket_cap_mb': 'None', 'dataloader_pin_memory': True, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': 'None', 'hub_model_id': 'None', 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'gradient_checkpointing': True, 'fp16_backend': 'auto', 'push_to_hub_model_id': 'None', 'push_to_hub_organization': 'None', 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', '_n_gpu': 1, 'mp_parameters': '', 'train_batch_size': 64, 'eval_batch_size': 64}
2022-01-29 21:54:53,634 INFO    MainThread:7453 [wandb_watch.py:watch():43] Watching
2022-01-30 12:56:04,998 INFO    MainThread:7453 [wandb_run.py:_atexit_cleanup():1780] got exitcode: 0
2022-01-30 12:56:05,003 INFO    MainThread:7453 [wandb_run.py:_restore():1752] restore
2022-01-30 12:56:07,450 INFO    MainThread:7453 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 1
}
pusher_stats {
  uploaded_bytes: 2179
  total_bytes: 2179
}

2022-01-30 12:56:07,569 INFO    MainThread:7453 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 1
}
pusher_stats {
  uploaded_bytes: 2179
  total_bytes: 2179
}

2022-01-30 12:56:08,270 INFO    MainThread:7453 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 1
}
pusher_stats {
  uploaded_bytes: 2179
  total_bytes: 2179
}

2022-01-30 12:56:08,707 INFO    MainThread:7453 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 3
}
pusher_stats {
  uploaded_bytes: 2179
  total_bytes: 251909
}

2022-01-30 12:56:08,810 INFO    MainThread:7453 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 5
}
pusher_stats {
  uploaded_bytes: 2179
  total_bytes: 1004734
}

2022-01-30 12:56:08,928 INFO    MainThread:7453 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 5
}
pusher_stats {
  uploaded_bytes: 2179
  total_bytes: 1004734
}

2022-01-30 12:56:09,032 INFO    MainThread:7453 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 5
}
pusher_stats {
  uploaded_bytes: 1004734
  total_bytes: 1004734
}

2022-01-30 12:56:09,135 INFO    MainThread:7453 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 5
}
pusher_stats {
  uploaded_bytes: 1004734
  total_bytes: 1004734
}

2022-01-30 12:56:09,238 INFO    MainThread:7453 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 5
}
pusher_stats {
  uploaded_bytes: 1004734
  total_bytes: 1004734
}

2022-01-30 12:56:09,341 INFO    MainThread:7453 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 5
}
pusher_stats {
  uploaded_bytes: 1004734
  total_bytes: 1004734
}

2022-01-30 12:56:09,444 INFO    MainThread:7453 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 5
}
pusher_stats {
  uploaded_bytes: 1004734
  total_bytes: 1004734
}

2022-01-30 12:56:09,548 INFO    MainThread:7453 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 5
}
pusher_stats {
  uploaded_bytes: 1004734
  total_bytes: 1004734
}

2022-01-30 12:56:09,651 INFO    MainThread:7453 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 5
}
pusher_stats {
  uploaded_bytes: 1004734
  total_bytes: 1004734
}

2022-01-30 12:56:10,713 INFO    MainThread:7453 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 5
}
pusher_stats {
  uploaded_bytes: 1004734
  total_bytes: 1004734
}

2022-01-30 12:56:11,122 INFO    MainThread:7453 [wandb_run.py:_wait_for_finish():1912] got exit ret: done: true
exit_result {
}
file_counts {
  wandb_count: 5
}
pusher_stats {
  uploaded_bytes: 1004734
  total_bytes: 1004734
}
local_info {
}

2022-01-30 12:56:12,305 INFO    MainThread:7453 [wandb_run.py:_append_history():2130] rendering history
2022-01-30 12:56:12,307 INFO    MainThread:7453 [wandb_run.py:_append_summary():2085] rendering summary
2022-01-30 12:56:12,308 INFO    MainThread:7453 [wandb_run.py:_append_files():2180] logging synced files