File size: 10,753 Bytes
4641f67
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
6f51b2d
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
2022-02-01 14:24:47,902 INFO    MainThread:27520 [wandb_setup.py:_flush():71] setting env: {}
2022-02-01 14:24:47,902 INFO    MainThread:27520 [wandb_setup.py:_flush():71] setting login settings: {}
2022-02-01 14:24:47,902 INFO    MainThread:27520 [wandb_init.py:_log_setup():371] Logging user logs to /workspace/xls-r-300m-fr/wandb/run-20220201_142447-3hqii6h6/logs/debug.log
2022-02-01 14:24:47,902 INFO    MainThread:27520 [wandb_init.py:_log_setup():372] Logging internal logs to /workspace/xls-r-300m-fr/wandb/run-20220201_142447-3hqii6h6/logs/debug-internal.log
2022-02-01 14:24:47,903 INFO    MainThread:27520 [wandb_init.py:init():404] calling init triggers
2022-02-01 14:24:47,903 INFO    MainThread:27520 [wandb_init.py:init():409] wandb.init called with sweep_config: {}
config: {}
2022-02-01 14:24:47,903 INFO    MainThread:27520 [wandb_init.py:init():460] starting backend
2022-02-01 14:24:47,903 INFO    MainThread:27520 [backend.py:_multiprocessing_setup():99] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
2022-02-01 14:24:48,004 INFO    MainThread:27520 [backend.py:ensure_launched():216] starting backend process...
2022-02-01 14:24:48,099 INFO    MainThread:27520 [backend.py:ensure_launched():221] started backend process with pid: 29947
2022-02-01 14:24:48,102 INFO    MainThread:27520 [wandb_init.py:init():469] backend started and connected
2022-02-01 14:24:48,111 INFO    MainThread:27520 [wandb_init.py:init():533] updated telemetry
2022-02-01 14:24:48,319 INFO    MainThread:27520 [wandb_init.py:init():563] communicating current version
2022-02-01 14:24:49,079 INFO    MainThread:27520 [wandb_init.py:init():568] got version response 
2022-02-01 14:24:49,079 INFO    MainThread:27520 [wandb_init.py:init():578] communicating run to backend with 30 second timeout
2022-02-01 14:24:49,275 INFO    MainThread:27520 [wandb_init.py:init():606] starting run threads in backend
2022-02-01 14:24:50,008 INFO    MainThread:27520 [wandb_run.py:_console_start():1810] atexit reg
2022-02-01 14:24:50,009 INFO    MainThread:27520 [wandb_run.py:_redirect():1684] redirect: SettingsConsole.REDIRECT
2022-02-01 14:24:50,010 INFO    MainThread:27520 [wandb_run.py:_redirect():1689] Redirecting console.
2022-02-01 14:24:50,018 INFO    MainThread:27520 [wandb_run.py:_redirect():1745] Redirects installed.
2022-02-01 14:24:50,018 INFO    MainThread:27520 [wandb_init.py:init():633] run started, returning control to user process
2022-02-01 14:24:50,021 INFO    MainThread:27520 [wandb_run.py:_config_callback():956] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'float32', 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['Wav2Vec2ForCTC'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 218, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': './checkpoint-18000', 'transformers_version': '4.17.0.dev0', 'feat_extract_dropout': 0.0, 'model_type': 'wav2vec2', 'num_feat_extract_layers': 7, 'hidden_size': 1024, 'feat_extract_norm': 'layer', 'feat_extract_activation': 'gelu', 'conv_dim': [512, 512, 512, 512, 512, 512, 512], 'conv_stride': [5, 2, 2, 2, 2, 2, 2], 'conv_kernel': [10, 3, 3, 3, 3, 2, 2], 'conv_bias': True, 'num_conv_pos_embeddings': 128, 'num_conv_pos_embedding_groups': 16, 'num_hidden_layers': 24, 'intermediate_size': 4096, 'hidden_act': 'gelu', 'num_attention_heads': 16, 'hidden_dropout': 0.0, 'attention_dropout': 0.0, 'activation_dropout': 0.05, 'feat_proj_dropout': 0.0, 'final_dropout': 0.0, 'layerdrop': 0.0, 'layer_norm_eps': 1e-05, 'initializer_range': 0.02, 'vocab_size': 219, 'do_stable_layer_norm': True, 'use_weighted_layer_sum': False, 'apply_spec_augment': True, 'mask_time_prob': 0.75, 'mask_time_length': 10, 'mask_time_min_masks': 2, 'mask_feature_prob': 0.25, 'mask_feature_length': 10, 'mask_feature_min_masks': 0, 'num_codevectors_per_group': 320, 'num_codevector_groups': 2, 'contrastive_logits_temperature': 0.1, 'feat_quantizer_dropout': 0.0, 'num_negatives': 100, 'codevector_dim': 768, 'proj_codevector_dim': 768, 'diversity_loss_weight': 0.1, 'ctc_loss_reduction': 'mean', 'ctc_zero_infinity': False, 'add_adapter': False, 'adapter_kernel_size': 3, 'adapter_stride': 2, 'num_adapter_layers': 3, 'output_hidden_size': 1024, 'classifier_proj_size': 256, 'tdnn_dim': [512, 512, 512, 512, 1500], 'tdnn_kernel': [5, 3, 3, 1, 1], 'tdnn_dilation': [1, 2, 3, 1, 1], 'xvector_output_dim': 512, 'output_dir': './', 'overwrite_output_dir': True, 'do_train': False, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 64, 'per_device_eval_batch_size': 64, 'per_gpu_train_batch_size': 'None', 'per_gpu_eval_batch_size': 'None', 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': 'None', 'learning_rate': 0.0001, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 1.0, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'warmup_ratio': 0.0, 'warmup_steps': 2700, 'log_level': -1, 'log_level_replica': -1, 'log_on_each_node': True, 'logging_dir': './runs/Feb01_14-18-51_job-1abccd0a-3293-4ffe-8274-9e8f841f653f', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 100, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 2, 'save_on_each_node': False, 'no_cuda': False, 'seed': 42, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'amp', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': 'None', 'local_rank': -1, 'xpu_backend': 'None', 'tpu_num_cores': 'None', 'tpu_metrics_debug': False, 'debug': '[]', 'dataloader_drop_last': False, 'eval_steps': 500, 'dataloader_num_workers': 0, 'past_index': -1, 'run_name': 'xls-r-300m-fr', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': 'None', 'load_best_model_at_end': True, 'metric_for_best_model': 'loss', 'greater_is_better': False, 'ignore_data_skip': False, 'sharded_ddp': '[]', 'deepspeed': 'None', 'label_smoothing_factor': 0.0, 'optim': 'adamw_hf', 'adafactor': False, 'group_by_length': True, 'length_column_name': 'input_length', 'report_to': "['wandb']", 'ddp_find_unused_parameters': 'None', 'ddp_bucket_cap_mb': 'None', 'dataloader_pin_memory': True, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': 'None', 'hub_model_id': 'None', 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'gradient_checkpointing': True, 'fp16_backend': 'auto', 'push_to_hub_model_id': 'None', 'push_to_hub_organization': 'None', 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', '_n_gpu': 1, 'mp_parameters': '', 'train_batch_size': 64, 'eval_batch_size': 64}
2022-02-01 14:24:50,026 INFO    MainThread:27520 [wandb_watch.py:watch():43] Watching
2022-02-01 14:26:20,525 INFO    MainThread:27520 [wandb_run.py:_atexit_cleanup():1780] got exitcode: 0
2022-02-01 14:26:20,528 INFO    MainThread:27520 [wandb_run.py:_restore():1752] restore
2022-02-01 14:26:23,413 INFO    MainThread:27520 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 1
}
pusher_stats {
  uploaded_bytes: 2177
  total_bytes: 2177
}

2022-02-01 14:26:23,726 INFO    MainThread:27520 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 1
}
pusher_stats {
  uploaded_bytes: 2177
  total_bytes: 2177
}

2022-02-01 14:26:24,351 INFO    MainThread:27520 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 4
}
pusher_stats {
  uploaded_bytes: 2177
  total_bytes: 14447
}

2022-02-01 14:26:24,454 INFO    MainThread:27520 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 5
}
pusher_stats {
  uploaded_bytes: 2177
  total_bytes: 17580
}

2022-02-01 14:26:24,566 INFO    MainThread:27520 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 5
}
pusher_stats {
  uploaded_bytes: 2177
  total_bytes: 17580
}

2022-02-01 14:26:24,669 INFO    MainThread:27520 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 5
}
pusher_stats {
  uploaded_bytes: 17580
  total_bytes: 17580
}

2022-02-01 14:26:24,772 INFO    MainThread:27520 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 5
}
pusher_stats {
  uploaded_bytes: 17580
  total_bytes: 17580
}

2022-02-01 14:26:24,875 INFO    MainThread:27520 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 5
}
pusher_stats {
  uploaded_bytes: 17580
  total_bytes: 17580
}

2022-02-01 14:26:24,978 INFO    MainThread:27520 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 5
}
pusher_stats {
  uploaded_bytes: 17580
  total_bytes: 17580
}

2022-02-01 14:26:25,080 INFO    MainThread:27520 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 5
}
pusher_stats {
  uploaded_bytes: 17580
  total_bytes: 17580
}

2022-02-01 14:26:25,555 INFO    MainThread:27520 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
  wandb_count: 5
}
pusher_stats {
  uploaded_bytes: 17580
  total_bytes: 17580
}

2022-02-01 14:26:25,964 INFO    MainThread:27520 [wandb_run.py:_wait_for_finish():1912] got exit ret: done: true
exit_result {
}
file_counts {
  wandb_count: 5
}
pusher_stats {
  uploaded_bytes: 17580
  total_bytes: 17580
}
local_info {
}

2022-02-01 14:26:27,100 INFO    MainThread:27520 [wandb_run.py:_append_history():2130] rendering history
2022-02-01 14:26:27,101 INFO    MainThread:27520 [wandb_run.py:_append_summary():2085] rendering summary
2022-02-01 14:26:27,101 INFO    MainThread:27520 [wandb_run.py:_append_files():2180] logging synced files