File size: 10,770 Bytes
f25a924 4641f67 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 |
2022-02-01 10:30:04,548 INFO MainThread:58652 [wandb_setup.py:_flush():71] setting env: {}
2022-02-01 10:30:04,549 INFO MainThread:58652 [wandb_setup.py:_flush():71] setting login settings: {}
2022-02-01 10:30:04,549 INFO MainThread:58652 [wandb_init.py:_log_setup():371] Logging user logs to /workspace/xls-r-300m-fr/wandb/run-20220201_103004-1yfj7vwy/logs/debug.log
2022-02-01 10:30:04,549 INFO MainThread:58652 [wandb_init.py:_log_setup():372] Logging internal logs to /workspace/xls-r-300m-fr/wandb/run-20220201_103004-1yfj7vwy/logs/debug-internal.log
2022-02-01 10:30:04,549 INFO MainThread:58652 [wandb_init.py:init():404] calling init triggers
2022-02-01 10:30:04,550 INFO MainThread:58652 [wandb_init.py:init():409] wandb.init called with sweep_config: {}
config: {}
2022-02-01 10:30:04,550 INFO MainThread:58652 [wandb_init.py:init():460] starting backend
2022-02-01 10:30:04,550 INFO MainThread:58652 [backend.py:_multiprocessing_setup():99] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
2022-02-01 10:30:04,647 INFO MainThread:58652 [backend.py:ensure_launched():216] starting backend process...
2022-02-01 10:30:04,775 INFO MainThread:58652 [backend.py:ensure_launched():221] started backend process with pid: 61045
2022-02-01 10:30:04,778 INFO MainThread:58652 [wandb_init.py:init():469] backend started and connected
2022-02-01 10:30:04,787 INFO MainThread:58652 [wandb_init.py:init():533] updated telemetry
2022-02-01 10:30:05,012 INFO MainThread:58652 [wandb_init.py:init():563] communicating current version
2022-02-01 10:30:05,703 INFO MainThread:58652 [wandb_init.py:init():568] got version response
2022-02-01 10:30:05,703 INFO MainThread:58652 [wandb_init.py:init():578] communicating run to backend with 30 second timeout
2022-02-01 10:30:05,897 INFO MainThread:58652 [wandb_init.py:init():606] starting run threads in backend
2022-02-01 10:30:06,508 INFO MainThread:58652 [wandb_run.py:_console_start():1810] atexit reg
2022-02-01 10:30:06,509 INFO MainThread:58652 [wandb_run.py:_redirect():1684] redirect: SettingsConsole.REDIRECT
2022-02-01 10:30:06,510 INFO MainThread:58652 [wandb_run.py:_redirect():1689] Redirecting console.
2022-02-01 10:30:06,516 INFO MainThread:58652 [wandb_run.py:_redirect():1745] Redirects installed.
2022-02-01 10:30:06,517 INFO MainThread:58652 [wandb_init.py:init():633] run started, returning control to user process
2022-02-01 10:30:06,519 INFO MainThread:58652 [wandb_run.py:_config_callback():956] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'float32', 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['Wav2Vec2ForPreTraining'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 218, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'facebook/wav2vec2-xls-r-300m', 'transformers_version': '4.17.0.dev0', 'feat_extract_dropout': 0.0, 'model_type': 'wav2vec2', 'num_feat_extract_layers': 7, 'hidden_size': 1024, 'feat_extract_norm': 'layer', 'feat_extract_activation': 'gelu', 'conv_dim': [512, 512, 512, 512, 512, 512, 512], 'conv_stride': [5, 2, 2, 2, 2, 2, 2], 'conv_kernel': [10, 3, 3, 3, 3, 2, 2], 'conv_bias': True, 'num_conv_pos_embeddings': 128, 'num_conv_pos_embedding_groups': 16, 'num_hidden_layers': 24, 'intermediate_size': 4096, 'hidden_act': 'gelu', 'num_attention_heads': 16, 'hidden_dropout': 0.0, 'attention_dropout': 0.0, 'activation_dropout': 0.05, 'feat_proj_dropout': 0.0, 'final_dropout': 0.0, 'layerdrop': 0.0, 'layer_norm_eps': 1e-05, 'initializer_range': 0.02, 'vocab_size': 219, 'do_stable_layer_norm': True, 'use_weighted_layer_sum': False, 'apply_spec_augment': True, 'mask_time_prob': 0.6, 'mask_time_length': 10, 'mask_time_min_masks': 2, 'mask_feature_prob': 0.25, 'mask_feature_length': 10, 'mask_feature_min_masks': 0, 'num_codevectors_per_group': 320, 'num_codevector_groups': 2, 'contrastive_logits_temperature': 0.1, 'feat_quantizer_dropout': 0.0, 'num_negatives': 100, 'codevector_dim': 768, 'proj_codevector_dim': 768, 'diversity_loss_weight': 0.1, 'ctc_loss_reduction': 'mean', 'ctc_zero_infinity': False, 'add_adapter': False, 'adapter_kernel_size': 3, 'adapter_stride': 2, 'num_adapter_layers': 3, 'output_hidden_size': 1024, 'classifier_proj_size': 256, 'tdnn_dim': [512, 512, 512, 512, 1500], 'tdnn_kernel': [5, 3, 3, 1, 1], 'tdnn_dilation': [1, 2, 3, 1, 1], 'xvector_output_dim': 512, 'output_dir': './', 'overwrite_output_dir': True, 'do_train': False, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 64, 'per_device_eval_batch_size': 64, 'per_gpu_train_batch_size': 'None', 'per_gpu_eval_batch_size': 'None', 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': 'None', 'learning_rate': 0.0001, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 4.0, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'warmup_ratio': 0.0, 'warmup_steps': 2700, 'log_level': -1, 'log_level_replica': -1, 'log_on_each_node': True, 'logging_dir': './runs/Feb01_10-24-12_job-1abccd0a-3293-4ffe-8274-9e8f841f653f', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 100, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 2, 'save_on_each_node': False, 'no_cuda': False, 'seed': 42, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'amp', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': 'None', 'local_rank': -1, 'xpu_backend': 'None', 'tpu_num_cores': 'None', 'tpu_metrics_debug': False, 'debug': '[]', 'dataloader_drop_last': False, 'eval_steps': 500, 'dataloader_num_workers': 0, 'past_index': -1, 'run_name': 'xls-r-300m-fr', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': 'None', 'load_best_model_at_end': True, 'metric_for_best_model': 'loss', 'greater_is_better': False, 'ignore_data_skip': False, 'sharded_ddp': '[]', 'deepspeed': 'None', 'label_smoothing_factor': 0.0, 'optim': 'adamw_hf', 'adafactor': False, 'group_by_length': True, 'length_column_name': 'input_length', 'report_to': "['wandb']", 'ddp_find_unused_parameters': 'None', 'ddp_bucket_cap_mb': 'None', 'dataloader_pin_memory': True, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': 'None', 'hub_model_id': 'None', 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'gradient_checkpointing': True, 'fp16_backend': 'auto', 'push_to_hub_model_id': 'None', 'push_to_hub_organization': 'None', 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', '_n_gpu': 1, 'mp_parameters': '', 'train_batch_size': 64, 'eval_batch_size': 64}
2022-02-01 10:30:06,524 INFO MainThread:58652 [wandb_watch.py:watch():43] Watching
2022-02-01 10:32:50,469 INFO MainThread:58652 [wandb_run.py:_atexit_cleanup():1780] got exitcode: 0
2022-02-01 10:32:50,473 INFO MainThread:58652 [wandb_run.py:_restore():1752] restore
2022-02-01 10:32:52,992 INFO MainThread:58652 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
wandb_count: 1
}
pusher_stats {
uploaded_bytes: 2158
total_bytes: 2158
}
2022-02-01 10:32:53,346 INFO MainThread:58652 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
wandb_count: 1
}
pusher_stats {
uploaded_bytes: 2158
total_bytes: 2158
}
2022-02-01 10:32:54,021 INFO MainThread:58652 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
wandb_count: 5
}
pusher_stats {
uploaded_bytes: 2158
total_bytes: 18907
}
2022-02-01 10:32:54,124 INFO MainThread:58652 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
wandb_count: 5
}
pusher_stats {
uploaded_bytes: 2158
total_bytes: 18907
}
2022-02-01 10:32:54,227 INFO MainThread:58652 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
wandb_count: 5
}
pusher_stats {
uploaded_bytes: 2158
total_bytes: 18907
}
2022-02-01 10:32:54,330 INFO MainThread:58652 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
wandb_count: 5
}
pusher_stats {
uploaded_bytes: 18907
total_bytes: 18907
}
2022-02-01 10:32:54,432 INFO MainThread:58652 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
wandb_count: 5
}
pusher_stats {
uploaded_bytes: 18907
total_bytes: 18907
}
2022-02-01 10:32:54,535 INFO MainThread:58652 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
wandb_count: 5
}
pusher_stats {
uploaded_bytes: 18907
total_bytes: 18907
}
2022-02-01 10:32:54,637 INFO MainThread:58652 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
wandb_count: 5
}
pusher_stats {
uploaded_bytes: 18907
total_bytes: 18907
}
2022-02-01 10:32:54,740 INFO MainThread:58652 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
wandb_count: 5
}
pusher_stats {
uploaded_bytes: 18907
total_bytes: 18907
}
2022-02-01 10:32:55,208 INFO MainThread:58652 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
wandb_count: 5
}
pusher_stats {
uploaded_bytes: 18907
total_bytes: 18907
}
2022-02-01 10:32:55,612 INFO MainThread:58652 [wandb_run.py:_wait_for_finish():1912] got exit ret: done: true
exit_result {
}
file_counts {
wandb_count: 5
}
pusher_stats {
uploaded_bytes: 18907
total_bytes: 18907
}
local_info {
}
2022-02-01 10:32:56,753 INFO MainThread:58652 [wandb_run.py:_append_history():2130] rendering history
2022-02-01 10:32:56,754 INFO MainThread:58652 [wandb_run.py:_append_summary():2085] rendering summary
2022-02-01 10:32:56,754 INFO MainThread:58652 [wandb_run.py:_append_files():2180] logging synced files
|