File size: 10,753 Bytes
4641f67 6f51b2d |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 |
2022-02-01 14:24:47,902 INFO MainThread:27520 [wandb_setup.py:_flush():71] setting env: {}
2022-02-01 14:24:47,902 INFO MainThread:27520 [wandb_setup.py:_flush():71] setting login settings: {}
2022-02-01 14:24:47,902 INFO MainThread:27520 [wandb_init.py:_log_setup():371] Logging user logs to /workspace/xls-r-300m-fr/wandb/run-20220201_142447-3hqii6h6/logs/debug.log
2022-02-01 14:24:47,902 INFO MainThread:27520 [wandb_init.py:_log_setup():372] Logging internal logs to /workspace/xls-r-300m-fr/wandb/run-20220201_142447-3hqii6h6/logs/debug-internal.log
2022-02-01 14:24:47,903 INFO MainThread:27520 [wandb_init.py:init():404] calling init triggers
2022-02-01 14:24:47,903 INFO MainThread:27520 [wandb_init.py:init():409] wandb.init called with sweep_config: {}
config: {}
2022-02-01 14:24:47,903 INFO MainThread:27520 [wandb_init.py:init():460] starting backend
2022-02-01 14:24:47,903 INFO MainThread:27520 [backend.py:_multiprocessing_setup():99] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
2022-02-01 14:24:48,004 INFO MainThread:27520 [backend.py:ensure_launched():216] starting backend process...
2022-02-01 14:24:48,099 INFO MainThread:27520 [backend.py:ensure_launched():221] started backend process with pid: 29947
2022-02-01 14:24:48,102 INFO MainThread:27520 [wandb_init.py:init():469] backend started and connected
2022-02-01 14:24:48,111 INFO MainThread:27520 [wandb_init.py:init():533] updated telemetry
2022-02-01 14:24:48,319 INFO MainThread:27520 [wandb_init.py:init():563] communicating current version
2022-02-01 14:24:49,079 INFO MainThread:27520 [wandb_init.py:init():568] got version response
2022-02-01 14:24:49,079 INFO MainThread:27520 [wandb_init.py:init():578] communicating run to backend with 30 second timeout
2022-02-01 14:24:49,275 INFO MainThread:27520 [wandb_init.py:init():606] starting run threads in backend
2022-02-01 14:24:50,008 INFO MainThread:27520 [wandb_run.py:_console_start():1810] atexit reg
2022-02-01 14:24:50,009 INFO MainThread:27520 [wandb_run.py:_redirect():1684] redirect: SettingsConsole.REDIRECT
2022-02-01 14:24:50,010 INFO MainThread:27520 [wandb_run.py:_redirect():1689] Redirecting console.
2022-02-01 14:24:50,018 INFO MainThread:27520 [wandb_run.py:_redirect():1745] Redirects installed.
2022-02-01 14:24:50,018 INFO MainThread:27520 [wandb_init.py:init():633] run started, returning control to user process
2022-02-01 14:24:50,021 INFO MainThread:27520 [wandb_run.py:_config_callback():956] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'float32', 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['Wav2Vec2ForCTC'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 218, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': './checkpoint-18000', 'transformers_version': '4.17.0.dev0', 'feat_extract_dropout': 0.0, 'model_type': 'wav2vec2', 'num_feat_extract_layers': 7, 'hidden_size': 1024, 'feat_extract_norm': 'layer', 'feat_extract_activation': 'gelu', 'conv_dim': [512, 512, 512, 512, 512, 512, 512], 'conv_stride': [5, 2, 2, 2, 2, 2, 2], 'conv_kernel': [10, 3, 3, 3, 3, 2, 2], 'conv_bias': True, 'num_conv_pos_embeddings': 128, 'num_conv_pos_embedding_groups': 16, 'num_hidden_layers': 24, 'intermediate_size': 4096, 'hidden_act': 'gelu', 'num_attention_heads': 16, 'hidden_dropout': 0.0, 'attention_dropout': 0.0, 'activation_dropout': 0.05, 'feat_proj_dropout': 0.0, 'final_dropout': 0.0, 'layerdrop': 0.0, 'layer_norm_eps': 1e-05, 'initializer_range': 0.02, 'vocab_size': 219, 'do_stable_layer_norm': True, 'use_weighted_layer_sum': False, 'apply_spec_augment': True, 'mask_time_prob': 0.75, 'mask_time_length': 10, 'mask_time_min_masks': 2, 'mask_feature_prob': 0.25, 'mask_feature_length': 10, 'mask_feature_min_masks': 0, 'num_codevectors_per_group': 320, 'num_codevector_groups': 2, 'contrastive_logits_temperature': 0.1, 'feat_quantizer_dropout': 0.0, 'num_negatives': 100, 'codevector_dim': 768, 'proj_codevector_dim': 768, 'diversity_loss_weight': 0.1, 'ctc_loss_reduction': 'mean', 'ctc_zero_infinity': False, 'add_adapter': False, 'adapter_kernel_size': 3, 'adapter_stride': 2, 'num_adapter_layers': 3, 'output_hidden_size': 1024, 'classifier_proj_size': 256, 'tdnn_dim': [512, 512, 512, 512, 1500], 'tdnn_kernel': [5, 3, 3, 1, 1], 'tdnn_dilation': [1, 2, 3, 1, 1], 'xvector_output_dim': 512, 'output_dir': './', 'overwrite_output_dir': True, 'do_train': False, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 64, 'per_device_eval_batch_size': 64, 'per_gpu_train_batch_size': 'None', 'per_gpu_eval_batch_size': 'None', 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': 'None', 'learning_rate': 0.0001, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 1.0, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'warmup_ratio': 0.0, 'warmup_steps': 2700, 'log_level': -1, 'log_level_replica': -1, 'log_on_each_node': True, 'logging_dir': './runs/Feb01_14-18-51_job-1abccd0a-3293-4ffe-8274-9e8f841f653f', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 100, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 2, 'save_on_each_node': False, 'no_cuda': False, 'seed': 42, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'amp', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': 'None', 'local_rank': -1, 'xpu_backend': 'None', 'tpu_num_cores': 'None', 'tpu_metrics_debug': False, 'debug': '[]', 'dataloader_drop_last': False, 'eval_steps': 500, 'dataloader_num_workers': 0, 'past_index': -1, 'run_name': 'xls-r-300m-fr', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': 'None', 'load_best_model_at_end': True, 'metric_for_best_model': 'loss', 'greater_is_better': False, 'ignore_data_skip': False, 'sharded_ddp': '[]', 'deepspeed': 'None', 'label_smoothing_factor': 0.0, 'optim': 'adamw_hf', 'adafactor': False, 'group_by_length': True, 'length_column_name': 'input_length', 'report_to': "['wandb']", 'ddp_find_unused_parameters': 'None', 'ddp_bucket_cap_mb': 'None', 'dataloader_pin_memory': True, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': 'None', 'hub_model_id': 'None', 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'gradient_checkpointing': True, 'fp16_backend': 'auto', 'push_to_hub_model_id': 'None', 'push_to_hub_organization': 'None', 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', '_n_gpu': 1, 'mp_parameters': '', 'train_batch_size': 64, 'eval_batch_size': 64}
2022-02-01 14:24:50,026 INFO MainThread:27520 [wandb_watch.py:watch():43] Watching
2022-02-01 14:26:20,525 INFO MainThread:27520 [wandb_run.py:_atexit_cleanup():1780] got exitcode: 0
2022-02-01 14:26:20,528 INFO MainThread:27520 [wandb_run.py:_restore():1752] restore
2022-02-01 14:26:23,413 INFO MainThread:27520 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
wandb_count: 1
}
pusher_stats {
uploaded_bytes: 2177
total_bytes: 2177
}
2022-02-01 14:26:23,726 INFO MainThread:27520 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
wandb_count: 1
}
pusher_stats {
uploaded_bytes: 2177
total_bytes: 2177
}
2022-02-01 14:26:24,351 INFO MainThread:27520 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
wandb_count: 4
}
pusher_stats {
uploaded_bytes: 2177
total_bytes: 14447
}
2022-02-01 14:26:24,454 INFO MainThread:27520 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
wandb_count: 5
}
pusher_stats {
uploaded_bytes: 2177
total_bytes: 17580
}
2022-02-01 14:26:24,566 INFO MainThread:27520 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
wandb_count: 5
}
pusher_stats {
uploaded_bytes: 2177
total_bytes: 17580
}
2022-02-01 14:26:24,669 INFO MainThread:27520 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
wandb_count: 5
}
pusher_stats {
uploaded_bytes: 17580
total_bytes: 17580
}
2022-02-01 14:26:24,772 INFO MainThread:27520 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
wandb_count: 5
}
pusher_stats {
uploaded_bytes: 17580
total_bytes: 17580
}
2022-02-01 14:26:24,875 INFO MainThread:27520 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
wandb_count: 5
}
pusher_stats {
uploaded_bytes: 17580
total_bytes: 17580
}
2022-02-01 14:26:24,978 INFO MainThread:27520 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
wandb_count: 5
}
pusher_stats {
uploaded_bytes: 17580
total_bytes: 17580
}
2022-02-01 14:26:25,080 INFO MainThread:27520 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
wandb_count: 5
}
pusher_stats {
uploaded_bytes: 17580
total_bytes: 17580
}
2022-02-01 14:26:25,555 INFO MainThread:27520 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
wandb_count: 5
}
pusher_stats {
uploaded_bytes: 17580
total_bytes: 17580
}
2022-02-01 14:26:25,964 INFO MainThread:27520 [wandb_run.py:_wait_for_finish():1912] got exit ret: done: true
exit_result {
}
file_counts {
wandb_count: 5
}
pusher_stats {
uploaded_bytes: 17580
total_bytes: 17580
}
local_info {
}
2022-02-01 14:26:27,100 INFO MainThread:27520 [wandb_run.py:_append_history():2130] rendering history
2022-02-01 14:26:27,101 INFO MainThread:27520 [wandb_run.py:_append_summary():2085] rendering summary
2022-02-01 14:26:27,101 INFO MainThread:27520 [wandb_run.py:_append_files():2180] logging synced files
|