{ "checkpoint_path": "/data_2to/devel_data/nn_pruning/output/squadv2_test_2/hp_mnop-bert-large-uncased-whole-word-masking_dn-squad_v2_v2wn1_od-__data_2to__devel_data__nn_pruning__output__squadv2_test_2___es-steps_pdtbs8_pdebs128_nte20_ls250_stl50_est5000_rn---58ebffa395d47d66/checkpoint-325000", "config": { "_name_or_path": "/tmp/tmpitf3rdr5", "architectures": ["BertForQuestionAnswering"], "attention_probs_dropout_prob": 0.1, "gradient_checkpointing": false, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 1024, "initializer_range": 0.02, "intermediate_size": 4096, "layer_norm_eps": 1e-12, "max_position_embeddings": 512, "model_type": "bert", "num_attention_heads": 16, "num_hidden_layers": 24, "pad_token_id": 0, "position_embedding_type": "absolute", "pruned_heads": { "0": [0, 1, 3, 4, 5, 8, 9, 13, 15], "1": [0, 1, 3, 5, 7, 9, 10, 13, 14], "10": [1, 2, 4, 5, 6, 8, 11, 13], "11": [0, 2, 5, 6, 7, 8, 10, 12, 15], "12": [0, 2, 6, 8, 9, 11, 13], "13": [2, 6, 10, 12, 15], "14": [1, 5, 6, 10, 11, 15], "15": [0, 9], "16": [5, 7], "17": [1, 4, 8, 12, 14], "18": [3, 4, 11], "19": [0, 5, 12], "2": [0, 1, 4, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15], "20": [0, 4, 10, 12], "21": [0, 2, 3, 4, 8, 11, 12, 15], "22": [0, 1, 3, 4, 7, 9, 10, 11, 13, 15], "23": [2, 4, 8, 9, 10, 13, 14, 15], "3": [0, 3, 5, 6, 7, 8, 9, 11, 13, 14, 15], "4": [0, 2, 3, 4, 5, 6, 7, 9, 10, 11, 12, 14], "5": [1, 3, 4, 5, 6, 7, 8, 9, 12, 13, 14, 15], "6": [0, 1, 2, 3, 4, 7, 8, 9, 10, 11, 12, 15], "7": [0, 3, 4, 5, 8, 9, 10, 11, 12, 13, 14], "8": [3, 4, 5, 7, 8, 9, 10, 11, 12], "9": [0, 1, 2, 3, 5, 6, 7, 9, 10, 13, 14, 15] }, "transformers_version": "4.5.1", "type_vocab_size": 2, "use_cache": true, "vocab_size": 30522 }, "eval_metrics": { "HasAns_exact": 74.8144399460189, "HasAns_f1": 80.555306012496, "HasAns_total": 5928, "NoAns_exact": 84.57527333894029, "NoAns_f1": 84.57527333894029, "NoAns_total": 5945, "best_exact": 79.70184452118251, "best_exact_thresh": 0.0, "best_f1": 82.56816761071966, "best_f1_thresh": 0.0, "exact": 79.70184452118251, "f1": 82.56816761071981, "main_metric": 82.56816761071981, "total": 11873 }, "model_args": { "cache_dir": null, "config_name": null, "model_name_or_path": "bert-large-uncased-whole-word-masking", "tokenizer_name": null, "use_fast_tokenizer": true }, "sparse_args": { "ampere_pruning_method": "disabled", "attention_block_cols": 32, "attention_block_rows": 32, "attention_lambda": 1.0, "attention_output_with_dense": 0, "attention_pruning_method": "sigmoied_threshold", "bias_mask": true, "dense_block_cols": 1, "dense_block_rows": 1, "dense_lambda": 1.0, "dense_pruning_method": "sigmoied_threshold:1d_alt", "distil_alpha_ce": 0.1, "distil_alpha_teacher": 0.9, "distil_teacher_name_or_path": "madlag/bert-large-uncased-whole-word-masking-finetuned-squadv2", "distil_temperature": 2.0, "eval_with_current_patch_params": false, "final_ampere_temperature": 20.0, "final_finetune": false, "final_threshold": 0.1, "final_warmup": 10, "gelu_patch": 0, "gelu_patch_steps": 50000, "initial_ampere_temperature": 0.0, "initial_threshold": 0, "initial_warmup": 1, "layer_norm_patch": 0, "layer_norm_patch_start_delta": 0.99, "layer_norm_patch_steps": 50000, "linear_min_parameters": 0.005, "mask_init": "constant", "mask_scale": 0.0, "mask_scores_learning_rate": 0.01, "qat": false, "qconfig": "default", "regularization": "l1", "regularization_final_lambda": 5, "rewind_model_name_or_path": null }, "speed": { "cuda_eval_elapsed_time": 52.126098907470706, "eval_elapsed_time": 56.62768604200005 }, "speedup": 0.8466287123910415, "stats": { "layers": { "0": { "linear_attention_nnz": 998400, "linear_attention_total": 4194304, "linear_dense_nnz": 587776, "linear_dense_total": 8388608, "linear_nnz": 1586176, "linear_total": 12582912, "nnz": 1593727, "total": 12594496 }, "1": { "linear_attention_nnz": 1025024, "linear_attention_total": 4194304, "linear_dense_nnz": 624640, "linear_dense_total": 8388608, "linear_nnz": 1649664, "linear_total": 12582912, "nnz": 1657297, "total": 12594496 }, "10": { "linear_attention_nnz": 1531904, "linear_attention_total": 4194304, "linear_dense_nnz": 1267712, "linear_dense_total": 8388608, "linear_nnz": 2799616, "linear_total": 12582912, "nnz": 2807915, "total": 12594688 }, "11": { "linear_attention_nnz": 1508352, "linear_attention_total": 4194304, "linear_dense_nnz": 1351680, "linear_dense_total": 8388608, "linear_nnz": 2860032, "linear_total": 12582912, "nnz": 2868180, "total": 12594496 }, "12": { "linear_attention_nnz": 1525760, "linear_attention_total": 4194304, "linear_dense_nnz": 1236992, "linear_dense_total": 8388608, "linear_nnz": 2762752, "linear_total": 12582912, "nnz": 2771132, "total": 12594880 }, "13": { "linear_attention_nnz": 1813504, "linear_attention_total": 4194304, "linear_dense_nnz": 1423360, "linear_dense_total": 8388608, "linear_nnz": 3236864, "linear_total": 12582912, "nnz": 3245559, "total": 12595264 }, "14": { "linear_attention_nnz": 1774592, "linear_attention_total": 4194304, "linear_dense_nnz": 1153024, "linear_dense_total": 8388608, "linear_nnz": 2927616, "linear_total": 12582912, "nnz": 2936051, "total": 12595072 }, "15": { "linear_attention_nnz": 1909760, "linear_attention_total": 4194304, "linear_dense_nnz": 1077248, "linear_dense_total": 8388608, "linear_nnz": 2987008, "linear_total": 12582912, "nnz": 2996110, "total": 12595840 }, "16": { "linear_attention_nnz": 2169856, "linear_attention_total": 4194304, "linear_dense_nnz": 1091584, "linear_dense_total": 8388608, "linear_nnz": 3261440, "linear_total": 12582912, "nnz": 3270645, "total": 12595840 }, "17": { "linear_attention_nnz": 1823744, "linear_attention_total": 4194304, "linear_dense_nnz": 1071104, "linear_dense_total": 8388608, "linear_nnz": 2894848, "linear_total": 12582912, "nnz": 2903531, "total": 12595264 }, "18": { "linear_attention_nnz": 1886208, "linear_attention_total": 4194304, "linear_dense_nnz": 774144, "linear_dense_total": 8388608, "linear_nnz": 2660352, "linear_total": 12582912, "nnz": 2669146, "total": 12595648 }, "19": { "linear_attention_nnz": 1472512, "linear_attention_total": 4194304, "linear_dense_nnz": 446464, "linear_dense_total": 8388608, "linear_nnz": 1918976, "linear_total": 12582912, "nnz": 1927354, "total": 12595648 }, "2": { "linear_attention_nnz": 595968, "linear_attention_total": 4194304, "linear_dense_nnz": 876544, "linear_dense_total": 8388608, "linear_nnz": 1472512, "linear_total": 12582912, "nnz": 1479660, "total": 12593728 }, "20": { "linear_attention_nnz": 1079296, "linear_attention_total": 4194304, "linear_dense_nnz": 299008, "linear_dense_total": 8388608, "linear_nnz": 1378304, "linear_total": 12582912, "nnz": 1386290, "total": 12595456 }, "21": { "linear_attention_nnz": 512000, "linear_attention_total": 4194304, "linear_dense_nnz": 155648, "linear_dense_total": 8388608, "linear_nnz": 667648, "linear_total": 12582912, "nnz": 674764, "total": 12594688 }, "22": { "linear_attention_nnz": 375808, "linear_attention_total": 4194304, "linear_dense_nnz": 90112, "linear_dense_total": 8388608, "linear_nnz": 465920, "linear_total": 12582912, "nnz": 472716, "total": 12594304 }, "23": { "linear_attention_nnz": 343040, "linear_attention_total": 4194304, "linear_dense_nnz": 221184, "linear_dense_total": 8388608, "linear_nnz": 564224, "linear_total": 12582912, "nnz": 571244, "total": 12594688 }, "3": { "linear_attention_nnz": 916480, "linear_attention_total": 4194304, "linear_dense_nnz": 1085440, "linear_dense_total": 8388608, "linear_nnz": 2001920, "linear_total": 12582912, "nnz": 2009554, "total": 12594112 }, "4": { "linear_attention_nnz": 678912, "linear_attention_total": 4194304, "linear_dense_nnz": 1146880, "linear_dense_total": 8388608, "linear_nnz": 1825792, "linear_total": 12582912, "nnz": 1833264, "total": 12593920 }, "5": { "linear_attention_nnz": 509952, "linear_attention_total": 4194304, "linear_dense_nnz": 1308672, "linear_dense_total": 8388608, "linear_nnz": 1818624, "linear_total": 12582912, "nnz": 1825983, "total": 12593920 }, "6": { "linear_attention_nnz": 717824, "linear_attention_total": 4194304, "linear_dense_nnz": 1441792, "linear_dense_total": 8388608, "linear_nnz": 2159616, "linear_total": 12582912, "nnz": 2167168, "total": 12593920 }, "7": { "linear_attention_nnz": 1009664, "linear_attention_total": 4194304, "linear_dense_nnz": 1468416, "linear_dense_total": 8388608, "linear_nnz": 2478080, "linear_total": 12582912, "nnz": 2485901, "total": 12594112 }, "8": { "linear_attention_nnz": 1327104, "linear_attention_total": 4194304, "linear_dense_nnz": 1468416, "linear_dense_total": 8388608, "linear_nnz": 2795520, "linear_total": 12582912, "nnz": 2803661, "total": 12594496 }, "9": { "linear_attention_nnz": 631808, "linear_attention_total": 4194304, "linear_dense_nnz": 1531904, "linear_dense_total": 8388608, "linear_nnz": 2163712, "linear_total": 12582912, "nnz": 2171276, "total": 12593920 } }, "linear_nnz": 51337216, "linear_sparsity": 83.00035264756944, "linear_total": 301989888, "nnz": 83313090, "pruned_heads": { "0": [0, 1, 3, 4, 5, 8, 9, 13, 15], "1": [0, 1, 3, 5, 7, 9, 10, 13, 14], "10": [1, 2, 4, 5, 6, 8, 11, 13], "11": [0, 2, 5, 6, 7, 8, 10, 12, 15], "12": [0, 2, 6, 8, 9, 11, 13], "13": [2, 6, 10, 12, 15], "14": [1, 5, 6, 10, 11, 15], "15": [0, 9], "16": [5, 7], "17": [1, 4, 8, 12, 14], "18": [3, 11, 4], "19": [0, 12, 5], "2": [0, 1, 4, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15], "20": [0, 10, 4, 12], "21": [0, 2, 3, 4, 8, 11, 12, 15], "22": [0, 1, 3, 4, 7, 9, 10, 11, 13, 15], "23": [2, 4, 8, 9, 10, 13, 14, 15], "3": [0, 3, 5, 6, 7, 8, 9, 11, 13, 14, 15], "4": [0, 2, 3, 4, 5, 6, 7, 9, 10, 11, 12, 14], "5": [1, 3, 4, 5, 6, 7, 8, 9, 12, 13, 14, 15], "6": [0, 1, 2, 3, 4, 7, 8, 9, 10, 11, 12, 15], "7": [0, 3, 4, 5, 8, 9, 10, 11, 12, 13, 14], "8": [3, 4, 5, 7, 8, 9, 10, 11, 12], "9": [0, 1, 2, 3, 5, 6, 7, 9, 10, 13, 14, 15] }, "total": 334057858, "total_sparsity": 75.06028132408129 }, "training_args": { "_n_gpu": 1, "adafactor": false, "adam_beta1": 0.9, "adam_beta2": 0.999, "adam_epsilon": 1e-08, "dataloader_drop_last": false, "dataloader_num_workers": 0, "dataloader_pin_memory": true, "ddp_find_unused_parameters": null, "debug": false, "deepspeed": null, "disable_tqdm": false, "do_eval": 1, "do_predict": false, "do_train": 1, "eval_accumulation_steps": null, "eval_steps": 5000, "evaluation_strategy": "steps", "fp16": false, "fp16_backend": "auto", "fp16_full_eval": false, "fp16_opt_level": "O1", "gradient_accumulation_steps": 1, "greater_is_better": null, "group_by_length": false, "ignore_data_skip": false, "label_names": null, "label_smoothing_factor": 0.0, "learning_rate": 3e-05, "length_column_name": "length", "load_best_model_at_end": false, "local_rank": -1, "logging_dir": "/data_2to/devel_data/nn_pruning/output/squadv2_test_2/", "logging_first_step": false, "logging_steps": 250, "logging_strategy": "steps", "lr_scheduler_type": "linear", "max_grad_norm": 1.0, "max_steps": -1, "metric_for_best_model": null, "mp_parameters": "", "no_cuda": false, "num_train_epochs": 20, "optimize_model_before_eval": "disabled", "output_dir": "/data_2to/devel_data/nn_pruning/output/squadv2_test_2/", "overwrite_output_dir": 1, "past_index": -1, "per_device_eval_batch_size": 128, "per_device_train_batch_size": 8, "per_gpu_eval_batch_size": null, "per_gpu_train_batch_size": null, "prediction_loss_only": false, "remove_unused_columns": true, "report_to": ["tensorboard", "wandb"], "run_name": "/data_2to/devel_data/nn_pruning/output/squadv2_test_2/", "save_steps": 5000, "save_strategy": "steps", "save_total_limit": 50, "seed": 17, "sharded_ddp": [], "skip_memory_metrics": false, "tpu_metrics_debug": false, "tpu_num_cores": null, "warmup_ratio": 0.0, "warmup_steps": 5400, "weight_decay": 0.0 } }