juliamendelsohn commited on
Commit
d2de03f
1 Parent(s): a3fbb71

commit from juliame

Browse files
config.json ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "RobertaForMultiLabelSequenceClassification"
4
+ ],
5
+ "attention_probs_dropout_prob": 0.1,
6
+ "bos_token_id": 0,
7
+ "eos_token_id": 2,
8
+ "gradient_checkpointing": false,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 3072,
14
+ "layer_norm_eps": 1e-05,
15
+ "max_position_embeddings": 514,
16
+ "model_type": "roberta",
17
+ "num_attention_heads": 12,
18
+ "num_hidden_layers": 12,
19
+ "pad_token_id": 1,
20
+ "type_vocab_size": 1,
21
+ "vocab_size": 50265
22
+ }
eval_results.txt ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ LRAP = 0.9511111111111111
2
+ eval_loss = 1.1304752279334425
3
+ macro_f1 = 0.8133553274682306
4
+ micro_f1 = 0.8133198789101919
5
+ weighted_f1 = 0.8179374389051809
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
model_args.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"adam_epsilon": 1e-08, "best_model_dir": "/shared/2/projects/framing/models/classify/Narrative/11-03-20_60_epochs_default_thresh_12_seed/best_model", "cache_dir": "cache_dir/", "config": {}, "custom_layer_parameters": [], "custom_parameter_groups": [], "dataloader_num_workers": 14, "do_lower_case": false, "dynamic_quantize": false, "early_stopping_consider_epochs": false, "early_stopping_delta": 0, "early_stopping_metric": "eval_loss", "early_stopping_metric_minimize": true, "early_stopping_patience": 20, "encoding": null, "eval_batch_size": 8, "evaluate_during_training": true, "evaluate_during_training_silent": true, "evaluate_during_training_steps": 100, "evaluate_during_training_verbose": false, "fp16": false, "gradient_accumulation_steps": 1, "learning_rate": 4e-05, "local_rank": -1, "logging_steps": 50, "manual_seed": 12, "max_grad_norm": 1.0, "max_seq_length": 128, "model_name": "/shared/2/projects/framing/models/finetune/roberta_cased_09-01-20", "model_type": "roberta", "multiprocessing_chunksize": 500, "n_gpu": 1, "no_cache": false, "no_save": false, "num_train_epochs": 60, "output_dir": "/shared/2/projects/framing/models/classify/Narrative/11-03-20_60_epochs_default_thresh_12_seed", "overwrite_output_dir": true, "process_count": 14, "quantized_model": false, "reprocess_input_data": true, "save_best_model": true, "save_eval_checkpoints": true, "save_model_every_epoch": true, "save_optimizer_and_scheduler": true, "save_steps": 2000, "silent": false, "tensorboard_dir": null, "thread_count": null, "train_batch_size": 8, "train_custom_parameters_only": false, "use_cached_eval_features": false, "use_early_stopping": true, "use_multiprocessing": true, "wandb_kwargs": {}, "wandb_project": null, "warmup_ratio": 0.06, "warmup_steps": 1620, "weight_decay": 0, "model_class": "MultiLabelClassificationModel", "sliding_window": false, "stride": 0.8, "threshold": 0.5, "tie_value": 1, "labels_list": [], "labels_map": {}, "lazy_loading": false}
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7182a22f4529200cb8721d9a30f573f54d2b5fd801ae283cf8292bd97b80adaa
3
+ size 501042459
special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"bos_token": "<s>", "eos_token": "</s>", "unk_token": "<unk>", "sep_token": "</s>", "pad_token": "<pad>", "cls_token": "<s>", "mask_token": "<mask>"}
tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"model_max_length": 512, "do_lower_case": false, "special_tokens_map_file": "/shared/2/projects/framing/models/finetune/roberta_cased_09-01-20/special_tokens_map.json", "full_tokenizer_file": null}
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7b0fcbc849a4c31b34a2ac685f8fa9e506c103a17fa21104895fb3b74634c95b
3
+ size 2479
training_progress_scores.csv ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ global_step,LRAP,train_loss,eval_loss,macro_f1,weighted_f1,micro_f1
2
+ 100,0.8688888888888889,0.6712647676467896,0.6655346887153492,0.629361647146451,0.6429480873405042,0.6604651162790698
3
+ 200,0.9022222222222223,0.622482180595398,0.6217048429606253,0.7149342891278375,0.7221650823333853,0.7140255009107467
4
+ 300,0.9211111111111111,0.645097017288208,0.564243764730922,0.747250821467689,0.7550740511452926,0.7442748091603052
5
+ 400,0.9433333333333334,0.4158073365688324,0.4596461151775561,0.8060647408391534,0.810685072013626,0.808870116156283
6
+ 450,0.9422222222222222,0.3783313035964966,0.4709558994100805,0.7914668045768958,0.7954028674459719,0.7925311203319503
7
+ 500,0.9488888888888889,0.19125902652740479,0.45353686155980094,0.7993172428788868,0.8032861586167126,0.8012486992715921
8
+ 600,0.9422222222222222,0.1630704402923584,0.4669703522272277,0.7968712889376086,0.8005667623847711,0.7982924226254001
9
+ 700,0.9444444444444444,0.22344070672988892,0.4686663888002697,0.8204055508930768,0.8245448370141117,0.8221993833504625
10
+ 800,0.95,0.08767425268888474,0.505050193975892,0.8055735662433097,0.8098452775976709,0.8065843621399177
11
+ 900,0.9344444444444444,0.4574398100376129,0.5292613248814616,0.802930543741412,0.8072047941873465,0.8028311425682507
12
+ 900,0.9344444444444444,0.4574398100376129,0.5292613248814616,0.802930543741412,0.8072047941873465,0.8028311425682507
13
+ 1000,0.9555555555555556,0.10432564467191696,0.5467683255933878,0.8145011850894204,0.8192593425329997,0.816580310880829
14
+ 1100,0.9366666666666666,0.04333396628499031,0.6997267740468184,0.7960599016880692,0.8010550225401721,0.7962962962962963
15
+ 1200,0.9477777777777778,0.15754951536655426,0.6409595489763377,0.8068247684585692,0.8118612268840607,0.81010101010101
16
+ 1300,0.9366666666666666,0.22498680651187897,0.838626898171609,0.7912846217585612,0.7930776858606029,0.7914893617021276
17
+ 1350,0.9533333333333334,0.2192697376012802,0.624743136920427,0.8270526726757312,0.8307004006781048,0.8299180327868853
18
+ 1400,0.9455555555555556,0.008564803749322891,0.6911759957540453,0.8140959679676233,0.8173227919009304,0.815734989648033
19
+ 1500,0.9511111111111111,0.13262632489204407,0.7170684868960004,0.8185130452507134,0.8229164820361936,0.8213552361396304
20
+ 1600,0.9555555555555556,0.02835235930979252,0.7308089047049483,0.8346854196137629,0.8381251585512342,0.836864406779661
21
+ 1700,0.9555555555555556,0.2004336565732956,0.7956759523796407,0.8368408232858361,0.841754168586149,0.8378650553877142
22
+ 1800,0.9566666666666667,0.20597997307777405,0.7102973188371643,0.8427885829754989,0.8469595967361104,0.8471337579617835
23
+ 1800,0.9566666666666667,0.20597997307777405,0.7102973188371643,0.8427885829754989,0.8469595967361104,0.8471337579617835
24
+ 1900,0.9555555555555556,0.009137887507677078,0.9947028399298066,0.8138726409888533,0.8202856964631027,0.8149568552253116
25
+ 2000,0.9522222222222222,0.28019455075263977,0.8701596499437041,0.8359751405246253,0.8405558287170655,0.8375893769152195
26
+ 2100,0.9511111111111111,0.0018466315232217312,0.9114740584249886,0.832292902715438,0.8360590895802165,0.8331606217616582
27
+ 2200,0.9433333333333334,0.4305149018764496,1.1260405475633186,0.8112835216731321,0.8154172197085806,0.8110882956878851
28
+ 2250,0.9588888888888889,0.002886707428842783,0.9006183988357518,0.834777455006845,0.8397805254439794,0.8382502543234995
29
+ 2300,0.9411111111111111,0.00110447162296623,1.0480767613146127,0.8279897216067429,0.8340268876809118,0.8296593186372746
30
+ 2400,0.9488888888888889,0.09333579987287521,1.0714588659946203,0.8278761264733209,0.8328137790540858,0.8285714285714286
31
+ 2500,0.9088888888888889,0.003728174138814211,1.496243638362278,0.7857108520610503,0.7912301510583488,0.7832167832167832
32
+ 2600,0.9511111111111111,0.0014892718754708767,1.1304752279334425,0.8133553274682306,0.8179374389051809,0.8133198789101919
vocab.json ADDED
The diff for this file is too large to render. See raw diff