juliamendelsohn
commited on
Commit
•
d2de03f
1
Parent(s):
a3fbb71
commit from juliame
Browse files- config.json +22 -0
- eval_results.txt +5 -0
- merges.txt +0 -0
- model_args.json +1 -0
- pytorch_model.bin +3 -0
- special_tokens_map.json +1 -0
- tokenizer_config.json +1 -0
- training_args.bin +3 -0
- training_progress_scores.csv +32 -0
- vocab.json +0 -0
config.json
ADDED
@@ -0,0 +1,22 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"architectures": [
|
3 |
+
"RobertaForMultiLabelSequenceClassification"
|
4 |
+
],
|
5 |
+
"attention_probs_dropout_prob": 0.1,
|
6 |
+
"bos_token_id": 0,
|
7 |
+
"eos_token_id": 2,
|
8 |
+
"gradient_checkpointing": false,
|
9 |
+
"hidden_act": "gelu",
|
10 |
+
"hidden_dropout_prob": 0.1,
|
11 |
+
"hidden_size": 768,
|
12 |
+
"initializer_range": 0.02,
|
13 |
+
"intermediate_size": 3072,
|
14 |
+
"layer_norm_eps": 1e-05,
|
15 |
+
"max_position_embeddings": 514,
|
16 |
+
"model_type": "roberta",
|
17 |
+
"num_attention_heads": 12,
|
18 |
+
"num_hidden_layers": 12,
|
19 |
+
"pad_token_id": 1,
|
20 |
+
"type_vocab_size": 1,
|
21 |
+
"vocab_size": 50265
|
22 |
+
}
|
eval_results.txt
ADDED
@@ -0,0 +1,5 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
LRAP = 0.9511111111111111
|
2 |
+
eval_loss = 1.1304752279334425
|
3 |
+
macro_f1 = 0.8133553274682306
|
4 |
+
micro_f1 = 0.8133198789101919
|
5 |
+
weighted_f1 = 0.8179374389051809
|
merges.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|
model_args.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"adam_epsilon": 1e-08, "best_model_dir": "/shared/2/projects/framing/models/classify/Narrative/11-03-20_60_epochs_default_thresh_12_seed/best_model", "cache_dir": "cache_dir/", "config": {}, "custom_layer_parameters": [], "custom_parameter_groups": [], "dataloader_num_workers": 14, "do_lower_case": false, "dynamic_quantize": false, "early_stopping_consider_epochs": false, "early_stopping_delta": 0, "early_stopping_metric": "eval_loss", "early_stopping_metric_minimize": true, "early_stopping_patience": 20, "encoding": null, "eval_batch_size": 8, "evaluate_during_training": true, "evaluate_during_training_silent": true, "evaluate_during_training_steps": 100, "evaluate_during_training_verbose": false, "fp16": false, "gradient_accumulation_steps": 1, "learning_rate": 4e-05, "local_rank": -1, "logging_steps": 50, "manual_seed": 12, "max_grad_norm": 1.0, "max_seq_length": 128, "model_name": "/shared/2/projects/framing/models/finetune/roberta_cased_09-01-20", "model_type": "roberta", "multiprocessing_chunksize": 500, "n_gpu": 1, "no_cache": false, "no_save": false, "num_train_epochs": 60, "output_dir": "/shared/2/projects/framing/models/classify/Narrative/11-03-20_60_epochs_default_thresh_12_seed", "overwrite_output_dir": true, "process_count": 14, "quantized_model": false, "reprocess_input_data": true, "save_best_model": true, "save_eval_checkpoints": true, "save_model_every_epoch": true, "save_optimizer_and_scheduler": true, "save_steps": 2000, "silent": false, "tensorboard_dir": null, "thread_count": null, "train_batch_size": 8, "train_custom_parameters_only": false, "use_cached_eval_features": false, "use_early_stopping": true, "use_multiprocessing": true, "wandb_kwargs": {}, "wandb_project": null, "warmup_ratio": 0.06, "warmup_steps": 1620, "weight_decay": 0, "model_class": "MultiLabelClassificationModel", "sliding_window": false, "stride": 0.8, "threshold": 0.5, "tie_value": 1, "labels_list": [], "labels_map": {}, "lazy_loading": false}
|
pytorch_model.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7182a22f4529200cb8721d9a30f573f54d2b5fd801ae283cf8292bd97b80adaa
|
3 |
+
size 501042459
|
special_tokens_map.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"bos_token": "<s>", "eos_token": "</s>", "unk_token": "<unk>", "sep_token": "</s>", "pad_token": "<pad>", "cls_token": "<s>", "mask_token": "<mask>"}
|
tokenizer_config.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"model_max_length": 512, "do_lower_case": false, "special_tokens_map_file": "/shared/2/projects/framing/models/finetune/roberta_cased_09-01-20/special_tokens_map.json", "full_tokenizer_file": null}
|
training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7b0fcbc849a4c31b34a2ac685f8fa9e506c103a17fa21104895fb3b74634c95b
|
3 |
+
size 2479
|
training_progress_scores.csv
ADDED
@@ -0,0 +1,32 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
global_step,LRAP,train_loss,eval_loss,macro_f1,weighted_f1,micro_f1
|
2 |
+
100,0.8688888888888889,0.6712647676467896,0.6655346887153492,0.629361647146451,0.6429480873405042,0.6604651162790698
|
3 |
+
200,0.9022222222222223,0.622482180595398,0.6217048429606253,0.7149342891278375,0.7221650823333853,0.7140255009107467
|
4 |
+
300,0.9211111111111111,0.645097017288208,0.564243764730922,0.747250821467689,0.7550740511452926,0.7442748091603052
|
5 |
+
400,0.9433333333333334,0.4158073365688324,0.4596461151775561,0.8060647408391534,0.810685072013626,0.808870116156283
|
6 |
+
450,0.9422222222222222,0.3783313035964966,0.4709558994100805,0.7914668045768958,0.7954028674459719,0.7925311203319503
|
7 |
+
500,0.9488888888888889,0.19125902652740479,0.45353686155980094,0.7993172428788868,0.8032861586167126,0.8012486992715921
|
8 |
+
600,0.9422222222222222,0.1630704402923584,0.4669703522272277,0.7968712889376086,0.8005667623847711,0.7982924226254001
|
9 |
+
700,0.9444444444444444,0.22344070672988892,0.4686663888002697,0.8204055508930768,0.8245448370141117,0.8221993833504625
|
10 |
+
800,0.95,0.08767425268888474,0.505050193975892,0.8055735662433097,0.8098452775976709,0.8065843621399177
|
11 |
+
900,0.9344444444444444,0.4574398100376129,0.5292613248814616,0.802930543741412,0.8072047941873465,0.8028311425682507
|
12 |
+
900,0.9344444444444444,0.4574398100376129,0.5292613248814616,0.802930543741412,0.8072047941873465,0.8028311425682507
|
13 |
+
1000,0.9555555555555556,0.10432564467191696,0.5467683255933878,0.8145011850894204,0.8192593425329997,0.816580310880829
|
14 |
+
1100,0.9366666666666666,0.04333396628499031,0.6997267740468184,0.7960599016880692,0.8010550225401721,0.7962962962962963
|
15 |
+
1200,0.9477777777777778,0.15754951536655426,0.6409595489763377,0.8068247684585692,0.8118612268840607,0.81010101010101
|
16 |
+
1300,0.9366666666666666,0.22498680651187897,0.838626898171609,0.7912846217585612,0.7930776858606029,0.7914893617021276
|
17 |
+
1350,0.9533333333333334,0.2192697376012802,0.624743136920427,0.8270526726757312,0.8307004006781048,0.8299180327868853
|
18 |
+
1400,0.9455555555555556,0.008564803749322891,0.6911759957540453,0.8140959679676233,0.8173227919009304,0.815734989648033
|
19 |
+
1500,0.9511111111111111,0.13262632489204407,0.7170684868960004,0.8185130452507134,0.8229164820361936,0.8213552361396304
|
20 |
+
1600,0.9555555555555556,0.02835235930979252,0.7308089047049483,0.8346854196137629,0.8381251585512342,0.836864406779661
|
21 |
+
1700,0.9555555555555556,0.2004336565732956,0.7956759523796407,0.8368408232858361,0.841754168586149,0.8378650553877142
|
22 |
+
1800,0.9566666666666667,0.20597997307777405,0.7102973188371643,0.8427885829754989,0.8469595967361104,0.8471337579617835
|
23 |
+
1800,0.9566666666666667,0.20597997307777405,0.7102973188371643,0.8427885829754989,0.8469595967361104,0.8471337579617835
|
24 |
+
1900,0.9555555555555556,0.009137887507677078,0.9947028399298066,0.8138726409888533,0.8202856964631027,0.8149568552253116
|
25 |
+
2000,0.9522222222222222,0.28019455075263977,0.8701596499437041,0.8359751405246253,0.8405558287170655,0.8375893769152195
|
26 |
+
2100,0.9511111111111111,0.0018466315232217312,0.9114740584249886,0.832292902715438,0.8360590895802165,0.8331606217616582
|
27 |
+
2200,0.9433333333333334,0.4305149018764496,1.1260405475633186,0.8112835216731321,0.8154172197085806,0.8110882956878851
|
28 |
+
2250,0.9588888888888889,0.002886707428842783,0.9006183988357518,0.834777455006845,0.8397805254439794,0.8382502543234995
|
29 |
+
2300,0.9411111111111111,0.00110447162296623,1.0480767613146127,0.8279897216067429,0.8340268876809118,0.8296593186372746
|
30 |
+
2400,0.9488888888888889,0.09333579987287521,1.0714588659946203,0.8278761264733209,0.8328137790540858,0.8285714285714286
|
31 |
+
2500,0.9088888888888889,0.003728174138814211,1.496243638362278,0.7857108520610503,0.7912301510583488,0.7832167832167832
|
32 |
+
2600,0.9511111111111111,0.0014892718754708767,1.1304752279334425,0.8133553274682306,0.8179374389051809,0.8133198789101919
|
vocab.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|