pietrolesci commited on
Commit
af9eba7
1 Parent(s): 7afe7e8

Upload folder using huggingface_hub

Browse files
README.md ADDED
@@ -0,0 +1,48 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ## Run info
2
+
3
+ complete_hash: 53fb0af286257eb95c3298a4ff2761e0
4
+ short_hash: 53fb0761e0
5
+
6
+ ### Configuration
7
+
8
+ ```yaml
9
+ data:
10
+ batch_size: 32
11
+ data_seed: 42
12
+ drop_last: false
13
+ eval_batch_size: 128
14
+ max_length: 512
15
+ multiprocessing_context: null
16
+ num_workers: 8
17
+ persistent_workers: false
18
+ pin_memory: true
19
+ replacement: false
20
+ shuffle: true
21
+ dataset: mnli
22
+ estimator:
23
+ accelerator: gpu
24
+ convert_to_bettertransformer: false
25
+ deterministic: true
26
+ precision: bf16-true
27
+ tf32_mode: high
28
+ fit:
29
+ enable_progress_bar: true
30
+ limit_train_batches: null
31
+ limit_validation_batches: null
32
+ log_interval: 100
33
+ max_epochs: 20
34
+ optimizer_kwargs:
35
+ init_kwargs:
36
+ fused: true
37
+ lr: 3.0e-05
38
+ name: adamw
39
+ scheduler_kwargs:
40
+ name: constant_schedule_with_warmup
41
+ num_warmup_steps: 2000
42
+ model:
43
+ name: bert-base-uncased
44
+ revision: null
45
+ seed: 42
46
+ seed: 42
47
+
48
+ ```
hparams.yaml ADDED
@@ -0,0 +1,60 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ estimator:
2
+ accelerator: gpu
3
+ precision: bf16-true
4
+ deterministic: true
5
+ tf32_mode: high
6
+ convert_to_bettertransformer: false
7
+ callbacks:
8
+ timer:
9
+ _target_: energizer.active_learning.callbacks.Timer
10
+ lr_monitor:
11
+ _target_: energizer.callbacks.lr_monitor.LearningRateMonitor
12
+ model_checkpoint:
13
+ _target_: energizer.callbacks.model_checkpoint.ModelCheckpoint
14
+ dirpath: .checkpoints
15
+ stage: train
16
+ frequency: 1:epoch
17
+ loggers:
18
+ tensorboard:
19
+ _target_: energizer.loggers.TensorBoardLogger
20
+ root_dir: ./
21
+ name: tb_logs
22
+ version: null
23
+ data:
24
+ batch_size: 32
25
+ eval_batch_size: 128
26
+ shuffle: true
27
+ replacement: false
28
+ data_seed: 42
29
+ drop_last: false
30
+ num_workers: 8
31
+ pin_memory: true
32
+ persistent_workers: false
33
+ multiprocessing_context: null
34
+ max_length: 512
35
+ fit:
36
+ max_epochs: 20
37
+ optimizer_kwargs:
38
+ name: adamw
39
+ lr: 3.0e-05
40
+ init_kwargs:
41
+ fused: true
42
+ scheduler_kwargs:
43
+ name: constant_schedule_with_warmup
44
+ num_warmup_steps: 2000
45
+ log_interval: 100
46
+ enable_progress_bar: true
47
+ limit_train_batches: null
48
+ limit_validation_batches: null
49
+ model:
50
+ name: bert-base-uncased
51
+ revision: null
52
+ seed: 42
53
+ log_interval: 100
54
+ enable_progress_bar: true
55
+ limit_batches: null
56
+ seed: 42
57
+ experiment_group: training
58
+ run_name: bert-base-uncased_2024-02-28T14-37-16
59
+ data_path: /home/pl487/coreset-project/data/processed
60
+ dataset: mnli
tb_logs/version_0/events.out.tfevents.1709131048.dev-gpu-pl487.689954.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:86784ec596a8ef052c2556c40735fe550329af65024345d19579f0ac63e138e7
3
+ size 404506
tensorboard_logs.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:be203a8ce81ef65a419144cf03ac657d433d12be70c87bad3e51eb3438ddc2ca
3
+ size 40873
train.log ADDED
@@ -0,0 +1,77 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2024-02-28 14:37:17,121][hydra][INFO] -
2
+ estimator:
3
+ accelerator: gpu
4
+ precision: bf16-true
5
+ deterministic: true
6
+ tf32_mode: high
7
+ convert_to_bettertransformer: false
8
+ callbacks:
9
+ timer:
10
+ _target_: energizer.active_learning.callbacks.Timer
11
+ lr_monitor:
12
+ _target_: energizer.callbacks.lr_monitor.LearningRateMonitor
13
+ model_checkpoint:
14
+ _target_: energizer.callbacks.model_checkpoint.ModelCheckpoint
15
+ dirpath: .checkpoints
16
+ stage: train
17
+ frequency: 1:epoch
18
+ loggers:
19
+ tensorboard:
20
+ _target_: energizer.loggers.TensorBoardLogger
21
+ root_dir: ./
22
+ name: tb_logs
23
+ version: null
24
+ data:
25
+ batch_size: 32
26
+ eval_batch_size: 128
27
+ shuffle: true
28
+ replacement: false
29
+ data_seed: 42
30
+ drop_last: false
31
+ num_workers: 8
32
+ pin_memory: true
33
+ persistent_workers: false
34
+ multiprocessing_context: null
35
+ max_length: 512
36
+ fit:
37
+ max_epochs: 20
38
+ optimizer_kwargs:
39
+ name: adamw
40
+ lr: 3.0e-05
41
+ init_kwargs:
42
+ fused: true
43
+ scheduler_kwargs:
44
+ name: constant_schedule_with_warmup
45
+ num_warmup_steps: 2000
46
+ log_interval: 100
47
+ enable_progress_bar: true
48
+ limit_train_batches: null
49
+ limit_validation_batches: null
50
+ model:
51
+ name: bert-base-uncased
52
+ revision: null
53
+ seed: 42
54
+ log_interval: 100
55
+ enable_progress_bar: true
56
+ limit_batches: null
57
+ seed: 42
58
+ experiment_group: training
59
+ run_name: bert-base-uncased_2024-02-28T14-37-16
60
+ data_path: /home/pl487/coreset-project/data/processed
61
+ dataset: mnli
62
+
63
+ ======================================================================
64
+ [2024-02-28 14:37:17,122][hydra][INFO] - Seed enabled: 42
65
+ [2024-02-28 14:37:17,496][hydra][INFO] - Label distribution:
66
+ {<RunningStage.TRAIN: 'train'>: {'0-(entailment)': 130899, '1-(neutral)': 130900, '2-(contradiction)': 130903}}
67
+ [2024-02-28 14:37:20,088][hydra][INFO] - Loggers: [<energizer.loggers.tensorboard.TensorBoardLogger object at 0x7efde1ce9ea0>]
68
+ [2024-02-28 14:37:20,088][hydra][INFO] - Callbacks: [<energizer.active_learning.callbacks.Timer object at 0x7efde11b6770>, <energizer.callbacks.lr_monitor.LearningRateMonitor object at 0x7efde11b67d0>, <energizer.callbacks.model_checkpoint.ModelCheckpoint object at 0x7efde11b5cf0>]
69
+ [2024-02-28 14:37:20,091][hydra][INFO] - Model summary:
70
+ Total num params: 109.5M
71
+ Of which trainable: 109.5M
72
+ With a memory footprint of 0.22GB
73
+ Total memory allocated 0.68GB
74
+ [2024-02-28 14:37:20,092][hydra][INFO] - Dataloading params:
75
+ SequenceClassificationDataloaderArgs(batch_size=32, eval_batch_size=128, num_workers=8, pin_memory=True, drop_last=False, persistent_workers=False, shuffle=True, replacement=False, data_seed=42, multiprocessing_context=None, max_length=512)
76
+ [2024-02-28 14:37:20,095][hydra][INFO] - Batch:
77
+ {<InputKeys.INPUT_IDS: 'input_ids'>: tensor([[ 101, 17842, 102, 5368, 102]]), <InputKeys.ATT_MASK: 'attention_mask'>: tensor([[1, 1, 1, 1, 1]]), <InputKeys.LABELS: 'labels'>: tensor([0]), <InputKeys.ON_CPU: 'on_cpu'>: {<SpecialKeys.ID: 'uid'>: [278780]}}