pietrolesci
commited on
Commit
•
af9eba7
1
Parent(s):
7afe7e8
Upload folder using huggingface_hub
Browse files- README.md +48 -0
- hparams.yaml +60 -0
- tb_logs/version_0/events.out.tfevents.1709131048.dev-gpu-pl487.689954.0 +3 -0
- tensorboard_logs.parquet +3 -0
- train.log +77 -0
README.md
ADDED
@@ -0,0 +1,48 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
## Run info
|
2 |
+
|
3 |
+
complete_hash: 53fb0af286257eb95c3298a4ff2761e0
|
4 |
+
short_hash: 53fb0761e0
|
5 |
+
|
6 |
+
### Configuration
|
7 |
+
|
8 |
+
```yaml
|
9 |
+
data:
|
10 |
+
batch_size: 32
|
11 |
+
data_seed: 42
|
12 |
+
drop_last: false
|
13 |
+
eval_batch_size: 128
|
14 |
+
max_length: 512
|
15 |
+
multiprocessing_context: null
|
16 |
+
num_workers: 8
|
17 |
+
persistent_workers: false
|
18 |
+
pin_memory: true
|
19 |
+
replacement: false
|
20 |
+
shuffle: true
|
21 |
+
dataset: mnli
|
22 |
+
estimator:
|
23 |
+
accelerator: gpu
|
24 |
+
convert_to_bettertransformer: false
|
25 |
+
deterministic: true
|
26 |
+
precision: bf16-true
|
27 |
+
tf32_mode: high
|
28 |
+
fit:
|
29 |
+
enable_progress_bar: true
|
30 |
+
limit_train_batches: null
|
31 |
+
limit_validation_batches: null
|
32 |
+
log_interval: 100
|
33 |
+
max_epochs: 20
|
34 |
+
optimizer_kwargs:
|
35 |
+
init_kwargs:
|
36 |
+
fused: true
|
37 |
+
lr: 3.0e-05
|
38 |
+
name: adamw
|
39 |
+
scheduler_kwargs:
|
40 |
+
name: constant_schedule_with_warmup
|
41 |
+
num_warmup_steps: 2000
|
42 |
+
model:
|
43 |
+
name: bert-base-uncased
|
44 |
+
revision: null
|
45 |
+
seed: 42
|
46 |
+
seed: 42
|
47 |
+
|
48 |
+
```
|
hparams.yaml
ADDED
@@ -0,0 +1,60 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
estimator:
|
2 |
+
accelerator: gpu
|
3 |
+
precision: bf16-true
|
4 |
+
deterministic: true
|
5 |
+
tf32_mode: high
|
6 |
+
convert_to_bettertransformer: false
|
7 |
+
callbacks:
|
8 |
+
timer:
|
9 |
+
_target_: energizer.active_learning.callbacks.Timer
|
10 |
+
lr_monitor:
|
11 |
+
_target_: energizer.callbacks.lr_monitor.LearningRateMonitor
|
12 |
+
model_checkpoint:
|
13 |
+
_target_: energizer.callbacks.model_checkpoint.ModelCheckpoint
|
14 |
+
dirpath: .checkpoints
|
15 |
+
stage: train
|
16 |
+
frequency: 1:epoch
|
17 |
+
loggers:
|
18 |
+
tensorboard:
|
19 |
+
_target_: energizer.loggers.TensorBoardLogger
|
20 |
+
root_dir: ./
|
21 |
+
name: tb_logs
|
22 |
+
version: null
|
23 |
+
data:
|
24 |
+
batch_size: 32
|
25 |
+
eval_batch_size: 128
|
26 |
+
shuffle: true
|
27 |
+
replacement: false
|
28 |
+
data_seed: 42
|
29 |
+
drop_last: false
|
30 |
+
num_workers: 8
|
31 |
+
pin_memory: true
|
32 |
+
persistent_workers: false
|
33 |
+
multiprocessing_context: null
|
34 |
+
max_length: 512
|
35 |
+
fit:
|
36 |
+
max_epochs: 20
|
37 |
+
optimizer_kwargs:
|
38 |
+
name: adamw
|
39 |
+
lr: 3.0e-05
|
40 |
+
init_kwargs:
|
41 |
+
fused: true
|
42 |
+
scheduler_kwargs:
|
43 |
+
name: constant_schedule_with_warmup
|
44 |
+
num_warmup_steps: 2000
|
45 |
+
log_interval: 100
|
46 |
+
enable_progress_bar: true
|
47 |
+
limit_train_batches: null
|
48 |
+
limit_validation_batches: null
|
49 |
+
model:
|
50 |
+
name: bert-base-uncased
|
51 |
+
revision: null
|
52 |
+
seed: 42
|
53 |
+
log_interval: 100
|
54 |
+
enable_progress_bar: true
|
55 |
+
limit_batches: null
|
56 |
+
seed: 42
|
57 |
+
experiment_group: training
|
58 |
+
run_name: bert-base-uncased_2024-02-28T14-37-16
|
59 |
+
data_path: /home/pl487/coreset-project/data/processed
|
60 |
+
dataset: mnli
|
tb_logs/version_0/events.out.tfevents.1709131048.dev-gpu-pl487.689954.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:86784ec596a8ef052c2556c40735fe550329af65024345d19579f0ac63e138e7
|
3 |
+
size 404506
|
tensorboard_logs.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:be203a8ce81ef65a419144cf03ac657d433d12be70c87bad3e51eb3438ddc2ca
|
3 |
+
size 40873
|
train.log
ADDED
@@ -0,0 +1,77 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
[2024-02-28 14:37:17,121][hydra][INFO] -
|
2 |
+
estimator:
|
3 |
+
accelerator: gpu
|
4 |
+
precision: bf16-true
|
5 |
+
deterministic: true
|
6 |
+
tf32_mode: high
|
7 |
+
convert_to_bettertransformer: false
|
8 |
+
callbacks:
|
9 |
+
timer:
|
10 |
+
_target_: energizer.active_learning.callbacks.Timer
|
11 |
+
lr_monitor:
|
12 |
+
_target_: energizer.callbacks.lr_monitor.LearningRateMonitor
|
13 |
+
model_checkpoint:
|
14 |
+
_target_: energizer.callbacks.model_checkpoint.ModelCheckpoint
|
15 |
+
dirpath: .checkpoints
|
16 |
+
stage: train
|
17 |
+
frequency: 1:epoch
|
18 |
+
loggers:
|
19 |
+
tensorboard:
|
20 |
+
_target_: energizer.loggers.TensorBoardLogger
|
21 |
+
root_dir: ./
|
22 |
+
name: tb_logs
|
23 |
+
version: null
|
24 |
+
data:
|
25 |
+
batch_size: 32
|
26 |
+
eval_batch_size: 128
|
27 |
+
shuffle: true
|
28 |
+
replacement: false
|
29 |
+
data_seed: 42
|
30 |
+
drop_last: false
|
31 |
+
num_workers: 8
|
32 |
+
pin_memory: true
|
33 |
+
persistent_workers: false
|
34 |
+
multiprocessing_context: null
|
35 |
+
max_length: 512
|
36 |
+
fit:
|
37 |
+
max_epochs: 20
|
38 |
+
optimizer_kwargs:
|
39 |
+
name: adamw
|
40 |
+
lr: 3.0e-05
|
41 |
+
init_kwargs:
|
42 |
+
fused: true
|
43 |
+
scheduler_kwargs:
|
44 |
+
name: constant_schedule_with_warmup
|
45 |
+
num_warmup_steps: 2000
|
46 |
+
log_interval: 100
|
47 |
+
enable_progress_bar: true
|
48 |
+
limit_train_batches: null
|
49 |
+
limit_validation_batches: null
|
50 |
+
model:
|
51 |
+
name: bert-base-uncased
|
52 |
+
revision: null
|
53 |
+
seed: 42
|
54 |
+
log_interval: 100
|
55 |
+
enable_progress_bar: true
|
56 |
+
limit_batches: null
|
57 |
+
seed: 42
|
58 |
+
experiment_group: training
|
59 |
+
run_name: bert-base-uncased_2024-02-28T14-37-16
|
60 |
+
data_path: /home/pl487/coreset-project/data/processed
|
61 |
+
dataset: mnli
|
62 |
+
|
63 |
+
======================================================================
|
64 |
+
[2024-02-28 14:37:17,122][hydra][INFO] - Seed enabled: 42
|
65 |
+
[2024-02-28 14:37:17,496][hydra][INFO] - Label distribution:
|
66 |
+
{<RunningStage.TRAIN: 'train'>: {'0-(entailment)': 130899, '1-(neutral)': 130900, '2-(contradiction)': 130903}}
|
67 |
+
[2024-02-28 14:37:20,088][hydra][INFO] - Loggers: [<energizer.loggers.tensorboard.TensorBoardLogger object at 0x7efde1ce9ea0>]
|
68 |
+
[2024-02-28 14:37:20,088][hydra][INFO] - Callbacks: [<energizer.active_learning.callbacks.Timer object at 0x7efde11b6770>, <energizer.callbacks.lr_monitor.LearningRateMonitor object at 0x7efde11b67d0>, <energizer.callbacks.model_checkpoint.ModelCheckpoint object at 0x7efde11b5cf0>]
|
69 |
+
[2024-02-28 14:37:20,091][hydra][INFO] - Model summary:
|
70 |
+
Total num params: 109.5M
|
71 |
+
Of which trainable: 109.5M
|
72 |
+
With a memory footprint of 0.22GB
|
73 |
+
Total memory allocated 0.68GB
|
74 |
+
[2024-02-28 14:37:20,092][hydra][INFO] - Dataloading params:
|
75 |
+
SequenceClassificationDataloaderArgs(batch_size=32, eval_batch_size=128, num_workers=8, pin_memory=True, drop_last=False, persistent_workers=False, shuffle=True, replacement=False, data_seed=42, multiprocessing_context=None, max_length=512)
|
76 |
+
[2024-02-28 14:37:20,095][hydra][INFO] - Batch:
|
77 |
+
{<InputKeys.INPUT_IDS: 'input_ids'>: tensor([[ 101, 17842, 102, 5368, 102]]), <InputKeys.ATT_MASK: 'attention_mask'>: tensor([[1, 1, 1, 1, 1]]), <InputKeys.LABELS: 'labels'>: tensor([0]), <InputKeys.ON_CPU: 'on_cpu'>: {<SpecialKeys.ID: 'uid'>: [278780]}}
|