Femboyuwu2000 commited on
Commit
f71cf9b
1 Parent(s): c59484b

falcon1b_h_story

Browse files
README.md CHANGED
@@ -5,7 +5,7 @@ tags:
5
  - trl
6
  - sft
7
  - generated_from_trainer
8
- base_model: TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T
9
  model-index:
10
  - name: working
11
  results: []
@@ -16,7 +16,7 @@ should probably proofread and complete it, then remove this comment. -->
16
 
17
  # working
18
 
19
- This model is a fine-tuned version of [TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T](https://huggingface.co/TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T) on an unknown dataset.
20
 
21
  ## Model description
22
 
@@ -36,15 +36,13 @@ More information needed
36
 
37
  The following hyperparameters were used during training:
38
  - learning_rate: 3e-05
39
- - train_batch_size: 8
40
- - eval_batch_size: 8
41
  - seed: 42
42
- - gradient_accumulation_steps: 4
43
- - total_train_batch_size: 32
44
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
  - lr_scheduler_type: cosine
46
- - lr_scheduler_warmup_ratio: 0.05
47
- - training_steps: 10
48
 
49
  ### Training results
50
 
 
5
  - trl
6
  - sft
7
  - generated_from_trainer
8
+ base_model: tiiuae/falcon-rw-1b
9
  model-index:
10
  - name: working
11
  results: []
 
16
 
17
  # working
18
 
19
+ This model is a fine-tuned version of [tiiuae/falcon-rw-1b](https://huggingface.co/tiiuae/falcon-rw-1b) on an unknown dataset.
20
 
21
  ## Model description
22
 
 
36
 
37
  The following hyperparameters were used during training:
38
  - learning_rate: 3e-05
39
+ - train_batch_size: 16
40
+ - eval_batch_size: 16
41
  - seed: 42
 
 
42
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
43
  - lr_scheduler_type: cosine
44
+ - lr_scheduler_warmup_ratio: 0.03
45
+ - training_steps: 20
46
 
47
  ### Training results
48
 
adapter_config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
- "base_model_name_or_path": "TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T",
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
@@ -10,7 +10,7 @@
10
  "layers_pattern": null,
11
  "layers_to_transform": null,
12
  "loftq_config": {},
13
- "lora_alpha": 64,
14
  "lora_dropout": 0.01,
15
  "megatron_config": null,
16
  "megatron_core": "megatron.core",
@@ -20,15 +20,13 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "up_proj",
24
- "gate_proj",
25
- "down_proj",
26
- "q_proj",
27
- "k_proj",
28
- "o_proj",
29
- "v_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
33
- "use_rslora": false
34
  }
 
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
+ "base_model_name_or_path": "tiiuae/falcon-rw-1b",
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
 
10
  "layers_pattern": null,
11
  "layers_to_transform": null,
12
  "loftq_config": {},
13
+ "lora_alpha": 16,
14
  "lora_dropout": 0.01,
15
  "megatron_config": null,
16
  "megatron_core": "megatron.core",
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
+ "dense",
24
+ "query_key_value",
25
+ "lm_head",
26
+ "dense_h_to_4h",
27
+ "word_embeddings"
 
 
28
  ],
29
  "task_type": "CAUSAL_LM",
30
  "use_dora": false,
31
+ "use_rslora": true
32
  }
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:93c117436ccb37985fb8a42daf5f2c6bdb4eb43a9625f8c045a2ceccc233c581
3
- size 25271744
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d9aea127eb50666b10f052869f64e016f82c077f73c58abb8489889d039a0447
3
+ size 844854136
runs/Apr10_21-42-21_4bf7c58fba03/events.out.tfevents.1712785831.4bf7c58fba03.227.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:18b743e6893e182f3f924d7175c1d6deeaa7c8ab85dcdf66048a82bae84f27cb
3
+ size 5679
runs/Apr10_21-52-40_4bf7c58fba03/events.out.tfevents.1712786444.4bf7c58fba03.227.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bee9179f34f9b37bd36b80d35f140c3cc45a06567ca2e12ae7ba94dea3f762b7
3
+ size 17416
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -1,6 +1,5 @@
1
  {
2
- "add_bos_token": true,
3
- "add_eos_token": false,
4
  "added_tokens_decoder": {
5
  "0": {
6
  "content": "<unk>",
@@ -25,17 +24,24 @@
25
  "rstrip": false,
26
  "single_word": false,
27
  "special": true
 
 
 
 
 
 
 
 
28
  }
29
  },
30
  "bos_token": "<s>",
31
  "clean_up_tokenization_spaces": false,
32
  "eos_token": "</s>",
33
- "legacy": false,
34
- "model_max_length": 1000000000000000019884624838656,
35
  "pad_token": "</s>",
36
  "padding_side": "right",
37
- "sp_model_kwargs": {},
38
- "tokenizer_class": "LlamaTokenizer",
39
- "unk_token": "<unk>",
40
- "use_default_system_prompt": false
41
  }
 
1
  {
2
+ "add_prefix_space": false,
 
3
  "added_tokens_decoder": {
4
  "0": {
5
  "content": "<unk>",
 
24
  "rstrip": false,
25
  "single_word": false,
26
  "special": true
27
+ },
28
+ "3": {
29
+ "content": "<pad>",
30
+ "lstrip": false,
31
+ "normalized": false,
32
+ "rstrip": false,
33
+ "single_word": false,
34
+ "special": true
35
  }
36
  },
37
  "bos_token": "<s>",
38
  "clean_up_tokenization_spaces": false,
39
  "eos_token": "</s>",
40
+ "max_length": 2048,
41
+ "model_max_length": 8192,
42
  "pad_token": "</s>",
43
  "padding_side": "right",
44
+ "tokenizer_class": "BloomTokenizer",
45
+ "truncation": true,
46
+ "unk_token": "<unk>"
 
47
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:13609889d8491533770ceab667ce50ea137267ebb81ebecd751651b23e13c7ab
3
  size 4920
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:71013af53dd90b2f29c64c2465ed2e7f850cd56a40c3c90ec075b87d0cf5e42c
3
  size 4920
wandb/debug-internal.log CHANGED
@@ -1,32 +1,32 @@
1
- 2024-04-10 21:00:30,782 INFO StreamThr :603 [internal.py:wandb_internal():86] W&B internal server running at pid: 603, started at: 2024-04-10 21:00:30.781407
2
- 2024-04-10 21:00:30,783 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: status
3
- 2024-04-10 21:00:31,093 INFO WriterThread:603 [datastore.py:open_for_write():87] open: /kaggle/working/wandb/run-20240410_210030-g0qpa26d/run-g0qpa26d.wandb
4
- 2024-04-10 21:00:31,094 DEBUG SenderThread:603 [sender.py:send():379] send: header
5
- 2024-04-10 21:00:31,097 DEBUG SenderThread:603 [sender.py:send():379] send: run
6
- 2024-04-10 21:00:31,284 INFO SenderThread:603 [dir_watcher.py:__init__():211] watching files in: /kaggle/working/wandb/run-20240410_210030-g0qpa26d/files
7
- 2024-04-10 21:00:31,284 INFO SenderThread:603 [sender.py:_start_run_threads():1124] run started: g0qpa26d with start time 1712782830.78121
8
- 2024-04-10 21:00:31,291 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: check_version
9
- 2024-04-10 21:00:31,292 DEBUG SenderThread:603 [sender.py:send_request():406] send_request: check_version
10
- 2024-04-10 21:00:31,382 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: run_start
11
- 2024-04-10 21:00:31,393 DEBUG HandlerThread:603 [system_info.py:__init__():26] System info init
12
- 2024-04-10 21:00:31,394 DEBUG HandlerThread:603 [system_info.py:__init__():41] System info init done
13
- 2024-04-10 21:00:31,394 INFO HandlerThread:603 [system_monitor.py:start():194] Starting system monitor
14
- 2024-04-10 21:00:31,394 INFO SystemMonitor:603 [system_monitor.py:_start():158] Starting system asset monitoring threads
15
- 2024-04-10 21:00:31,394 INFO HandlerThread:603 [system_monitor.py:probe():214] Collecting system info
16
- 2024-04-10 21:00:31,395 INFO SystemMonitor:603 [interfaces.py:start():190] Started cpu monitoring
17
- 2024-04-10 21:00:31,396 INFO SystemMonitor:603 [interfaces.py:start():190] Started disk monitoring
18
- 2024-04-10 21:00:31,396 INFO SystemMonitor:603 [interfaces.py:start():190] Started gpu monitoring
19
- 2024-04-10 21:00:31,398 INFO SystemMonitor:603 [interfaces.py:start():190] Started memory monitoring
20
- 2024-04-10 21:00:31,399 INFO SystemMonitor:603 [interfaces.py:start():190] Started network monitoring
21
- 2024-04-10 21:00:31,409 DEBUG HandlerThread:603 [system_info.py:probe():150] Probing system
22
- 2024-04-10 21:00:31,411 DEBUG HandlerThread:603 [gitlib.py:_init_repo():56] git repository is invalid
23
- 2024-04-10 21:00:31,411 DEBUG HandlerThread:603 [system_info.py:probe():198] Probing system done
24
- 2024-04-10 21:00:31,411 DEBUG HandlerThread:603 [system_monitor.py:probe():223] {'os': 'Linux-5.15.133+-x86_64-with-glibc2.31', 'python': '3.10.13', 'heartbeatAt': '2024-04-10T21:00:31.409648', 'startedAt': '2024-04-10T21:00:30.774771', 'docker': None, 'cuda': None, 'args': (), 'state': 'running', 'program': 'kaggle.ipynb', 'codePathLocal': None, 'root': '/kaggle/working', 'host': '57bca537cf5f', 'username': 'root', 'executable': '/opt/conda/bin/python3.10', 'cpu_count': 2, 'cpu_count_logical': 4, 'cpu_freq': {'current': 2000.142, 'min': 0.0, 'max': 0.0}, 'cpu_freq_per_core': [{'current': 2000.142, 'min': 0.0, 'max': 0.0}, {'current': 2000.142, 'min': 0.0, 'max': 0.0}, {'current': 2000.142, 'min': 0.0, 'max': 0.0}, {'current': 2000.142, 'min': 0.0, 'max': 0.0}], 'disk': {'/': {'total': 8062.387607574463, 'used': 5569.839115142822}}, 'gpu': 'Tesla T4', 'gpu_count': 2, 'gpu_devices': [{'name': 'Tesla T4', 'memory_total': 16106127360}, {'name': 'Tesla T4', 'memory_total': 16106127360}], 'memory': {'total': 31.357559204101562}}
25
- 2024-04-10 21:00:31,411 INFO HandlerThread:603 [system_monitor.py:probe():224] Finished collecting system info
26
- 2024-04-10 21:00:31,412 INFO HandlerThread:603 [system_monitor.py:probe():227] Publishing system info
27
- 2024-04-10 21:00:31,412 DEBUG HandlerThread:603 [system_info.py:_save_conda():207] Saving list of conda packages installed into the current environment
28
- 2024-04-10 21:00:32,286 INFO Thread-12 :603 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240410_210030-g0qpa26d/files/conda-environment.yaml
29
- 2024-04-10 21:00:46,426 ERROR HandlerThread:603 [system_info.py:_save_conda():221] Error saving conda packages: Command '['conda', 'env', 'export']' timed out after 15 seconds
30
  Traceback (most recent call last):
31
  File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/internal/system/system_info.py", line 214, in _save_conda
32
  subprocess.call(
@@ -37,101 +37,370 @@ Traceback (most recent call last):
37
  File "/opt/conda/lib/python3.10/subprocess.py", line 1951, in _wait
38
  raise TimeoutExpired(self.args, timeout)
39
  subprocess.TimeoutExpired: Command '['conda', 'env', 'export']' timed out after 15 seconds
40
- 2024-04-10 21:00:46,427 DEBUG HandlerThread:603 [system_info.py:_save_conda():222] Saving conda packages done
41
- 2024-04-10 21:00:46,427 INFO HandlerThread:603 [system_monitor.py:probe():229] Finished publishing system info
42
- 2024-04-10 21:00:46,432 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: status_report
43
- 2024-04-10 21:00:46,433 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: keepalive
44
- 2024-04-10 21:00:46,433 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: status_report
45
- 2024-04-10 21:00:46,433 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: keepalive
46
- 2024-04-10 21:00:46,433 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: status_report
47
- 2024-04-10 21:00:46,433 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: keepalive
48
- 2024-04-10 21:00:46,434 DEBUG SenderThread:603 [sender.py:send():379] send: files
49
- 2024-04-10 21:00:46,434 INFO SenderThread:603 [sender.py:_save_file():1390] saving file wandb-metadata.json with policy now
50
- 2024-04-10 21:00:46,686 INFO wandb-upload_0:603 [upload_job.py:push():131] Uploaded file /tmp/tmptu_ruxaxwandb/y8c9lp5m-wandb-metadata.json
51
- 2024-04-10 21:00:47,289 INFO Thread-12 :603 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240410_210030-g0qpa26d/files/wandb-metadata.json
52
- 2024-04-10 21:00:47,385 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: python_packages
53
- 2024-04-10 21:00:47,385 DEBUG SenderThread:603 [sender.py:send_request():406] send_request: python_packages
54
- 2024-04-10 21:00:47,389 DEBUG SenderThread:603 [sender.py:send():379] send: telemetry
55
- 2024-04-10 21:00:47,401 DEBUG SenderThread:603 [sender.py:send():379] send: config
56
- 2024-04-10 21:00:47,402 DEBUG SenderThread:603 [sender.py:send():379] send: metric
57
- 2024-04-10 21:00:47,402 DEBUG SenderThread:603 [sender.py:send():379] send: telemetry
58
- 2024-04-10 21:00:47,402 DEBUG SenderThread:603 [sender.py:send():379] send: metric
59
- 2024-04-10 21:00:47,403 WARNING SenderThread:603 [sender.py:send_metric():1341] Seen metric with glob (shouldn't happen)
60
- 2024-04-10 21:00:47,403 DEBUG SenderThread:603 [sender.py:send():379] send: telemetry
61
- 2024-04-10 21:00:47,403 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: stop_status
62
- 2024-04-10 21:00:47,404 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: internal_messages
63
- 2024-04-10 21:00:47,406 DEBUG SenderThread:603 [sender.py:send_request():406] send_request: stop_status
64
- 2024-04-10 21:00:48,290 INFO Thread-12 :603 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240410_210030-g0qpa26d/files/requirements.txt
65
- 2024-04-10 21:00:48,290 INFO Thread-12 :603 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240410_210030-g0qpa26d/files/output.log
66
- 2024-04-10 21:00:50,291 INFO Thread-12 :603 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240410_210030-g0qpa26d/files/output.log
67
- 2024-04-10 21:00:51,626 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: status_report
68
- 2024-04-10 21:00:56,627 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: status_report
69
- 2024-04-10 21:01:01,633 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: status_report
70
- 2024-04-10 21:01:02,296 INFO Thread-12 :603 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240410_210030-g0qpa26d/files/config.yaml
71
- 2024-04-10 21:01:02,387 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: stop_status
72
- 2024-04-10 21:01:02,388 DEBUG SenderThread:603 [sender.py:send_request():406] send_request: stop_status
73
- 2024-04-10 21:01:02,391 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: internal_messages
74
- 2024-04-10 21:01:07,482 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: status_report
75
- 2024-04-10 21:01:12,483 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: status_report
76
- 2024-04-10 21:01:17,386 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: stop_status
77
- 2024-04-10 21:01:17,387 DEBUG SenderThread:603 [sender.py:send_request():406] send_request: stop_status
78
- 2024-04-10 21:01:17,428 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: internal_messages
79
- 2024-04-10 21:01:17,493 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: status_report
80
- 2024-04-10 21:01:22,494 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: status_report
81
- 2024-04-10 21:01:27,495 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: status_report
82
- 2024-04-10 21:01:31,399 DEBUG SystemMonitor:603 [system_monitor.py:_start():172] Starting system metrics aggregation loop
83
- 2024-04-10 21:01:31,400 DEBUG SenderThread:603 [sender.py:send():379] send: stats
84
- 2024-04-10 21:01:32,387 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: stop_status
85
- 2024-04-10 21:01:32,387 DEBUG SenderThread:603 [sender.py:send_request():406] send_request: stop_status
86
- 2024-04-10 21:01:32,428 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: internal_messages
87
- 2024-04-10 21:01:32,732 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: status_report
88
- 2024-04-10 21:01:37,733 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: status_report
89
- 2024-04-10 21:01:42,734 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: status_report
90
- 2024-04-10 21:01:47,387 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: stop_status
91
- 2024-04-10 21:01:47,387 DEBUG SenderThread:603 [sender.py:send_request():406] send_request: stop_status
92
- 2024-04-10 21:01:47,428 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: internal_messages
93
- 2024-04-10 21:01:48,467 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: status_report
94
- 2024-04-10 21:01:53,468 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: status_report
95
- 2024-04-10 21:01:58,469 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: status_report
96
- 2024-04-10 21:02:01,401 DEBUG SenderThread:603 [sender.py:send():379] send: stats
97
- 2024-04-10 21:02:02,387 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: stop_status
98
- 2024-04-10 21:02:02,387 DEBUG SenderThread:603 [sender.py:send_request():406] send_request: stop_status
99
- 2024-04-10 21:02:02,428 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: internal_messages
100
- 2024-04-10 21:02:04,452 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: status_report
101
- 2024-04-10 21:02:09,454 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: status_report
102
- 2024-04-10 21:02:14,455 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: status_report
103
- 2024-04-10 21:02:17,387 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: stop_status
104
- 2024-04-10 21:02:17,388 DEBUG SenderThread:603 [sender.py:send_request():406] send_request: stop_status
105
- 2024-04-10 21:02:17,428 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: internal_messages
106
- 2024-04-10 21:02:19,519 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: status_report
107
- 2024-04-10 21:02:24,520 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: status_report
108
- 2024-04-10 21:02:29,521 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: status_report
109
- 2024-04-10 21:02:31,403 DEBUG SenderThread:603 [sender.py:send():379] send: stats
110
- 2024-04-10 21:02:31,907 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: partial_history
111
- 2024-04-10 21:02:31,910 DEBUG SenderThread:603 [sender.py:send():379] send: metric
112
- 2024-04-10 21:02:31,910 DEBUG SenderThread:603 [sender.py:send():379] send: metric
113
- 2024-04-10 21:02:31,910 DEBUG SenderThread:603 [sender.py:send():379] send: metric
114
- 2024-04-10 21:02:31,910 DEBUG SenderThread:603 [sender.py:send():379] send: metric
115
- 2024-04-10 21:02:31,910 DEBUG SenderThread:603 [sender.py:send():379] send: history
116
- 2024-04-10 21:02:31,910 DEBUG SenderThread:603 [sender.py:send_request():406] send_request: summary_record
117
- 2024-04-10 21:02:31,910 INFO SenderThread:603 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
118
- 2024-04-10 21:02:32,337 INFO Thread-12 :603 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240410_210030-g0qpa26d/files/wandb-summary.json
119
- 2024-04-10 21:02:32,392 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: stop_status
120
- 2024-04-10 21:02:32,392 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: internal_messages
121
- 2024-04-10 21:02:32,392 DEBUG SenderThread:603 [sender.py:send_request():406] send_request: stop_status
122
- 2024-04-10 21:02:32,482 DEBUG SenderThread:603 [sender.py:send():379] send: telemetry
123
- 2024-04-10 21:02:32,524 DEBUG HandlerThread:603 [handler.py:handle_request():146] handle_request: partial_history
124
- 2024-04-10 21:02:32,525 DEBUG SenderThread:603 [sender.py:send_request():406] send_request: summary_record
125
- 2024-04-10 21:02:32,525 INFO SenderThread:603 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
126
- 2024-04-10 21:02:32,526 DEBUG SenderThread:603 [sender.py:send_request():406] send_request: summary_record
127
- 2024-04-10 21:02:32,526 INFO SenderThread:603 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
128
- 2024-04-10 21:02:32,526 DEBUG SenderThread:603 [sender.py:send_request():406] send_request: summary_record
129
- 2024-04-10 21:02:32,527 INFO SenderThread:603 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
130
- 2024-04-10 21:02:32,527 DEBUG SenderThread:603 [sender.py:send_request():406] send_request: summary_record
131
- 2024-04-10 21:02:32,527 INFO SenderThread:603 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
132
- 2024-04-10 21:02:32,527 DEBUG SenderThread:603 [sender.py:send_request():406] send_request: summary_record
133
- 2024-04-10 21:02:32,528 INFO SenderThread:603 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
134
- 2024-04-10 21:02:32,528 DEBUG SenderThread:603 [sender.py:send():379] send: history
135
- 2024-04-10 21:02:32,528 DEBUG SenderThread:603 [sender.py:send_request():406] send_request: summary_record
136
- 2024-04-10 21:02:32,528 INFO SenderThread:603 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
137
- 2024-04-10 21:02:33,337 INFO Thread-12 :603 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240410_210030-g0qpa26d/files/wandb-summary.json
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-04-10 21:51:01,073 INFO StreamThr :274 [internal.py:wandb_internal():86] W&B internal server running at pid: 274, started at: 2024-04-10 21:51:01.072932
2
+ 2024-04-10 21:51:01,075 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status
3
+ 2024-04-10 21:51:01,447 INFO WriterThread:274 [datastore.py:open_for_write():87] open: /kaggle/working/wandb/run-20240410_215101-ta1pahei/run-ta1pahei.wandb
4
+ 2024-04-10 21:51:01,448 DEBUG SenderThread:274 [sender.py:send():379] send: header
5
+ 2024-04-10 21:51:01,459 DEBUG SenderThread:274 [sender.py:send():379] send: run
6
+ 2024-04-10 21:51:01,670 INFO SenderThread:274 [dir_watcher.py:__init__():211] watching files in: /kaggle/working/wandb/run-20240410_215101-ta1pahei/files
7
+ 2024-04-10 21:51:01,670 INFO SenderThread:274 [sender.py:_start_run_threads():1124] run started: ta1pahei with start time 1712785861.074028
8
+ 2024-04-10 21:51:01,678 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: check_version
9
+ 2024-04-10 21:51:01,678 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: check_version
10
+ 2024-04-10 21:51:01,777 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: run_start
11
+ 2024-04-10 21:51:01,788 DEBUG HandlerThread:274 [system_info.py:__init__():26] System info init
12
+ 2024-04-10 21:51:01,789 DEBUG HandlerThread:274 [system_info.py:__init__():41] System info init done
13
+ 2024-04-10 21:51:01,789 INFO HandlerThread:274 [system_monitor.py:start():194] Starting system monitor
14
+ 2024-04-10 21:51:01,789 INFO SystemMonitor:274 [system_monitor.py:_start():158] Starting system asset monitoring threads
15
+ 2024-04-10 21:51:01,789 INFO HandlerThread:274 [system_monitor.py:probe():214] Collecting system info
16
+ 2024-04-10 21:51:01,789 INFO SystemMonitor:274 [interfaces.py:start():190] Started cpu monitoring
17
+ 2024-04-10 21:51:01,790 INFO SystemMonitor:274 [interfaces.py:start():190] Started disk monitoring
18
+ 2024-04-10 21:51:01,791 INFO SystemMonitor:274 [interfaces.py:start():190] Started gpu monitoring
19
+ 2024-04-10 21:51:01,792 INFO SystemMonitor:274 [interfaces.py:start():190] Started memory monitoring
20
+ 2024-04-10 21:51:01,792 INFO SystemMonitor:274 [interfaces.py:start():190] Started network monitoring
21
+ 2024-04-10 21:51:01,802 DEBUG HandlerThread:274 [system_info.py:probe():150] Probing system
22
+ 2024-04-10 21:51:01,805 DEBUG HandlerThread:274 [gitlib.py:_init_repo():56] git repository is invalid
23
+ 2024-04-10 21:51:01,805 DEBUG HandlerThread:274 [system_info.py:probe():198] Probing system done
24
+ 2024-04-10 21:51:01,805 DEBUG HandlerThread:274 [system_monitor.py:probe():223] {'os': 'Linux-5.15.133+-x86_64-with-glibc2.31', 'python': '3.10.13', 'heartbeatAt': '2024-04-10T21:51:01.802955', 'startedAt': '2024-04-10T21:51:01.066772', 'docker': None, 'cuda': None, 'args': (), 'state': 'running', 'program': 'kaggle.ipynb', 'codePathLocal': None, 'root': '/kaggle/working', 'host': '4bf7c58fba03', 'username': 'root', 'executable': '/opt/conda/bin/python3.10', 'cpu_count': 2, 'cpu_count_logical': 4, 'cpu_freq': {'current': 2000.192, 'min': 0.0, 'max': 0.0}, 'cpu_freq_per_core': [{'current': 2000.192, 'min': 0.0, 'max': 0.0}, {'current': 2000.192, 'min': 0.0, 'max': 0.0}, {'current': 2000.192, 'min': 0.0, 'max': 0.0}, {'current': 2000.192, 'min': 0.0, 'max': 0.0}], 'disk': {'/': {'total': 8062.387607574463, 'used': 5578.723789215088}}, 'gpu': 'Tesla T4', 'gpu_count': 2, 'gpu_devices': [{'name': 'Tesla T4', 'memory_total': 16106127360}, {'name': 'Tesla T4', 'memory_total': 16106127360}], 'memory': {'total': 31.357559204101562}}
25
+ 2024-04-10 21:51:01,805 INFO HandlerThread:274 [system_monitor.py:probe():224] Finished collecting system info
26
+ 2024-04-10 21:51:01,805 INFO HandlerThread:274 [system_monitor.py:probe():227] Publishing system info
27
+ 2024-04-10 21:51:01,805 DEBUG HandlerThread:274 [system_info.py:_save_conda():207] Saving list of conda packages installed into the current environment
28
+ 2024-04-10 21:51:02,672 INFO Thread-12 :274 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240410_215101-ta1pahei/files/conda-environment.yaml
29
+ 2024-04-10 21:51:16,821 ERROR HandlerThread:274 [system_info.py:_save_conda():221] Error saving conda packages: Command '['conda', 'env', 'export']' timed out after 15 seconds
30
  Traceback (most recent call last):
31
  File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/internal/system/system_info.py", line 214, in _save_conda
32
  subprocess.call(
 
37
  File "/opt/conda/lib/python3.10/subprocess.py", line 1951, in _wait
38
  raise TimeoutExpired(self.args, timeout)
39
  subprocess.TimeoutExpired: Command '['conda', 'env', 'export']' timed out after 15 seconds
40
+ 2024-04-10 21:51:16,824 DEBUG HandlerThread:274 [system_info.py:_save_conda():222] Saving conda packages done
41
+ 2024-04-10 21:51:16,825 INFO HandlerThread:274 [system_monitor.py:probe():229] Finished publishing system info
42
+ 2024-04-10 21:51:16,831 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
43
+ 2024-04-10 21:51:16,832 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: keepalive
44
+ 2024-04-10 21:51:16,832 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
45
+ 2024-04-10 21:51:16,832 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: keepalive
46
+ 2024-04-10 21:51:16,832 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
47
+ 2024-04-10 21:51:16,832 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: keepalive
48
+ 2024-04-10 21:51:16,833 DEBUG SenderThread:274 [sender.py:send():379] send: files
49
+ 2024-04-10 21:51:16,833 INFO SenderThread:274 [sender.py:_save_file():1390] saving file wandb-metadata.json with policy now
50
+ 2024-04-10 21:51:17,270 INFO wandb-upload_0:274 [upload_job.py:push():131] Uploaded file /tmp/tmphet9wukmwandb/tfhoobcd-wandb-metadata.json
51
+ 2024-04-10 21:51:17,675 INFO Thread-12 :274 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240410_215101-ta1pahei/files/wandb-metadata.json
52
+ 2024-04-10 21:51:17,776 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: python_packages
53
+ 2024-04-10 21:51:17,776 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: python_packages
54
+ 2024-04-10 21:51:17,779 DEBUG SenderThread:274 [sender.py:send():379] send: telemetry
55
+ 2024-04-10 21:51:17,789 DEBUG SenderThread:274 [sender.py:send():379] send: config
56
+ 2024-04-10 21:51:17,791 DEBUG SenderThread:274 [sender.py:send():379] send: metric
57
+ 2024-04-10 21:51:17,792 DEBUG SenderThread:274 [sender.py:send():379] send: telemetry
58
+ 2024-04-10 21:51:17,792 DEBUG SenderThread:274 [sender.py:send():379] send: metric
59
+ 2024-04-10 21:51:17,792 WARNING SenderThread:274 [sender.py:send_metric():1341] Seen metric with glob (shouldn't happen)
60
+ 2024-04-10 21:51:17,792 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
61
+ 2024-04-10 21:51:17,793 DEBUG SenderThread:274 [sender.py:send():379] send: telemetry
62
+ 2024-04-10 21:51:17,793 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
63
+ 2024-04-10 21:51:17,793 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
64
+ 2024-04-10 21:51:18,676 INFO Thread-12 :274 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240410_215101-ta1pahei/files/requirements.txt
65
+ 2024-04-10 21:51:18,676 INFO Thread-12 :274 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240410_215101-ta1pahei/files/output.log
66
+ 2024-04-10 21:51:20,677 INFO Thread-12 :274 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240410_215101-ta1pahei/files/output.log
67
+ 2024-04-10 21:51:22,163 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
68
+ 2024-04-10 21:51:22,849 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: pause
69
+ 2024-04-10 21:51:22,849 INFO HandlerThread:274 [handler.py:handle_request_pause():708] stopping system metrics thread
70
+ 2024-04-10 21:51:22,849 INFO HandlerThread:274 [system_monitor.py:finish():203] Stopping system monitor
71
+ 2024-04-10 21:51:22,849 DEBUG SystemMonitor:274 [system_monitor.py:_start():172] Starting system metrics aggregation loop
72
+ 2024-04-10 21:51:22,849 DEBUG SystemMonitor:274 [system_monitor.py:_start():179] Finished system metrics aggregation loop
73
+ 2024-04-10 21:51:22,849 DEBUG SystemMonitor:274 [system_monitor.py:_start():183] Publishing last batch of metrics
74
+ 2024-04-10 21:51:22,850 INFO HandlerThread:274 [interfaces.py:finish():202] Joined cpu monitor
75
+ 2024-04-10 21:51:22,850 INFO HandlerThread:274 [interfaces.py:finish():202] Joined disk monitor
76
+ 2024-04-10 21:51:22,861 INFO HandlerThread:274 [interfaces.py:finish():202] Joined gpu monitor
77
+ 2024-04-10 21:51:22,861 INFO HandlerThread:274 [interfaces.py:finish():202] Joined memory monitor
78
+ 2024-04-10 21:51:22,861 INFO HandlerThread:274 [interfaces.py:finish():202] Joined network monitor
79
+ 2024-04-10 21:51:22,861 DEBUG SenderThread:274 [sender.py:send():379] send: stats
80
+ 2024-04-10 21:51:27,862 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
81
+ 2024-04-10 21:51:32,778 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
82
+ 2024-04-10 21:51:32,779 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
83
+ 2024-04-10 21:51:32,779 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
84
+ 2024-04-10 21:51:33,855 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
85
+ 2024-04-10 21:51:34,682 INFO Thread-12 :274 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240410_215101-ta1pahei/files/config.yaml
86
+ 2024-04-10 21:51:38,978 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
87
+ 2024-04-10 21:51:43,979 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
88
+ 2024-04-10 21:51:47,777 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
89
+ 2024-04-10 21:51:47,778 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
90
+ 2024-04-10 21:51:47,817 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
91
+ 2024-04-10 21:51:49,947 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
92
+ 2024-04-10 21:51:54,947 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
93
+ 2024-04-10 21:51:59,948 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
94
+ 2024-04-10 21:52:02,777 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
95
+ 2024-04-10 21:52:02,778 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
96
+ 2024-04-10 21:52:02,818 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
97
+ 2024-04-10 21:52:05,863 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
98
+ 2024-04-10 21:52:10,864 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
99
+ 2024-04-10 21:52:15,865 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
100
+ 2024-04-10 21:52:17,777 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
101
+ 2024-04-10 21:52:17,778 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
102
+ 2024-04-10 21:52:17,818 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
103
+ 2024-04-10 21:52:20,900 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
104
+ 2024-04-10 21:52:25,901 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
105
+ 2024-04-10 21:52:30,902 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
106
+ 2024-04-10 21:52:32,777 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
107
+ 2024-04-10 21:52:32,778 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
108
+ 2024-04-10 21:52:32,818 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
109
+ 2024-04-10 21:52:36,843 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
110
+ 2024-04-10 21:52:40,759 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: resume
111
+ 2024-04-10 21:52:40,759 INFO HandlerThread:274 [handler.py:handle_request_resume():699] starting system metrics thread
112
+ 2024-04-10 21:52:40,759 INFO HandlerThread:274 [system_monitor.py:start():194] Starting system monitor
113
+ 2024-04-10 21:52:40,759 INFO SystemMonitor:274 [system_monitor.py:_start():158] Starting system asset monitoring threads
114
+ 2024-04-10 21:52:40,760 INFO SystemMonitor:274 [interfaces.py:start():190] Started cpu monitoring
115
+ 2024-04-10 21:52:40,761 INFO SystemMonitor:274 [interfaces.py:start():190] Started disk monitoring
116
+ 2024-04-10 21:52:40,762 INFO SystemMonitor:274 [interfaces.py:start():190] Started gpu monitoring
117
+ 2024-04-10 21:52:40,763 INFO SystemMonitor:274 [interfaces.py:start():190] Started memory monitoring
118
+ 2024-04-10 21:52:40,764 INFO SystemMonitor:274 [interfaces.py:start():190] Started network monitoring
119
+ 2024-04-10 21:52:41,844 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
120
+ 2024-04-10 21:52:46,708 INFO Thread-12 :274 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240410_215101-ta1pahei/files/output.log
121
+ 2024-04-10 21:52:47,665 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
122
+ 2024-04-10 21:52:47,789 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
123
+ 2024-04-10 21:52:47,789 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
124
+ 2024-04-10 21:52:47,829 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
125
+ 2024-04-10 21:52:48,709 INFO Thread-12 :274 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240410_215101-ta1pahei/files/output.log
126
+ 2024-04-10 21:52:52,915 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
127
+ 2024-04-10 21:52:57,916 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
128
+ 2024-04-10 21:53:02,783 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
129
+ 2024-04-10 21:53:02,784 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
130
+ 2024-04-10 21:53:02,784 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
131
+ 2024-04-10 21:53:04,789 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
132
+ 2024-04-10 21:53:09,789 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
133
+ 2024-04-10 21:53:14,790 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
134
+ 2024-04-10 21:53:17,800 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
135
+ 2024-04-10 21:53:17,802 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
136
+ 2024-04-10 21:53:17,802 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
137
+ 2024-04-10 21:53:19,871 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
138
+ 2024-04-10 21:53:24,874 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
139
+ 2024-04-10 21:53:29,875 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
140
+ 2024-04-10 21:53:32,801 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
141
+ 2024-04-10 21:53:32,801 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
142
+ 2024-04-10 21:53:32,802 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
143
+ 2024-04-10 21:53:34,952 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
144
+ 2024-04-10 21:53:39,953 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
145
+ 2024-04-10 21:53:40,767 DEBUG SystemMonitor:274 [system_monitor.py:_start():172] Starting system metrics aggregation loop
146
+ 2024-04-10 21:53:40,768 DEBUG SenderThread:274 [sender.py:send():379] send: stats
147
+ 2024-04-10 21:53:45,770 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
148
+ 2024-04-10 21:53:48,058 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
149
+ 2024-04-10 21:53:48,058 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
150
+ 2024-04-10 21:53:48,106 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
151
+ 2024-04-10 21:53:51,112 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
152
+ 2024-04-10 21:53:56,113 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
153
+ 2024-04-10 21:54:01,113 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
154
+ 2024-04-10 21:54:03,034 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
155
+ 2024-04-10 21:54:03,035 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
156
+ 2024-04-10 21:54:03,075 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
157
+ 2024-04-10 21:54:06,135 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
158
+ 2024-04-10 21:54:10,769 DEBUG SenderThread:274 [sender.py:send():379] send: stats
159
+ 2024-04-10 21:54:11,770 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
160
+ 2024-04-10 21:54:16,771 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
161
+ 2024-04-10 21:54:18,045 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
162
+ 2024-04-10 21:54:18,045 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
163
+ 2024-04-10 21:54:18,087 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
164
+ 2024-04-10 21:54:22,200 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
165
+ 2024-04-10 21:54:27,201 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
166
+ 2024-04-10 21:54:32,202 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
167
+ 2024-04-10 21:54:33,045 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
168
+ 2024-04-10 21:54:33,045 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
169
+ 2024-04-10 21:54:33,085 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
170
+ 2024-04-10 21:54:37,214 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
171
+ 2024-04-10 21:54:40,770 DEBUG SenderThread:274 [sender.py:send():379] send: stats
172
+ 2024-04-10 21:54:42,771 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
173
+ 2024-04-10 21:54:47,772 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
174
+ 2024-04-10 21:54:48,045 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
175
+ 2024-04-10 21:54:48,046 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
176
+ 2024-04-10 21:54:48,087 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
177
+ 2024-04-10 21:54:53,167 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
178
+ 2024-04-10 21:54:58,167 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
179
+ 2024-04-10 21:55:03,168 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
180
+ 2024-04-10 21:55:04,522 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
181
+ 2024-04-10 21:55:04,523 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
182
+ 2024-04-10 21:55:04,524 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
183
+ 2024-04-10 21:55:08,693 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
184
+ 2024-04-10 21:55:10,771 DEBUG SenderThread:274 [sender.py:send():379] send: stats
185
+ 2024-04-10 21:55:13,772 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
186
+ 2024-04-10 21:55:18,288 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
187
+ 2024-04-10 21:55:18,288 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
188
+ 2024-04-10 21:55:18,399 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
189
+ 2024-04-10 21:55:19,387 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
190
+ 2024-04-10 21:55:24,388 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
191
+ 2024-04-10 21:55:29,389 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
192
+ 2024-04-10 21:55:33,895 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
193
+ 2024-04-10 21:55:33,895 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
194
+ 2024-04-10 21:55:33,896 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
195
+ 2024-04-10 21:55:34,984 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
196
+ 2024-04-10 21:55:39,985 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
197
+ 2024-04-10 21:55:40,772 DEBUG SenderThread:274 [sender.py:send():379] send: stats
198
+ 2024-04-10 21:55:45,774 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
199
+ 2024-04-10 21:55:48,531 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
200
+ 2024-04-10 21:55:48,581 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
201
+ 2024-04-10 21:55:48,582 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
202
+ 2024-04-10 21:55:50,895 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
203
+ 2024-04-10 21:55:55,897 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
204
+ 2024-04-10 21:56:00,897 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
205
+ 2024-04-10 21:56:03,531 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
206
+ 2024-04-10 21:56:03,581 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
207
+ 2024-04-10 21:56:03,582 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
208
+ 2024-04-10 21:56:06,895 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
209
+ 2024-04-10 21:56:10,774 DEBUG SenderThread:274 [sender.py:send():379] send: stats
210
+ 2024-04-10 21:56:12,775 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
211
+ 2024-04-10 21:56:17,775 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
212
+ 2024-04-10 21:56:18,531 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
213
+ 2024-04-10 21:56:18,581 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
214
+ 2024-04-10 21:56:18,582 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
215
+ 2024-04-10 21:56:22,895 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
216
+ 2024-04-10 21:56:27,896 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
217
+ 2024-04-10 21:56:32,898 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
218
+ 2024-04-10 21:56:33,722 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
219
+ 2024-04-10 21:56:33,729 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
220
+ 2024-04-10 21:56:33,729 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
221
+ 2024-04-10 21:56:38,159 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
222
+ 2024-04-10 21:56:40,775 DEBUG SenderThread:274 [sender.py:send():379] send: stats
223
+ 2024-04-10 21:56:43,776 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
224
+ 2024-04-10 21:56:48,546 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
225
+ 2024-04-10 21:56:48,585 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
226
+ 2024-04-10 21:56:48,586 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
227
+ 2024-04-10 21:56:49,717 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
228
+ 2024-04-10 21:56:55,577 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
229
+ 2024-04-10 21:57:00,578 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
230
+ 2024-04-10 21:57:03,546 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
231
+ 2024-04-10 21:57:03,586 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
232
+ 2024-04-10 21:57:03,586 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
233
+ 2024-04-10 21:57:05,636 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
234
+ 2024-04-10 21:57:10,637 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
235
+ 2024-04-10 21:57:10,776 DEBUG SenderThread:274 [sender.py:send():379] send: stats
236
+ 2024-04-10 21:57:15,777 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
237
+ 2024-04-10 21:57:18,547 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
238
+ 2024-04-10 21:57:18,586 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
239
+ 2024-04-10 21:57:18,586 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
240
+ 2024-04-10 21:57:21,585 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
241
+ 2024-04-10 21:57:26,586 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
242
+ 2024-04-10 21:57:31,587 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
243
+ 2024-04-10 21:57:33,567 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
244
+ 2024-04-10 21:57:33,588 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
245
+ 2024-04-10 21:57:33,589 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
246
+ 2024-04-10 21:57:36,605 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
247
+ 2024-04-10 21:57:40,777 DEBUG SenderThread:274 [sender.py:send():379] send: stats
248
+ 2024-04-10 21:57:41,778 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
249
+ 2024-04-10 21:57:46,779 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
250
+ 2024-04-10 21:57:48,567 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
251
+ 2024-04-10 21:57:48,588 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
252
+ 2024-04-10 21:57:48,589 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
253
+ 2024-04-10 21:57:52,586 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
254
+ 2024-04-10 21:57:57,587 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
255
+ 2024-04-10 21:58:02,588 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
256
+ 2024-04-10 21:58:03,567 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
257
+ 2024-04-10 21:58:03,588 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
258
+ 2024-04-10 21:58:03,589 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
259
+ 2024-04-10 21:58:08,587 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
260
+ 2024-04-10 21:58:10,778 DEBUG SenderThread:274 [sender.py:send():379] send: stats
261
+ 2024-04-10 21:58:13,779 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
262
+ 2024-04-10 21:58:18,567 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
263
+ 2024-04-10 21:58:18,588 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
264
+ 2024-04-10 21:58:18,589 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
265
+ 2024-04-10 21:58:19,703 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
266
+ 2024-04-10 21:58:25,586 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
267
+ 2024-04-10 21:58:30,587 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
268
+ 2024-04-10 21:58:33,567 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
269
+ 2024-04-10 21:58:33,589 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
270
+ 2024-04-10 21:58:33,589 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
271
+ 2024-04-10 21:58:36,586 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
272
+ 2024-04-10 21:58:40,779 DEBUG SenderThread:274 [sender.py:send():379] send: stats
273
+ 2024-04-10 21:58:41,780 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
274
+ 2024-04-10 21:58:46,780 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
275
+ 2024-04-10 21:58:48,567 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
276
+ 2024-04-10 21:58:48,602 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
277
+ 2024-04-10 21:58:48,602 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
278
+ 2024-04-10 21:58:52,596 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
279
+ 2024-04-10 21:58:57,597 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
280
+ 2024-04-10 21:59:02,598 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
281
+ 2024-04-10 21:59:03,568 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
282
+ 2024-04-10 21:59:03,589 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
283
+ 2024-04-10 21:59:03,589 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
284
+ 2024-04-10 21:59:08,596 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
285
+ 2024-04-10 21:59:10,780 DEBUG SenderThread:274 [sender.py:send():379] send: stats
286
+ 2024-04-10 21:59:13,781 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
287
+ 2024-04-10 21:59:18,568 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
288
+ 2024-04-10 21:59:18,589 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
289
+ 2024-04-10 21:59:18,590 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
290
+ 2024-04-10 21:59:19,663 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
291
+ 2024-04-10 21:59:25,597 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
292
+ 2024-04-10 21:59:30,598 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
293
+ 2024-04-10 21:59:33,568 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
294
+ 2024-04-10 21:59:33,589 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
295
+ 2024-04-10 21:59:33,590 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
296
+ 2024-04-10 21:59:35,711 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
297
+ 2024-04-10 21:59:40,781 DEBUG SenderThread:274 [sender.py:send():379] send: stats
298
+ 2024-04-10 21:59:40,783 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
299
+ 2024-04-10 21:59:45,783 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
300
+ 2024-04-10 21:59:48,568 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
301
+ 2024-04-10 21:59:48,589 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
302
+ 2024-04-10 21:59:48,590 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
303
+ 2024-04-10 21:59:51,602 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
304
+ 2024-04-10 21:59:56,604 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
305
+ 2024-04-10 22:00:01,605 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
306
+ 2024-04-10 22:00:03,568 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
307
+ 2024-04-10 22:00:03,589 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
308
+ 2024-04-10 22:00:03,590 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
309
+ 2024-04-10 22:00:06,608 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
310
+ 2024-04-10 22:00:10,782 DEBUG SenderThread:274 [sender.py:send():379] send: stats
311
+ 2024-04-10 22:00:11,783 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
312
+ 2024-04-10 22:00:16,784 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
313
+ 2024-04-10 22:00:18,568 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
314
+ 2024-04-10 22:00:18,590 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
315
+ 2024-04-10 22:00:18,590 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
316
+ 2024-04-10 22:00:22,603 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
317
+ 2024-04-10 22:00:27,604 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
318
+ 2024-04-10 22:00:32,605 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
319
+ 2024-04-10 22:00:33,584 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
320
+ 2024-04-10 22:00:33,614 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
321
+ 2024-04-10 22:00:33,615 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
322
+ 2024-04-10 22:00:38,603 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
323
+ 2024-04-10 22:00:40,783 DEBUG SenderThread:274 [sender.py:send():379] send: stats
324
+ 2024-04-10 22:00:44,504 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
325
+ 2024-04-10 22:00:44,558 DEBUG SenderThread:274 [sender.py:send():379] send: config
326
+ 2024-04-10 22:00:44,560 DEBUG SenderThread:274 [sender.py:send():379] send: metric
327
+ 2024-04-10 22:00:44,560 DEBUG SenderThread:274 [sender.py:send():379] send: metric
328
+ 2024-04-10 22:00:44,560 WARNING SenderThread:274 [sender.py:send_metric():1341] Seen metric with glob (shouldn't happen)
329
+ 2024-04-10 22:00:44,919 INFO Thread-12 :274 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240410_215101-ta1pahei/files/output.log
330
+ 2024-04-10 22:00:46,920 INFO Thread-12 :274 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240410_215101-ta1pahei/files/output.log
331
+ 2024-04-10 22:00:48,585 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
332
+ 2024-04-10 22:00:48,615 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
333
+ 2024-04-10 22:00:48,615 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
334
+ 2024-04-10 22:00:48,921 INFO Thread-12 :274 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240410_215101-ta1pahei/files/output.log
335
+ 2024-04-10 22:00:49,760 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
336
+ 2024-04-10 22:00:55,604 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
337
+ 2024-04-10 22:00:56,923 INFO Thread-12 :274 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240410_215101-ta1pahei/files/output.log
338
+ 2024-04-10 22:00:58,924 INFO Thread-12 :274 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240410_215101-ta1pahei/files/output.log
339
+ 2024-04-10 22:01:00,664 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
340
+ 2024-04-10 22:01:03,585 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
341
+ 2024-04-10 22:01:03,615 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
342
+ 2024-04-10 22:01:03,615 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
343
+ 2024-04-10 22:01:05,691 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
344
+ 2024-04-10 22:01:05,927 INFO Thread-12 :274 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240410_215101-ta1pahei/files/config.yaml
345
+ 2024-04-10 22:01:08,928 INFO Thread-12 :274 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240410_215101-ta1pahei/files/output.log
346
+ 2024-04-10 22:01:10,784 DEBUG SenderThread:274 [sender.py:send():379] send: stats
347
+ 2024-04-10 22:01:11,785 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
348
+ 2024-04-10 22:01:16,786 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
349
+ 2024-04-10 22:01:18,585 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
350
+ 2024-04-10 22:01:18,615 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
351
+ 2024-04-10 22:01:18,616 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
352
+ 2024-04-10 22:01:22,656 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
353
+ 2024-04-10 22:01:27,657 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
354
+ 2024-04-10 22:01:32,658 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
355
+ 2024-04-10 22:01:33,585 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
356
+ 2024-04-10 22:01:33,615 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
357
+ 2024-04-10 22:01:33,615 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
358
+ 2024-04-10 22:01:38,656 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
359
+ 2024-04-10 22:01:40,785 DEBUG SenderThread:274 [sender.py:send():379] send: stats
360
+ 2024-04-10 22:01:43,786 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
361
+ 2024-04-10 22:01:48,588 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
362
+ 2024-04-10 22:01:48,615 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
363
+ 2024-04-10 22:01:48,615 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
364
+ 2024-04-10 22:01:49,779 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
365
+ 2024-04-10 22:01:54,904 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: partial_history
366
+ 2024-04-10 22:01:54,907 DEBUG SenderThread:274 [sender.py:send():379] send: metric
367
+ 2024-04-10 22:01:54,907 DEBUG SenderThread:274 [sender.py:send():379] send: metric
368
+ 2024-04-10 22:01:54,907 DEBUG SenderThread:274 [sender.py:send():379] send: metric
369
+ 2024-04-10 22:01:54,907 DEBUG SenderThread:274 [sender.py:send():379] send: metric
370
+ 2024-04-10 22:01:54,907 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
371
+ 2024-04-10 22:01:54,907 DEBUG SenderThread:274 [sender.py:send():379] send: history
372
+ 2024-04-10 22:01:54,908 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: summary_record
373
+ 2024-04-10 22:01:54,908 INFO SenderThread:274 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
374
+ 2024-04-10 22:01:54,945 INFO Thread-12 :274 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240410_215101-ta1pahei/files/wandb-summary.json
375
+ 2024-04-10 22:01:56,946 INFO Thread-12 :274 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240410_215101-ta1pahei/files/output.log
376
+ 2024-04-10 22:01:57,338 DEBUG SenderThread:274 [sender.py:send():379] send: telemetry
377
+ 2024-04-10 22:01:57,338 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: summary_record
378
+ 2024-04-10 22:01:57,340 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: partial_history
379
+ 2024-04-10 22:01:57,341 INFO SenderThread:274 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
380
+ 2024-04-10 22:01:57,342 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: summary_record
381
+ 2024-04-10 22:01:57,342 INFO SenderThread:274 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
382
+ 2024-04-10 22:01:57,342 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: summary_record
383
+ 2024-04-10 22:01:57,343 INFO SenderThread:274 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
384
+ 2024-04-10 22:01:57,343 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: summary_record
385
+ 2024-04-10 22:01:57,343 INFO SenderThread:274 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
386
+ 2024-04-10 22:01:57,344 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: summary_record
387
+ 2024-04-10 22:01:57,344 INFO SenderThread:274 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
388
+ 2024-04-10 22:01:57,344 DEBUG SenderThread:274 [sender.py:send():379] send: history
389
+ 2024-04-10 22:01:57,345 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: summary_record
390
+ 2024-04-10 22:01:57,345 INFO SenderThread:274 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
391
+ 2024-04-10 22:01:57,946 INFO Thread-12 :274 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240410_215101-ta1pahei/files/wandb-summary.json
392
+ 2024-04-10 22:01:58,946 INFO Thread-12 :274 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240410_215101-ta1pahei/files/output.log
393
+ 2024-04-10 22:02:00,489 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
394
+ 2024-04-10 22:02:03,585 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
395
+ 2024-04-10 22:02:03,615 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
396
+ 2024-04-10 22:02:03,616 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
397
+ 2024-04-10 22:02:05,656 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
398
+ 2024-04-10 22:02:10,662 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
399
+ 2024-04-10 22:02:10,786 DEBUG SenderThread:274 [sender.py:send():379] send: stats
400
+ 2024-04-10 22:02:10,950 INFO Thread-12 :274 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240410_215101-ta1pahei/files/config.yaml
401
+ 2024-04-10 22:02:15,787 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
402
+ 2024-04-10 22:02:18,586 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
403
+ 2024-04-10 22:02:18,615 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
404
+ 2024-04-10 22:02:18,616 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
405
+ 2024-04-10 22:02:21,656 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
406
+ 2024-04-10 22:02:26,657 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
wandb/debug.log CHANGED
@@ -1,31 +1,36 @@
1
- 2024-04-10 21:00:30,776 INFO MainThread:556 [wandb_setup.py:_flush():76] Current SDK version is 0.16.5
2
- 2024-04-10 21:00:30,776 INFO MainThread:556 [wandb_setup.py:_flush():76] Configure stats pid to 556
3
- 2024-04-10 21:00:30,776 INFO MainThread:556 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
4
- 2024-04-10 21:00:30,776 INFO MainThread:556 [wandb_setup.py:_flush():76] Loading settings from /kaggle/working/wandb/settings
5
- 2024-04-10 21:00:30,776 INFO MainThread:556 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
6
- 2024-04-10 21:00:30,777 INFO MainThread:556 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
7
- 2024-04-10 21:00:30,777 INFO MainThread:556 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program': '<python with no main file>'}
8
- 2024-04-10 21:00:30,777 INFO MainThread:556 [wandb_setup.py:_flush():76] Applying login settings: {}
9
- 2024-04-10 21:00:30,777 INFO MainThread:556 [wandb_init.py:_log_setup():527] Logging user logs to /kaggle/working/wandb/run-20240410_210030-g0qpa26d/logs/debug.log
10
- 2024-04-10 21:00:30,777 INFO MainThread:556 [wandb_init.py:_log_setup():528] Logging internal logs to /kaggle/working/wandb/run-20240410_210030-g0qpa26d/logs/debug-internal.log
11
- 2024-04-10 21:00:30,777 INFO MainThread:556 [wandb_init.py:_jupyter_setup():473] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x7ce03cb61630>
12
- 2024-04-10 21:00:30,777 INFO MainThread:556 [wandb_init.py:init():567] calling init triggers
13
- 2024-04-10 21:00:30,777 INFO MainThread:556 [wandb_init.py:init():574] wandb.init called with sweep_config: {}
 
14
  config: {}
15
- 2024-04-10 21:00:30,777 INFO MainThread:556 [wandb_init.py:init():617] starting backend
16
- 2024-04-10 21:00:30,777 INFO MainThread:556 [wandb_init.py:init():621] setting up manager
17
- 2024-04-10 21:00:30,779 INFO MainThread:556 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
18
- 2024-04-10 21:00:30,780 INFO MainThread:556 [wandb_init.py:init():629] backend started and connected
19
- 2024-04-10 21:00:30,794 INFO MainThread:556 [wandb_run.py:_label_probe_notebook():1299] probe notebook
20
- 2024-04-10 21:00:31,093 INFO MainThread:556 [wandb_init.py:init():721] updated telemetry
21
- 2024-04-10 21:00:31,096 INFO MainThread:556 [wandb_init.py:init():754] communicating run to backend with 90.0 second timeout
22
- 2024-04-10 21:00:31,291 INFO MainThread:556 [wandb_run.py:_on_init():2344] communicating current version
23
- 2024-04-10 21:00:31,376 INFO MainThread:556 [wandb_run.py:_on_init():2353] got version response upgrade_message: "wandb version 0.16.6 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
24
 
25
- 2024-04-10 21:00:31,376 INFO MainThread:556 [wandb_init.py:init():805] starting run threads in backend
26
- 2024-04-10 21:00:47,386 INFO MainThread:556 [wandb_run.py:_console_start():2323] atexit reg
27
- 2024-04-10 21:00:47,386 INFO MainThread:556 [wandb_run.py:_redirect():2178] redirect: wrap_raw
28
- 2024-04-10 21:00:47,388 INFO MainThread:556 [wandb_run.py:_redirect():2243] Wrapping output streams.
29
- 2024-04-10 21:00:47,388 INFO MainThread:556 [wandb_run.py:_redirect():2268] Redirects installed.
30
- 2024-04-10 21:00:47,389 INFO MainThread:556 [wandb_init.py:init():848] run started, returning control to user process
31
- 2024-04-10 21:00:47,395 INFO MainThread:556 [wandb_run.py:_config_callback():1347] config_cb None None {'vocab_size': 32000, 'max_position_embeddings': 2048, 'hidden_size': 2048, 'intermediate_size': 5632, 'num_hidden_layers': 22, 'num_attention_heads': 32, 'num_key_value_heads': 4, 'hidden_act': 'silu', 'initializer_range': 0.02, 'rms_norm_eps': 1e-05, 'pretraining_tp': 1, 'use_cache': False, 'rope_theta': 10000.0, 'rope_scaling': None, 'attention_bias': False, 'attention_dropout': 0.0, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'float32', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['LlamaForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': None, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T', 'transformers_version': '4.39.3', 'model_type': 'llama', 'quantization_config': {'quant_method': 'QuantizationMethod.BITS_AND_BYTES', '_load_in_8bit': False, '_load_in_4bit': True, 'llm_int8_threshold': 6.0, 'llm_int8_skip_modules': None, 'llm_int8_enable_fp32_cpu_offload': False, 'llm_int8_has_fp16_weight': False, 'bnb_4bit_quant_type': 'nf4', 'bnb_4bit_use_double_quant': False, 'bnb_4bit_compute_dtype': 'float16', 'bnb_4bit_quant_storage': 'uint8', 'load_in_4bit': True, 'load_in_8bit': False}, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': False, 'do_predict': False, 'evaluation_strategy': 'no', 'prediction_loss_only': False, 'per_device_train_batch_size': 8, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 4, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 3e-05, 'weight_decay': 0.0001, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 0.01, 'num_train_epochs': 5, 'max_steps': 10, 'lr_scheduler_type': 'cosine', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.05, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Apr10_20-58-11_57bca537cf5f', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 10, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 10, 'save_total_limit': 1, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 8, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'paged_adamw_8bit', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': True, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': True, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None}
 
 
 
 
 
1
+ 2024-04-10 21:51:01,068 INFO MainThread:227 [wandb_setup.py:_flush():76] Current SDK version is 0.16.5
2
+ 2024-04-10 21:51:01,068 INFO MainThread:227 [wandb_setup.py:_flush():76] Configure stats pid to 227
3
+ 2024-04-10 21:51:01,068 INFO MainThread:227 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
4
+ 2024-04-10 21:51:01,068 INFO MainThread:227 [wandb_setup.py:_flush():76] Loading settings from /kaggle/working/wandb/settings
5
+ 2024-04-10 21:51:01,068 INFO MainThread:227 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
6
+ 2024-04-10 21:51:01,068 INFO MainThread:227 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
7
+ 2024-04-10 21:51:01,068 INFO MainThread:227 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program': '<python with no main file>'}
8
+ 2024-04-10 21:51:01,068 INFO MainThread:227 [wandb_setup.py:_flush():76] Applying login settings: {}
9
+ 2024-04-10 21:51:01,068 INFO MainThread:227 [wandb_setup.py:_flush():76] Applying login settings: {'api_key': '***REDACTED***'}
10
+ 2024-04-10 21:51:01,068 INFO MainThread:227 [wandb_init.py:_log_setup():527] Logging user logs to /kaggle/working/wandb/run-20240410_215101-ta1pahei/logs/debug.log
11
+ 2024-04-10 21:51:01,069 INFO MainThread:227 [wandb_init.py:_log_setup():528] Logging internal logs to /kaggle/working/wandb/run-20240410_215101-ta1pahei/logs/debug-internal.log
12
+ 2024-04-10 21:51:01,069 INFO MainThread:227 [wandb_init.py:_jupyter_setup():473] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x7b92ac4655a0>
13
+ 2024-04-10 21:51:01,069 INFO MainThread:227 [wandb_init.py:init():567] calling init triggers
14
+ 2024-04-10 21:51:01,069 INFO MainThread:227 [wandb_init.py:init():574] wandb.init called with sweep_config: {}
15
  config: {}
16
+ 2024-04-10 21:51:01,069 INFO MainThread:227 [wandb_init.py:init():617] starting backend
17
+ 2024-04-10 21:51:01,069 INFO MainThread:227 [wandb_init.py:init():621] setting up manager
18
+ 2024-04-10 21:51:01,071 INFO MainThread:227 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
19
+ 2024-04-10 21:51:01,073 INFO MainThread:227 [wandb_init.py:init():629] backend started and connected
20
+ 2024-04-10 21:51:01,085 INFO MainThread:227 [wandb_run.py:_label_probe_notebook():1299] probe notebook
21
+ 2024-04-10 21:51:01,447 INFO MainThread:227 [wandb_init.py:init():721] updated telemetry
22
+ 2024-04-10 21:51:01,458 INFO MainThread:227 [wandb_init.py:init():754] communicating run to backend with 90.0 second timeout
23
+ 2024-04-10 21:51:01,677 INFO MainThread:227 [wandb_run.py:_on_init():2344] communicating current version
24
+ 2024-04-10 21:51:01,771 INFO MainThread:227 [wandb_run.py:_on_init():2353] got version response upgrade_message: "wandb version 0.16.6 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
25
 
26
+ 2024-04-10 21:51:01,771 INFO MainThread:227 [wandb_init.py:init():805] starting run threads in backend
27
+ 2024-04-10 21:51:17,777 INFO MainThread:227 [wandb_run.py:_console_start():2323] atexit reg
28
+ 2024-04-10 21:51:17,777 INFO MainThread:227 [wandb_run.py:_redirect():2178] redirect: wrap_raw
29
+ 2024-04-10 21:51:17,777 INFO MainThread:227 [wandb_run.py:_redirect():2243] Wrapping output streams.
30
+ 2024-04-10 21:51:17,778 INFO MainThread:227 [wandb_run.py:_redirect():2268] Redirects installed.
31
+ 2024-04-10 21:51:17,779 INFO MainThread:227 [wandb_init.py:init():848] run started, returning control to user process
32
+ 2024-04-10 21:51:17,784 INFO MainThread:227 [wandb_run.py:_config_callback():1347] config_cb None None {'vocab_size': 50304, 'hidden_size': 2048, 'num_hidden_layers': 24, 'num_attention_heads': 32, 'layer_norm_epsilon': 1e-05, 'initializer_range': 0.02, 'use_cache': False, 'hidden_dropout': 0.0, 'attention_dropout': 0.0, 'bos_token_id': 1, 'eos_token_id': 2, 'num_kv_heads': 32, 'alibi': True, 'new_decoder_architecture': False, 'multi_query': False, 'parallel_attn': False, 'bias': True, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'bfloat16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['FalconForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'pad_token_id': None, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'tiiuae/falcon-rw-1b', 'transformers_version': '4.39.3', 'apply_residual_connection_post_layernorm': False, 'auto_map': {'AutoConfig': 'tiiuae/falcon-rw-1b--configuration_falcon.FalconConfig', 'AutoModel': 'tiiuae/falcon-rw-1b--modeling_falcon.FalconModel', 'AutoModelForSequenceClassification': 'tiiuae/falcon-rw-1b--modeling_falcon.FalconForSequenceClassification', 'AutoModelForTokenClassification': 'tiiuae/falcon-rw-1b--modeling_falcon.FalconForTokenClassification', 'AutoModelForQuestionAnswering': 'tiiuae/falcon-rw-1b--modeling_falcon.FalconForQuestionAnswering', 'AutoModelForCausalLM': 'tiiuae/falcon-rw-1b--modeling_falcon.FalconForCausalLM'}, 'model_type': 'falcon', 'quantization_config': {'quant_method': 'QuantizationMethod.BITS_AND_BYTES', '_load_in_8bit': False, '_load_in_4bit': True, 'llm_int8_threshold': 6.0, 'llm_int8_skip_modules': None, 'llm_int8_enable_fp32_cpu_offload': False, 'llm_int8_has_fp16_weight': False, 'bnb_4bit_quant_type': 'nf4', 'bnb_4bit_use_double_quant': False, 'bnb_4bit_compute_dtype': 'float16', 'bnb_4bit_quant_storage': 'uint8', 'load_in_4bit': True, 'load_in_8bit': False}, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': False, 'do_predict': False, 'evaluation_strategy': 'no', 'prediction_loss_only': False, 'per_device_train_batch_size': 8, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 3e-05, 'weight_decay': 0.0001, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 0.3, 'num_train_epochs': 5, 'max_steps': 20, 'lr_scheduler_type': 'cosine', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.03, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Apr10_21-42-21_4bf7c58fba03', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 20, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 20, 'save_total_limit': 1, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 8, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'paged_adamw_8bit', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': True, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': True, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None}
33
+ 2024-04-10 21:51:22,848 INFO MainThread:227 [jupyter.py:save_ipynb():373] not saving jupyter notebook
34
+ 2024-04-10 21:51:22,848 INFO MainThread:227 [wandb_init.py:_pause_backend():438] pausing backend
35
+ 2024-04-10 21:52:40,758 INFO MainThread:227 [wandb_init.py:_resume_backend():443] resuming backend
36
+ 2024-04-10 22:00:44,550 INFO MainThread:227 [wandb_run.py:_config_callback():1347] config_cb None None {'vocab_size': 50304, 'hidden_size': 2048, 'num_hidden_layers': 24, 'num_attention_heads': 32, 'layer_norm_epsilon': 1e-05, 'initializer_range': 0.02, 'use_cache': False, 'hidden_dropout': 0.0, 'attention_dropout': 0.0, 'bos_token_id': 1, 'eos_token_id': 2, 'num_kv_heads': 32, 'alibi': True, 'new_decoder_architecture': False, 'multi_query': False, 'parallel_attn': False, 'bias': True, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'bfloat16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['FalconForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'pad_token_id': None, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'tiiuae/falcon-rw-1b', 'transformers_version': '4.39.3', 'apply_residual_connection_post_layernorm': False, 'auto_map': {'AutoConfig': 'tiiuae/falcon-rw-1b--configuration_falcon.FalconConfig', 'AutoModel': 'tiiuae/falcon-rw-1b--modeling_falcon.FalconModel', 'AutoModelForSequenceClassification': 'tiiuae/falcon-rw-1b--modeling_falcon.FalconForSequenceClassification', 'AutoModelForTokenClassification': 'tiiuae/falcon-rw-1b--modeling_falcon.FalconForTokenClassification', 'AutoModelForQuestionAnswering': 'tiiuae/falcon-rw-1b--modeling_falcon.FalconForQuestionAnswering', 'AutoModelForCausalLM': 'tiiuae/falcon-rw-1b--modeling_falcon.FalconForCausalLM'}, 'model_type': 'falcon', 'quantization_config': {'quant_method': 'QuantizationMethod.BITS_AND_BYTES', '_load_in_8bit': False, '_load_in_4bit': True, 'llm_int8_threshold': 6.0, 'llm_int8_skip_modules': None, 'llm_int8_enable_fp32_cpu_offload': False, 'llm_int8_has_fp16_weight': False, 'bnb_4bit_quant_type': 'nf4', 'bnb_4bit_use_double_quant': False, 'bnb_4bit_compute_dtype': 'float16', 'bnb_4bit_quant_storage': 'uint8', 'load_in_4bit': True, 'load_in_8bit': False}, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': False, 'do_predict': False, 'evaluation_strategy': 'no', 'prediction_loss_only': False, 'per_device_train_batch_size': 8, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 3e-05, 'weight_decay': 0.0001, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 0.3, 'num_train_epochs': 5, 'max_steps': 20, 'lr_scheduler_type': 'cosine', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.03, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Apr10_21-52-40_4bf7c58fba03', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 20, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 20, 'save_total_limit': 1, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 8, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'paged_adamw_8bit', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': True, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': True, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None}
wandb/run-20240410_215101-ta1pahei/files/conda-environment.yaml ADDED
File without changes
wandb/run-20240410_215101-ta1pahei/files/config.yaml ADDED
@@ -0,0 +1,696 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wandb_version: 1
2
+
3
+ _wandb:
4
+ desc: null
5
+ value:
6
+ python_version: 3.10.13
7
+ cli_version: 0.16.5
8
+ framework: huggingface
9
+ huggingface_version: 4.39.3
10
+ is_jupyter_run: true
11
+ is_kaggle_kernel: true
12
+ start_time: 1712785861.0
13
+ t:
14
+ 1:
15
+ - 1
16
+ - 2
17
+ - 3
18
+ - 5
19
+ - 11
20
+ - 12
21
+ - 49
22
+ - 51
23
+ - 53
24
+ - 55
25
+ - 71
26
+ - 84
27
+ - 98
28
+ - 105
29
+ 2:
30
+ - 1
31
+ - 2
32
+ - 3
33
+ - 5
34
+ - 11
35
+ - 12
36
+ - 49
37
+ - 51
38
+ - 53
39
+ - 55
40
+ - 71
41
+ - 84
42
+ - 98
43
+ - 105
44
+ 3:
45
+ - 7
46
+ - 23
47
+ - 62
48
+ 4: 3.10.13
49
+ 5: 0.16.5
50
+ 6: 4.39.3
51
+ 8:
52
+ - 1
53
+ - 2
54
+ - 5
55
+ 9:
56
+ 1: transformers_trainer
57
+ 13: linux-x86_64
58
+ m:
59
+ - 1: train/global_step
60
+ 6:
61
+ - 3
62
+ - 1: train/loss
63
+ 5: 1
64
+ 6:
65
+ - 1
66
+ - 1: train/grad_norm
67
+ 5: 1
68
+ 6:
69
+ - 1
70
+ - 1: train/learning_rate
71
+ 5: 1
72
+ 6:
73
+ - 1
74
+ - 1: train/epoch
75
+ 5: 1
76
+ 6:
77
+ - 1
78
+ vocab_size:
79
+ desc: null
80
+ value: 50304
81
+ hidden_size:
82
+ desc: null
83
+ value: 2048
84
+ num_hidden_layers:
85
+ desc: null
86
+ value: 24
87
+ num_attention_heads:
88
+ desc: null
89
+ value: 32
90
+ layer_norm_epsilon:
91
+ desc: null
92
+ value: 1.0e-05
93
+ initializer_range:
94
+ desc: null
95
+ value: 0.02
96
+ use_cache:
97
+ desc: null
98
+ value: false
99
+ hidden_dropout:
100
+ desc: null
101
+ value: 0.0
102
+ attention_dropout:
103
+ desc: null
104
+ value: 0.0
105
+ bos_token_id:
106
+ desc: null
107
+ value: 1
108
+ eos_token_id:
109
+ desc: null
110
+ value: 2
111
+ num_kv_heads:
112
+ desc: null
113
+ value: 32
114
+ alibi:
115
+ desc: null
116
+ value: true
117
+ new_decoder_architecture:
118
+ desc: null
119
+ value: false
120
+ multi_query:
121
+ desc: null
122
+ value: false
123
+ parallel_attn:
124
+ desc: null
125
+ value: false
126
+ bias:
127
+ desc: null
128
+ value: true
129
+ return_dict:
130
+ desc: null
131
+ value: true
132
+ output_hidden_states:
133
+ desc: null
134
+ value: false
135
+ output_attentions:
136
+ desc: null
137
+ value: false
138
+ torchscript:
139
+ desc: null
140
+ value: false
141
+ torch_dtype:
142
+ desc: null
143
+ value: bfloat16
144
+ use_bfloat16:
145
+ desc: null
146
+ value: false
147
+ tf_legacy_loss:
148
+ desc: null
149
+ value: false
150
+ pruned_heads:
151
+ desc: null
152
+ value: {}
153
+ tie_word_embeddings:
154
+ desc: null
155
+ value: true
156
+ chunk_size_feed_forward:
157
+ desc: null
158
+ value: 0
159
+ is_encoder_decoder:
160
+ desc: null
161
+ value: false
162
+ is_decoder:
163
+ desc: null
164
+ value: false
165
+ cross_attention_hidden_size:
166
+ desc: null
167
+ value: null
168
+ add_cross_attention:
169
+ desc: null
170
+ value: false
171
+ tie_encoder_decoder:
172
+ desc: null
173
+ value: false
174
+ max_length:
175
+ desc: null
176
+ value: 20
177
+ min_length:
178
+ desc: null
179
+ value: 0
180
+ do_sample:
181
+ desc: null
182
+ value: false
183
+ early_stopping:
184
+ desc: null
185
+ value: false
186
+ num_beams:
187
+ desc: null
188
+ value: 1
189
+ num_beam_groups:
190
+ desc: null
191
+ value: 1
192
+ diversity_penalty:
193
+ desc: null
194
+ value: 0.0
195
+ temperature:
196
+ desc: null
197
+ value: 1.0
198
+ top_k:
199
+ desc: null
200
+ value: 50
201
+ top_p:
202
+ desc: null
203
+ value: 1.0
204
+ typical_p:
205
+ desc: null
206
+ value: 1.0
207
+ repetition_penalty:
208
+ desc: null
209
+ value: 1.0
210
+ length_penalty:
211
+ desc: null
212
+ value: 1.0
213
+ no_repeat_ngram_size:
214
+ desc: null
215
+ value: 0
216
+ encoder_no_repeat_ngram_size:
217
+ desc: null
218
+ value: 0
219
+ bad_words_ids:
220
+ desc: null
221
+ value: null
222
+ num_return_sequences:
223
+ desc: null
224
+ value: 1
225
+ output_scores:
226
+ desc: null
227
+ value: false
228
+ return_dict_in_generate:
229
+ desc: null
230
+ value: false
231
+ forced_bos_token_id:
232
+ desc: null
233
+ value: null
234
+ forced_eos_token_id:
235
+ desc: null
236
+ value: null
237
+ remove_invalid_values:
238
+ desc: null
239
+ value: false
240
+ exponential_decay_length_penalty:
241
+ desc: null
242
+ value: null
243
+ suppress_tokens:
244
+ desc: null
245
+ value: null
246
+ begin_suppress_tokens:
247
+ desc: null
248
+ value: null
249
+ architectures:
250
+ desc: null
251
+ value:
252
+ - FalconForCausalLM
253
+ finetuning_task:
254
+ desc: null
255
+ value: null
256
+ id2label:
257
+ desc: null
258
+ value:
259
+ '0': LABEL_0
260
+ '1': LABEL_1
261
+ label2id:
262
+ desc: null
263
+ value:
264
+ LABEL_0: 0
265
+ LABEL_1: 1
266
+ tokenizer_class:
267
+ desc: null
268
+ value: null
269
+ prefix:
270
+ desc: null
271
+ value: null
272
+ pad_token_id:
273
+ desc: null
274
+ value: null
275
+ sep_token_id:
276
+ desc: null
277
+ value: null
278
+ decoder_start_token_id:
279
+ desc: null
280
+ value: null
281
+ task_specific_params:
282
+ desc: null
283
+ value: null
284
+ problem_type:
285
+ desc: null
286
+ value: null
287
+ _name_or_path:
288
+ desc: null
289
+ value: tiiuae/falcon-rw-1b
290
+ transformers_version:
291
+ desc: null
292
+ value: 4.39.3
293
+ apply_residual_connection_post_layernorm:
294
+ desc: null
295
+ value: false
296
+ auto_map:
297
+ desc: null
298
+ value:
299
+ AutoConfig: tiiuae/falcon-rw-1b--configuration_falcon.FalconConfig
300
+ AutoModel: tiiuae/falcon-rw-1b--modeling_falcon.FalconModel
301
+ AutoModelForSequenceClassification: tiiuae/falcon-rw-1b--modeling_falcon.FalconForSequenceClassification
302
+ AutoModelForTokenClassification: tiiuae/falcon-rw-1b--modeling_falcon.FalconForTokenClassification
303
+ AutoModelForQuestionAnswering: tiiuae/falcon-rw-1b--modeling_falcon.FalconForQuestionAnswering
304
+ AutoModelForCausalLM: tiiuae/falcon-rw-1b--modeling_falcon.FalconForCausalLM
305
+ model_type:
306
+ desc: null
307
+ value: falcon
308
+ quantization_config:
309
+ desc: null
310
+ value:
311
+ quant_method: QuantizationMethod.BITS_AND_BYTES
312
+ _load_in_8bit: false
313
+ _load_in_4bit: true
314
+ llm_int8_threshold: 6.0
315
+ llm_int8_skip_modules: null
316
+ llm_int8_enable_fp32_cpu_offload: false
317
+ llm_int8_has_fp16_weight: false
318
+ bnb_4bit_quant_type: nf4
319
+ bnb_4bit_use_double_quant: false
320
+ bnb_4bit_compute_dtype: float16
321
+ bnb_4bit_quant_storage: uint8
322
+ load_in_4bit: true
323
+ load_in_8bit: false
324
+ output_dir:
325
+ desc: null
326
+ value: /kaggle/working/
327
+ overwrite_output_dir:
328
+ desc: null
329
+ value: false
330
+ do_train:
331
+ desc: null
332
+ value: false
333
+ do_eval:
334
+ desc: null
335
+ value: false
336
+ do_predict:
337
+ desc: null
338
+ value: false
339
+ evaluation_strategy:
340
+ desc: null
341
+ value: 'no'
342
+ prediction_loss_only:
343
+ desc: null
344
+ value: false
345
+ per_device_train_batch_size:
346
+ desc: null
347
+ value: 8
348
+ per_device_eval_batch_size:
349
+ desc: null
350
+ value: 8
351
+ per_gpu_train_batch_size:
352
+ desc: null
353
+ value: null
354
+ per_gpu_eval_batch_size:
355
+ desc: null
356
+ value: null
357
+ gradient_accumulation_steps:
358
+ desc: null
359
+ value: 1
360
+ eval_accumulation_steps:
361
+ desc: null
362
+ value: null
363
+ eval_delay:
364
+ desc: null
365
+ value: 0
366
+ learning_rate:
367
+ desc: null
368
+ value: 3.0e-05
369
+ weight_decay:
370
+ desc: null
371
+ value: 0.0001
372
+ adam_beta1:
373
+ desc: null
374
+ value: 0.9
375
+ adam_beta2:
376
+ desc: null
377
+ value: 0.999
378
+ adam_epsilon:
379
+ desc: null
380
+ value: 1.0e-08
381
+ max_grad_norm:
382
+ desc: null
383
+ value: 0.3
384
+ num_train_epochs:
385
+ desc: null
386
+ value: 5
387
+ max_steps:
388
+ desc: null
389
+ value: 20
390
+ lr_scheduler_type:
391
+ desc: null
392
+ value: cosine
393
+ lr_scheduler_kwargs:
394
+ desc: null
395
+ value: {}
396
+ warmup_ratio:
397
+ desc: null
398
+ value: 0.03
399
+ warmup_steps:
400
+ desc: null
401
+ value: 0
402
+ log_level:
403
+ desc: null
404
+ value: passive
405
+ log_level_replica:
406
+ desc: null
407
+ value: warning
408
+ log_on_each_node:
409
+ desc: null
410
+ value: true
411
+ logging_dir:
412
+ desc: null
413
+ value: /kaggle/working/runs/Apr10_21-52-40_4bf7c58fba03
414
+ logging_strategy:
415
+ desc: null
416
+ value: steps
417
+ logging_first_step:
418
+ desc: null
419
+ value: false
420
+ logging_steps:
421
+ desc: null
422
+ value: 20
423
+ logging_nan_inf_filter:
424
+ desc: null
425
+ value: true
426
+ save_strategy:
427
+ desc: null
428
+ value: steps
429
+ save_steps:
430
+ desc: null
431
+ value: 20
432
+ save_total_limit:
433
+ desc: null
434
+ value: 1
435
+ save_safetensors:
436
+ desc: null
437
+ value: true
438
+ save_on_each_node:
439
+ desc: null
440
+ value: false
441
+ save_only_model:
442
+ desc: null
443
+ value: false
444
+ no_cuda:
445
+ desc: null
446
+ value: false
447
+ use_cpu:
448
+ desc: null
449
+ value: false
450
+ use_mps_device:
451
+ desc: null
452
+ value: false
453
+ seed:
454
+ desc: null
455
+ value: 42
456
+ data_seed:
457
+ desc: null
458
+ value: null
459
+ jit_mode_eval:
460
+ desc: null
461
+ value: false
462
+ use_ipex:
463
+ desc: null
464
+ value: false
465
+ bf16:
466
+ desc: null
467
+ value: false
468
+ fp16:
469
+ desc: null
470
+ value: false
471
+ fp16_opt_level:
472
+ desc: null
473
+ value: O1
474
+ half_precision_backend:
475
+ desc: null
476
+ value: auto
477
+ bf16_full_eval:
478
+ desc: null
479
+ value: false
480
+ fp16_full_eval:
481
+ desc: null
482
+ value: false
483
+ tf32:
484
+ desc: null
485
+ value: null
486
+ local_rank:
487
+ desc: null
488
+ value: 0
489
+ ddp_backend:
490
+ desc: null
491
+ value: null
492
+ tpu_num_cores:
493
+ desc: null
494
+ value: null
495
+ tpu_metrics_debug:
496
+ desc: null
497
+ value: false
498
+ debug:
499
+ desc: null
500
+ value: []
501
+ dataloader_drop_last:
502
+ desc: null
503
+ value: false
504
+ eval_steps:
505
+ desc: null
506
+ value: null
507
+ dataloader_num_workers:
508
+ desc: null
509
+ value: 8
510
+ dataloader_prefetch_factor:
511
+ desc: null
512
+ value: null
513
+ past_index:
514
+ desc: null
515
+ value: -1
516
+ run_name:
517
+ desc: null
518
+ value: /kaggle/working/
519
+ disable_tqdm:
520
+ desc: null
521
+ value: false
522
+ remove_unused_columns:
523
+ desc: null
524
+ value: true
525
+ label_names:
526
+ desc: null
527
+ value: null
528
+ load_best_model_at_end:
529
+ desc: null
530
+ value: false
531
+ metric_for_best_model:
532
+ desc: null
533
+ value: null
534
+ greater_is_better:
535
+ desc: null
536
+ value: null
537
+ ignore_data_skip:
538
+ desc: null
539
+ value: false
540
+ fsdp:
541
+ desc: null
542
+ value: []
543
+ fsdp_min_num_params:
544
+ desc: null
545
+ value: 0
546
+ fsdp_config:
547
+ desc: null
548
+ value:
549
+ min_num_params: 0
550
+ xla: false
551
+ xla_fsdp_v2: false
552
+ xla_fsdp_grad_ckpt: false
553
+ fsdp_transformer_layer_cls_to_wrap:
554
+ desc: null
555
+ value: null
556
+ accelerator_config:
557
+ desc: null
558
+ value:
559
+ split_batches: false
560
+ dispatch_batches: null
561
+ even_batches: true
562
+ use_seedable_sampler: true
563
+ deepspeed:
564
+ desc: null
565
+ value: null
566
+ label_smoothing_factor:
567
+ desc: null
568
+ value: 0.0
569
+ optim:
570
+ desc: null
571
+ value: paged_adamw_8bit
572
+ optim_args:
573
+ desc: null
574
+ value: null
575
+ adafactor:
576
+ desc: null
577
+ value: false
578
+ group_by_length:
579
+ desc: null
580
+ value: false
581
+ length_column_name:
582
+ desc: null
583
+ value: length
584
+ report_to:
585
+ desc: null
586
+ value:
587
+ - tensorboard
588
+ - wandb
589
+ ddp_find_unused_parameters:
590
+ desc: null
591
+ value: null
592
+ ddp_bucket_cap_mb:
593
+ desc: null
594
+ value: null
595
+ ddp_broadcast_buffers:
596
+ desc: null
597
+ value: null
598
+ dataloader_pin_memory:
599
+ desc: null
600
+ value: true
601
+ dataloader_persistent_workers:
602
+ desc: null
603
+ value: false
604
+ skip_memory_metrics:
605
+ desc: null
606
+ value: true
607
+ use_legacy_prediction_loop:
608
+ desc: null
609
+ value: false
610
+ push_to_hub:
611
+ desc: null
612
+ value: false
613
+ resume_from_checkpoint:
614
+ desc: null
615
+ value: null
616
+ hub_model_id:
617
+ desc: null
618
+ value: null
619
+ hub_strategy:
620
+ desc: null
621
+ value: every_save
622
+ hub_token:
623
+ desc: null
624
+ value: <HUB_TOKEN>
625
+ hub_private_repo:
626
+ desc: null
627
+ value: false
628
+ hub_always_push:
629
+ desc: null
630
+ value: false
631
+ gradient_checkpointing:
632
+ desc: null
633
+ value: true
634
+ gradient_checkpointing_kwargs:
635
+ desc: null
636
+ value: null
637
+ include_inputs_for_metrics:
638
+ desc: null
639
+ value: false
640
+ fp16_backend:
641
+ desc: null
642
+ value: auto
643
+ push_to_hub_model_id:
644
+ desc: null
645
+ value: null
646
+ push_to_hub_organization:
647
+ desc: null
648
+ value: null
649
+ push_to_hub_token:
650
+ desc: null
651
+ value: <PUSH_TO_HUB_TOKEN>
652
+ mp_parameters:
653
+ desc: null
654
+ value: ''
655
+ auto_find_batch_size:
656
+ desc: null
657
+ value: true
658
+ full_determinism:
659
+ desc: null
660
+ value: false
661
+ torchdynamo:
662
+ desc: null
663
+ value: null
664
+ ray_scope:
665
+ desc: null
666
+ value: last
667
+ ddp_timeout:
668
+ desc: null
669
+ value: 1800
670
+ torch_compile:
671
+ desc: null
672
+ value: false
673
+ torch_compile_backend:
674
+ desc: null
675
+ value: null
676
+ torch_compile_mode:
677
+ desc: null
678
+ value: null
679
+ dispatch_batches:
680
+ desc: null
681
+ value: null
682
+ split_batches:
683
+ desc: null
684
+ value: null
685
+ include_tokens_per_second:
686
+ desc: null
687
+ value: false
688
+ include_num_input_tokens_seen:
689
+ desc: null
690
+ value: false
691
+ neftune_noise_alpha:
692
+ desc: null
693
+ value: null
694
+ optim_target_modules:
695
+ desc: null
696
+ value: null
wandb/run-20240410_215101-ta1pahei/files/output.log ADDED
@@ -0,0 +1,146 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ /opt/conda/lib/python3.10/site-packages/torch/utils/data/dataloader.py:557: UserWarning: This DataLoader will create 8 worker processes in total. Our suggested max number of worker in current system is 4, which is smaller than what this DataLoader is going to create. Please be aware that excessive worker creation might get DataLoader running slow or even freeze, lower the worker number to avoid potential slowness/freeze if necessary.
2
+ warnings.warn(_create_warning_msg(
3
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
4
+ To disable this warning, you can either:
5
+ - Avoid using `tokenizers` before the fork if possible
6
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
7
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
8
+ To disable this warning, you can either:
9
+ - Avoid using `tokenizers` before the fork if possible
10
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
11
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
12
+ To disable this warning, you can either:
13
+ - Avoid using `tokenizers` before the fork if possible
14
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
15
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
16
+ To disable this warning, you can either:
17
+ - Avoid using `tokenizers` before the fork if possible
18
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
19
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
20
+ To disable this warning, you can either:
21
+ - Avoid using `tokenizers` before the fork if possible
22
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
23
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
24
+ To disable this warning, you can either:
25
+ - Avoid using `tokenizers` before the fork if possible
26
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
27
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
28
+ To disable this warning, you can either:
29
+ - Avoid using `tokenizers` before the fork if possible
30
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
31
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
32
+ To disable this warning, you can either:
33
+ - Avoid using `tokenizers` before the fork if possible
34
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
35
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
36
+ warnings.warn(
37
+ You are using an old version of the checkpointing format that is deprecated (We will also silently ignore `gradient_checkpointing_kwargs` in case you passed it).Please update to the new format on your modeling file. To use the new format, you need to completely remove the definition of the method `_set_gradient_checkpointing` in your model.
38
+ You are using an old version of the checkpointing format that is deprecated (We will also silently ignore `gradient_checkpointing_kwargs` in case you passed it).Please update to the new format on your modeling file. To use the new format, you need to completely remove the definition of the method `_set_gradient_checkpointing` in your model.
39
+ Token indices sequence length is longer than the specified maximum sequence length for this model (14974 > 8192). Running this sequence through the model will result in indexing errors
40
+ /opt/conda/lib/python3.10/site-packages/accelerate/accelerator.py:432: FutureWarning: Passing the following arguments to `Accelerator` is deprecated and will be removed in version 1.0 of Accelerate: dict_keys(['dispatch_batches', 'split_batches', 'even_batches', 'use_seedable_sampler']). Please pass an `accelerate.DataLoaderConfiguration` instead:
41
+ dataloader_config = DataLoaderConfiguration(dispatch_batches=None, split_batches=False, even_batches=True, use_seedable_sampler=True)
42
+ warnings.warn(
43
+ You are using an old version of the checkpointing format that is deprecated (We will also silently ignore `gradient_checkpointing_kwargs` in case you passed it).Please update to the new format on your modeling file. To use the new format, you need to completely remove the definition of the method `_set_gradient_checkpointing` in your model.
44
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
45
+ To disable this warning, you can either:
46
+ - Avoid using `tokenizers` before the fork if possible
47
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
48
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
49
+ To disable this warning, you can either:
50
+ - Avoid using `tokenizers` before the fork if possible
51
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
52
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
53
+ To disable this warning, you can either:
54
+ - Avoid using `tokenizers` before the fork if possible
55
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
56
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
57
+ To disable this warning, you can either:
58
+ - Avoid using `tokenizers` before the fork if possible
59
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
60
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
61
+ To disable this warning, you can either:
62
+ - Avoid using `tokenizers` before the fork if possible
63
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
64
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
65
+ To disable this warning, you can either:
66
+ - Avoid using `tokenizers` before the fork if possible
67
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
68
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
69
+ To disable this warning, you can either:
70
+ - Avoid using `tokenizers` before the fork if possible
71
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
72
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
73
+ To disable this warning, you can either:
74
+ - Avoid using `tokenizers` before the fork if possible
75
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
76
+ /opt/conda/lib/python3.10/site-packages/torch/nn/parallel/_functions.py:68: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
77
+ warnings.warn('Was asked to gather along dimension 0, but all '
78
+ You are using an old version of the checkpointing format that is deprecated (We will also silently ignore `gradient_checkpointing_kwargs` in case you passed it).Please update to the new format on your modeling file. To use the new format, you need to completely remove the definition of the method `_set_gradient_checkpointing` in your model.
79
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
80
+ To disable this warning, you can either:
81
+ - Avoid using `tokenizers` before the fork if possible
82
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
83
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
84
+ To disable this warning, you can either:
85
+ - Avoid using `tokenizers` before the fork if possible
86
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
87
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
88
+ To disable this warning, you can either:
89
+ - Avoid using `tokenizers` before the fork if possible
90
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
91
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
92
+ To disable this warning, you can either:
93
+ - Avoid using `tokenizers` before the fork if possible
94
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
95
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
96
+ To disable this warning, you can either:
97
+ - Avoid using `tokenizers` before the fork if possible
98
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
99
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
100
+ To disable this warning, you can either:
101
+ - Avoid using `tokenizers` before the fork if possible
102
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
103
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
104
+ To disable this warning, you can either:
105
+ - Avoid using `tokenizers` before the fork if possible
106
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
107
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
108
+ To disable this warning, you can either:
109
+ - Avoid using `tokenizers` before the fork if possible
110
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
111
+ You are using an old version of the checkpointing format that is deprecated (We will also silently ignore `gradient_checkpointing_kwargs` in case you passed it).Please update to the new format on your modeling file. To use the new format, you need to completely remove the definition of the method `_set_gradient_checkpointing` in your model.
112
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
113
+ To disable this warning, you can either:
114
+ - Avoid using `tokenizers` before the fork if possible
115
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
116
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
117
+ To disable this warning, you can either:
118
+ - Avoid using `tokenizers` before the fork if possible
119
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
120
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
121
+ To disable this warning, you can either:
122
+ - Avoid using `tokenizers` before the fork if possible
123
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
124
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
125
+ To disable this warning, you can either:
126
+ - Avoid using `tokenizers` before the fork if possible
127
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
128
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
129
+ To disable this warning, you can either:
130
+ - Avoid using `tokenizers` before the fork if possible
131
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
132
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
133
+ To disable this warning, you can either:
134
+ - Avoid using `tokenizers` before the fork if possible
135
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
136
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
137
+ To disable this warning, you can either:
138
+ - Avoid using `tokenizers` before the fork if possible
139
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
140
+ huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
141
+ To disable this warning, you can either:
142
+ - Avoid using `tokenizers` before the fork if possible
143
+ - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
144
+ /opt/conda/lib/python3.10/site-packages/peft/utils/save_and_load.py:139: UserWarning: Setting `save_embedding_layers` to `True` as embedding layers found in `target_modules`.
145
+ warnings.warn("Setting `save_embedding_layers` to `True` as embedding layers found in `target_modules`.")
146
+ /opt/conda/lib/python3.10/site-packages/peft/utils/save_and_load.py:139: UserWarning: Setting `save_embedding_layers` to `True` as embedding layers found in `target_modules`.
wandb/run-20240410_215101-ta1pahei/files/requirements.txt ADDED
@@ -0,0 +1,865 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Babel==2.14.0
2
+ Boruta==0.3
3
+ Brotli==1.0.9
4
+ CVXcanon==0.1.2
5
+ Cartopy==0.22.0
6
+ Cython==3.0.8
7
+ Deprecated==1.2.14
8
+ Farama-Notifications==0.0.4
9
+ Flask==3.0.2
10
+ Geohash==1.0
11
+ GitPython==3.1.41
12
+ ImageHash==4.3.1
13
+ Janome==0.5.0
14
+ Jinja2==3.1.2
15
+ LunarCalendar==0.0.9
16
+ Mako==1.3.2
17
+ Markdown==3.5.2
18
+ MarkupSafe==2.1.3
19
+ MarkupSafe==2.1.5
20
+ Pillow==9.5.0
21
+ PuLP==2.8.0
22
+ PyArabic==0.6.15
23
+ PyJWT==2.8.0
24
+ PyMeeus==0.5.12
25
+ PySocks==1.7.1
26
+ PyUpSet==0.1.1.post7
27
+ PyWavelets==1.5.0
28
+ PyYAML==6.0.1
29
+ Pygments==2.17.2
30
+ Pympler==1.0.1
31
+ QtPy==2.4.1
32
+ Rtree==1.2.0
33
+ SQLAlchemy==2.0.25
34
+ SecretStorage==3.3.3
35
+ Send2Trash==1.8.2
36
+ Shapely==1.8.5.post1
37
+ Shimmy==1.3.0
38
+ SimpleITK==2.3.1
39
+ TPOT==0.12.1
40
+ Theano-PyMC==1.1.2
41
+ Theano==1.0.5
42
+ Wand==0.6.13
43
+ Werkzeug==3.0.2
44
+ absl-py==1.4.0
45
+ accelerate==0.28.0
46
+ access==1.1.9
47
+ affine==2.4.0
48
+ aiobotocore==2.12.2
49
+ aiofiles==22.1.0
50
+ aiohttp-cors==0.7.0
51
+ aiohttp==3.9.1
52
+ aioitertools==0.11.0
53
+ aiorwlock==1.3.0
54
+ aiosignal==1.3.1
55
+ aiosqlite==0.19.0
56
+ albumentations==1.4.0
57
+ alembic==1.13.1
58
+ altair==5.3.0
59
+ annotated-types==0.6.0
60
+ annoy==1.17.3
61
+ anyio==4.2.0
62
+ apache-beam==2.46.0
63
+ aplus==0.11.0
64
+ appdirs==1.4.4
65
+ archspec==0.2.3
66
+ argon2-cffi-bindings==21.2.0
67
+ argon2-cffi==23.1.0
68
+ array-record==0.5.0
69
+ arrow==1.3.0
70
+ arviz==0.17.1
71
+ astroid==3.1.0
72
+ astropy-iers-data==0.2024.4.1.0.33.14
73
+ astropy==6.0.1
74
+ asttokens==2.4.1
75
+ astunparse==1.6.3
76
+ async-lru==2.0.4
77
+ async-timeout==4.0.3
78
+ attrs==23.2.0
79
+ audioread==3.0.1
80
+ autopep8==2.0.4
81
+ backoff==2.2.1
82
+ bayesian-optimization==1.4.3
83
+ beatrix_jupyterlab==2023.128.151533
84
+ beautifulsoup4==4.12.2
85
+ bitsandbytes==0.43.0
86
+ blake3==0.2.1
87
+ bleach==6.1.0
88
+ blessed==1.20.0
89
+ blinker==1.7.0
90
+ blis==0.7.10
91
+ blosc2==2.6.0
92
+ bokeh==3.3.4
93
+ boltons==23.1.1
94
+ boto3==1.26.100
95
+ botocore==1.34.51
96
+ bq_helper==0.4.1
97
+ bqplot==0.12.43
98
+ branca==0.7.1
99
+ brewer2mpl==1.4.1
100
+ brotlipy==0.7.0
101
+ cached-property==1.5.2
102
+ cachetools==4.2.4
103
+ cachetools==5.3.2
104
+ catalogue==2.0.10
105
+ catalyst==22.4
106
+ catboost==1.2.3
107
+ category-encoders==2.6.3
108
+ certifi==2024.2.2
109
+ cesium==0.12.1
110
+ cffi==1.16.0
111
+ charset-normalizer==3.3.2
112
+ chex==0.1.86
113
+ cleverhans==4.0.0
114
+ click-plugins==1.1.1
115
+ click==8.1.7
116
+ cligj==0.7.2
117
+ cloud-tpu-client==0.10
118
+ cloud-tpu-profiler==2.4.0
119
+ cloudpathlib==0.16.0
120
+ cloudpickle==2.2.1
121
+ cloudpickle==3.0.0
122
+ cmdstanpy==1.2.2
123
+ colorama==0.4.6
124
+ colorcet==3.1.0
125
+ colorful==0.5.6
126
+ colorlog==6.8.2
127
+ colorlover==0.3.0
128
+ comm==0.2.1
129
+ conda-libmamba-solver==23.7.0
130
+ conda-package-handling==2.2.0
131
+ conda==23.7.4
132
+ conda_package_streaming==0.9.0
133
+ confection==0.1.4
134
+ contextily==1.6.0
135
+ contourpy==1.2.0
136
+ convertdate==2.4.0
137
+ crcmod==1.7
138
+ cryptography==41.0.7
139
+ cuda-python==12.4.0
140
+ cudf==23.8.0
141
+ cufflinks==0.17.3
142
+ cuml==23.8.0
143
+ cupy==13.0.0
144
+ cycler==0.12.1
145
+ cymem==2.0.8
146
+ cytoolz==0.12.3
147
+ daal4py==2024.2.0
148
+ daal==2024.2.0
149
+ dacite==1.8.1
150
+ dask-cuda==23.8.0
151
+ dask-cudf==23.8.0
152
+ dask-expr==1.0.9
153
+ dask==2024.4.0
154
+ dataclasses-json==0.6.4
155
+ dataproc_jupyter_plugin==0.1.66
156
+ datasets==2.16.0
157
+ datashader==0.16.0
158
+ datatile==1.0.3
159
+ db-dtypes==1.2.0
160
+ deap==1.4.1
161
+ debugpy==1.8.0
162
+ decorator==5.1.1
163
+ deepdiff==6.7.1
164
+ defusedxml==0.7.1
165
+ deprecation==2.1.0
166
+ descartes==1.1.0
167
+ dill==0.3.7
168
+ dipy==1.9.0
169
+ distlib==0.3.8
170
+ distributed==2023.7.1
171
+ distro==1.9.0
172
+ dm-tree==0.1.8
173
+ docker-pycreds==0.4.0
174
+ docker==7.0.0
175
+ docopt==0.6.2
176
+ docstring-parser==0.15
177
+ docstring-to-markdown==0.15
178
+ docutils==0.20.1
179
+ earthengine-api==0.1.395
180
+ easydict==1.13
181
+ easyocr==1.7.1
182
+ ecos==2.0.13
183
+ einops==0.7.0
184
+ eli5==0.13.0
185
+ emoji==2.11.0
186
+ en-core-web-lg==3.7.1
187
+ en-core-web-sm==3.7.1
188
+ entrypoints==0.4
189
+ ephem==4.1.5
190
+ esda==2.5.1
191
+ essentia==2.1b6.dev1110
192
+ et-xmlfile==1.1.0
193
+ etils==1.6.0
194
+ exceptiongroup==1.2.0
195
+ executing==2.0.1
196
+ explainable-ai-sdk==1.3.3
197
+ fastai==2.7.14
198
+ fastapi==0.108.0
199
+ fastavro==1.9.3
200
+ fastcore==1.5.29
201
+ fastdownload==0.0.7
202
+ fasteners==0.19
203
+ fastjsonschema==2.19.1
204
+ fastprogress==1.0.3
205
+ fastrlock==0.8.2
206
+ fasttext==0.9.2
207
+ feather-format==0.4.1
208
+ featuretools==1.30.0
209
+ filelock==3.13.1
210
+ fiona==1.9.6
211
+ fitter==1.7.0
212
+ flake8==7.0.0
213
+ flashtext==2.7
214
+ flatbuffers==23.5.26
215
+ flax==0.8.2
216
+ folium==0.16.0
217
+ fonttools==4.47.0
218
+ fonttools==4.50.0
219
+ fqdn==1.5.1
220
+ frozendict==2.4.1
221
+ frozenlist==1.4.1
222
+ fsspec==2023.10.0
223
+ fsspec==2024.3.1
224
+ funcy==2.0
225
+ fury==0.10.0
226
+ future==1.0.0
227
+ fuzzywuzzy==0.18.0
228
+ gast==0.5.4
229
+ gatspy==0.3
230
+ gcsfs==2024.2.0
231
+ gensim==4.3.2
232
+ geographiclib==2.0
233
+ geojson==3.1.0
234
+ geopandas==0.14.3
235
+ geoplot==0.5.1
236
+ geopy==2.4.1
237
+ geoviews==1.11.1
238
+ ggplot==0.11.5
239
+ giddy==2.3.5
240
+ gitdb==4.0.11
241
+ google-ai-generativelanguage==0.4.0
242
+ google-api-core==2.11.1
243
+ google-api-core==2.18.0
244
+ google-api-python-client==2.125.0
245
+ google-apitools==0.5.31
246
+ google-auth-httplib2==0.2.0
247
+ google-auth-oauthlib==1.2.0
248
+ google-auth==2.26.1
249
+ google-cloud-aiplatform==0.6.0a1
250
+ google-cloud-artifact-registry==1.10.0
251
+ google-cloud-automl==1.0.1
252
+ google-cloud-bigquery==2.34.4
253
+ google-cloud-bigtable==1.7.3
254
+ google-cloud-core==2.4.1
255
+ google-cloud-datastore==2.19.0
256
+ google-cloud-dlp==3.14.0
257
+ google-cloud-jupyter-config==0.0.5
258
+ google-cloud-language==2.13.3
259
+ google-cloud-monitoring==2.18.0
260
+ google-cloud-pubsub==2.19.0
261
+ google-cloud-pubsublite==1.9.0
262
+ google-cloud-recommendations-ai==0.7.1
263
+ google-cloud-resource-manager==1.11.0
264
+ google-cloud-spanner==3.40.1
265
+ google-cloud-storage==1.44.0
266
+ google-cloud-translate==3.12.1
267
+ google-cloud-videointelligence==2.13.3
268
+ google-cloud-vision==2.8.0
269
+ google-crc32c==1.5.0
270
+ google-generativeai==0.4.1
271
+ google-pasta==0.2.0
272
+ google-resumable-media==2.7.0
273
+ googleapis-common-protos==1.62.0
274
+ gplearn==0.4.2
275
+ gpustat==1.0.0
276
+ gpxpy==1.6.2
277
+ graphviz==0.20.3
278
+ greenlet==3.0.3
279
+ grpc-google-iam-v1==0.12.7
280
+ grpcio-status==1.48.1
281
+ grpcio-status==1.48.2
282
+ grpcio==1.51.1
283
+ grpcio==1.60.0
284
+ gviz-api==1.10.0
285
+ gym-notices==0.0.8
286
+ gym==0.26.2
287
+ gymnasium==0.29.0
288
+ h11==0.14.0
289
+ h2o==3.46.0.1
290
+ h5netcdf==1.3.0
291
+ h5py==3.10.0
292
+ haversine==2.8.1
293
+ hdfs==2.7.3
294
+ hep-ml==0.7.2
295
+ hijri-converter==2.3.1
296
+ hmmlearn==0.3.2
297
+ holidays==0.24
298
+ holoviews==1.18.3
299
+ hpsklearn==0.1.0
300
+ html5lib==1.1
301
+ htmlmin==0.1.12
302
+ httpcore==1.0.5
303
+ httplib2==0.21.0
304
+ httptools==0.6.1
305
+ httpx==0.27.0
306
+ huggingface-hub==0.22.2
307
+ hunspell==0.5.5
308
+ hydra-slayer==0.5.0
309
+ hyperopt==0.2.7
310
+ hypertools==0.8.0
311
+ idna==3.6
312
+ igraph==0.11.4
313
+ imagecodecs==2024.1.1
314
+ imageio==2.33.1
315
+ imbalanced-learn==0.12.2
316
+ imgaug==0.4.0
317
+ importlib-metadata==6.11.0
318
+ importlib-metadata==7.0.1
319
+ importlib-resources==6.1.1
320
+ inequality==1.0.1
321
+ iniconfig==2.0.0
322
+ ipydatawidgets==4.3.5
323
+ ipykernel==6.28.0
324
+ ipyleaflet==0.18.2
325
+ ipympl==0.7.0
326
+ ipython-genutils==0.2.0
327
+ ipython-genutils==0.2.0
328
+ ipython-sql==0.5.0
329
+ ipython==8.20.0
330
+ ipyvolume==0.6.3
331
+ ipyvue==1.10.2
332
+ ipyvuetify==1.9.3
333
+ ipywebrtc==0.6.0
334
+ ipywidgets==7.7.1
335
+ isoduration==20.11.0
336
+ isort==5.13.2
337
+ isoweek==1.3.3
338
+ itsdangerous==2.1.2
339
+ jaraco.classes==3.3.0
340
+ jax-jumpy==1.0.0
341
+ jax==0.4.23
342
+ jaxlib==0.4.23.dev20240116
343
+ jedi==0.19.1
344
+ jeepney==0.8.0
345
+ jieba==0.42.1
346
+ jmespath==1.0.1
347
+ joblib==1.3.2
348
+ json5==0.9.14
349
+ jsonpatch==1.33
350
+ jsonpointer==2.4
351
+ jsonschema-specifications==2023.12.1
352
+ jsonschema==4.20.0
353
+ jupyter-console==6.6.3
354
+ jupyter-events==0.9.0
355
+ jupyter-http-over-ws==0.0.8
356
+ jupyter-lsp==1.5.1
357
+ jupyter-server-mathjax==0.2.6
358
+ jupyter-ydoc==0.2.5
359
+ jupyter_client==7.4.9
360
+ jupyter_client==8.6.0
361
+ jupyter_core==5.7.1
362
+ jupyter_server==2.13.0
363
+ jupyter_server_fileid==0.9.1
364
+ jupyter_server_proxy==4.1.0
365
+ jupyter_server_terminals==0.5.1
366
+ jupyter_server_ydoc==0.8.0
367
+ jupyterlab-lsp==5.1.0
368
+ jupyterlab-widgets==3.0.9
369
+ jupyterlab==4.1.5
370
+ jupyterlab_git==0.44.0
371
+ jupyterlab_pygments==0.3.0
372
+ jupyterlab_server==2.25.2
373
+ jupytext==1.16.0
374
+ kaggle-environments==1.14.3
375
+ kaggle==1.6.8
376
+ kagglehub==0.2.2
377
+ keras-cv==0.8.2
378
+ keras-nlp==0.8.2
379
+ keras-tuner==1.4.6
380
+ keras==3.1.1
381
+ kernels-mixer==0.0.7
382
+ keyring==24.3.0
383
+ keyrings.google-artifactregistry-auth==1.1.2
384
+ kfp-pipeline-spec==0.2.2
385
+ kfp-server-api==2.0.5
386
+ kfp==2.5.0
387
+ kiwisolver==1.4.5
388
+ kmapper==2.0.1
389
+ kmodes==0.12.2
390
+ korean-lunar-calendar==0.3.1
391
+ kornia==0.7.2
392
+ kornia_rs==0.1.3
393
+ kt-legacy==1.0.5
394
+ kubernetes==26.1.0
395
+ langcodes==3.3.0
396
+ langid==1.1.6
397
+ lazy_loader==0.3
398
+ learntools==0.3.4
399
+ leven==1.0.4
400
+ libclang==16.0.6
401
+ libmambapy==1.5.0
402
+ libpysal==4.9.2
403
+ librosa==0.10.1
404
+ lightgbm==4.2.0
405
+ lightning-utilities==0.11.2
406
+ lime==0.2.0.1
407
+ line-profiler==4.1.2
408
+ linkify-it-py==2.0.3
409
+ llvmlite==0.41.1
410
+ llvmlite==0.42.0
411
+ lml==0.1.0
412
+ locket==1.0.0
413
+ loguru==0.7.2
414
+ lxml==5.2.1
415
+ lz4==4.3.3
416
+ mamba==1.5.0
417
+ mapclassify==2.6.1
418
+ markdown-it-py==3.0.0
419
+ marshmallow==3.21.1
420
+ matplotlib-inline==0.1.6
421
+ matplotlib-venn==0.11.10
422
+ matplotlib==3.7.5
423
+ matplotlib==3.8.3
424
+ mccabe==0.7.0
425
+ mdit-py-plugins==0.4.0
426
+ mdurl==0.1.2
427
+ memory-profiler==0.61.0
428
+ menuinst==2.0.1
429
+ mercantile==1.2.1
430
+ mgwr==2.2.1
431
+ missingno==0.5.2
432
+ mistune==0.8.4
433
+ mizani==0.11.1
434
+ ml-dtypes==0.2.0
435
+ mlcrate==0.2.0
436
+ mlens==0.2.3
437
+ mlxtend==0.23.1
438
+ mne==1.6.1
439
+ mnist==0.2.2
440
+ momepy==0.7.0
441
+ more-itertools==10.2.0
442
+ mpld3==0.5.10
443
+ mpmath==1.3.0
444
+ msgpack==1.0.7
445
+ multidict==6.0.4
446
+ multimethod==1.10
447
+ multipledispatch==1.0.0
448
+ multiprocess==0.70.15
449
+ munkres==1.1.4
450
+ murmurhash==1.0.10
451
+ mypy-extensions==1.0.0
452
+ namex==0.0.7
453
+ nb-conda-kernels==2.3.1
454
+ nb_conda==2.2.1
455
+ nbclassic==1.0.0
456
+ nbclient==0.5.13
457
+ nbconvert==6.4.5
458
+ nbdime==3.2.0
459
+ nbformat==5.9.2
460
+ ndindex==1.8
461
+ nest-asyncio==1.5.8
462
+ networkx==3.2.1
463
+ nibabel==5.2.1
464
+ nilearn==0.10.3
465
+ ninja==1.11.1.1
466
+ nltk==3.2.4
467
+ nose==1.3.7
468
+ notebook==6.5.4
469
+ notebook==6.5.6
470
+ notebook_executor==0.2
471
+ notebook_shim==0.2.3
472
+ numba==0.58.1
473
+ numba==0.59.1
474
+ numexpr==2.10.0
475
+ numpy==1.26.4
476
+ nvidia-ml-py==11.495.46
477
+ nvtx==0.2.10
478
+ oauth2client==4.1.3
479
+ oauthlib==3.2.2
480
+ objsize==0.6.1
481
+ odfpy==1.4.1
482
+ olefile==0.47
483
+ onnx==1.16.0
484
+ opencensus-context==0.1.3
485
+ opencensus==0.11.4
486
+ opencv-contrib-python==4.9.0.80
487
+ opencv-python-headless==4.9.0.80
488
+ opencv-python==4.9.0.80
489
+ openpyxl==3.1.2
490
+ openslide-python==1.3.1
491
+ opentelemetry-api==1.22.0
492
+ opentelemetry-exporter-otlp-proto-common==1.22.0
493
+ opentelemetry-exporter-otlp-proto-grpc==1.22.0
494
+ opentelemetry-exporter-otlp-proto-http==1.22.0
495
+ opentelemetry-exporter-otlp==1.22.0
496
+ opentelemetry-proto==1.22.0
497
+ opentelemetry-sdk==1.22.0
498
+ opentelemetry-semantic-conventions==0.43b0
499
+ opt-einsum==3.3.0
500
+ optax==0.2.2
501
+ optree==0.11.0
502
+ optuna==3.6.1
503
+ orbax-checkpoint==0.5.7
504
+ ordered-set==4.1.0
505
+ orjson==3.9.10
506
+ ortools==9.4.1874
507
+ osmnx==1.9.2
508
+ overrides==7.4.0
509
+ packaging==21.3
510
+ pandas-datareader==0.10.0
511
+ pandas-profiling==3.6.6
512
+ pandas-summary==0.2.0
513
+ pandas==2.1.4
514
+ pandas==2.2.1
515
+ pandasql==0.7.3
516
+ pandocfilters==1.5.0
517
+ panel==1.3.8
518
+ papermill==2.5.0
519
+ param==2.1.0
520
+ parso==0.8.3
521
+ partd==1.4.1
522
+ path.py==12.5.0
523
+ path==16.10.0
524
+ pathos==0.3.2
525
+ pathy==0.10.3
526
+ patsy==0.5.6
527
+ pdf2image==1.17.0
528
+ peft==0.10.0
529
+ pettingzoo==1.24.0
530
+ pexpect==4.8.0
531
+ pexpect==4.9.0
532
+ phik==0.12.4
533
+ pickleshare==0.7.5
534
+ pillow==10.3.0
535
+ pip==23.3.2
536
+ pkgutil_resolve_name==1.3.10
537
+ platformdirs==4.2.0
538
+ plotly-express==0.4.1
539
+ plotly==5.18.0
540
+ plotnine==0.13.4
541
+ pluggy==1.4.0
542
+ pointpats==2.4.0
543
+ polars==0.20.18
544
+ polyglot==16.7.4
545
+ pooch==1.8.1
546
+ pox==0.3.4
547
+ ppca==0.0.4
548
+ ppft==1.7.6.8
549
+ preprocessing==0.1.13
550
+ preshed==3.0.9
551
+ prettytable==3.9.0
552
+ progressbar2==4.4.2
553
+ prometheus-client==0.19.0
554
+ promise==2.3
555
+ prompt-toolkit==3.0.42
556
+ prompt-toolkit==3.0.43
557
+ prophet==1.1.1
558
+ proto-plus==1.23.0
559
+ protobuf==3.20.3
560
+ protobuf==4.21.12
561
+ psutil==5.9.3
562
+ psutil==5.9.7
563
+ ptyprocess==0.7.0
564
+ pudb==2024.1
565
+ pure-eval==0.2.2
566
+ py-cpuinfo==9.0.0
567
+ py-spy==0.3.14
568
+ py4j==0.10.9.7
569
+ pyLDAvis==3.4.1
570
+ pyOpenSSL==23.3.0
571
+ pyaml==23.12.0
572
+ pyarrow-hotfix==0.6
573
+ pyarrow==15.0.2
574
+ pyasn1-modules==0.3.0
575
+ pyasn1==0.5.1
576
+ pybind11==2.12.0
577
+ pyclipper==1.3.0.post5
578
+ pycodestyle==2.11.1
579
+ pycosat==0.6.6
580
+ pycparser==2.21
581
+ pycryptodome==3.20.0
582
+ pyct==0.5.0
583
+ pycuda==2024.1
584
+ pydantic==2.5.3
585
+ pydantic==2.6.4
586
+ pydantic_core==2.14.6
587
+ pydantic_core==2.16.3
588
+ pydegensac==0.1.2
589
+ pydicom==2.4.4
590
+ pydocstyle==6.3.0
591
+ pydot==1.4.2
592
+ pydub==0.25.1
593
+ pyemd==1.0.0
594
+ pyerfa==2.0.1.1
595
+ pyexcel-io==0.6.6
596
+ pyexcel-ods==0.6.0
597
+ pyflakes==3.2.0
598
+ pygltflib==1.16.2
599
+ pykalman==0.9.7
600
+ pylibraft==23.8.0
601
+ pylint==3.1.0
602
+ pymc3==3.11.4
603
+ pymongo==3.13.0
604
+ pynndescent==0.5.12
605
+ pynvml==11.4.1
606
+ pynvrtc==9.2
607
+ pyparsing==3.1.1
608
+ pyparsing==3.1.2
609
+ pypdf==4.1.0
610
+ pyproj==3.6.1
611
+ pysal==24.1
612
+ pyshp==2.3.1
613
+ pytesseract==0.3.10
614
+ pytest==8.1.1
615
+ python-bidi==0.4.2
616
+ python-dateutil==2.9.0.post0
617
+ python-dotenv==1.0.0
618
+ python-json-logger==2.0.7
619
+ python-louvain==0.16
620
+ python-lsp-jsonrpc==1.1.2
621
+ python-lsp-server==1.11.0
622
+ python-slugify==8.0.4
623
+ python-utils==3.8.2
624
+ pythreejs==2.4.2
625
+ pytoolconfig==1.3.1
626
+ pytools==2024.1.1
627
+ pytorch-ignite==0.5.0.post2
628
+ pytorch-lightning==2.2.1
629
+ pytz==2023.3.post1
630
+ pytz==2024.1
631
+ pyu2f==0.1.5
632
+ pyviz_comms==3.0.2
633
+ pyzmq==24.0.1
634
+ pyzmq==25.1.2
635
+ qgrid==1.3.1
636
+ qtconsole==5.5.1
637
+ quantecon==0.7.2
638
+ qudida==0.0.4
639
+ raft-dask==23.8.0
640
+ rasterio==1.3.9
641
+ rasterstats==0.19.0
642
+ ray-cpp==2.9.0
643
+ ray==2.9.0
644
+ referencing==0.32.1
645
+ regex==2023.12.25
646
+ requests-oauthlib==1.3.1
647
+ requests-toolbelt==0.10.1
648
+ requests==2.31.0
649
+ retrying==1.3.3
650
+ retrying==1.3.4
651
+ rfc3339-validator==0.1.4
652
+ rfc3986-validator==0.1.1
653
+ rgf-python==3.12.0
654
+ rich-click==1.7.4
655
+ rich==13.7.0
656
+ rich==13.7.1
657
+ rmm==23.8.0
658
+ rope==1.13.0
659
+ rpds-py==0.16.2
660
+ rsa==4.9
661
+ ruamel-yaml-conda==0.15.100
662
+ ruamel.yaml.clib==0.2.7
663
+ ruamel.yaml==0.17.40
664
+ s2sphere==0.2.5
665
+ s3fs==2024.2.0
666
+ s3transfer==0.6.2
667
+ safetensors==0.4.2
668
+ scattertext==0.1.19
669
+ scikit-image==0.22.0
670
+ scikit-learn-intelex==2024.2.0
671
+ scikit-learn==1.2.2
672
+ scikit-multilearn==0.2.0
673
+ scikit-optimize==0.10.1
674
+ scikit-plot==0.3.7
675
+ scikit-surprise==1.1.3
676
+ scipy==1.11.4
677
+ scipy==1.12.0
678
+ seaborn==0.12.2
679
+ segment_anything==1.0
680
+ segregation==2.5
681
+ semver==3.0.2
682
+ sentencepiece==0.2.0
683
+ sentry-sdk==1.44.1
684
+ setproctitle==1.3.3
685
+ setuptools-git==1.2
686
+ setuptools-scm==8.0.4
687
+ setuptools==69.0.3
688
+ shap==0.44.1
689
+ shapely==2.0.3
690
+ shellingham==1.5.4
691
+ shtab==1.7.1
692
+ simpervisor==1.0.0
693
+ simplejson==3.19.2
694
+ six==1.16.0
695
+ sklearn-pandas==2.2.0
696
+ slicer==0.0.7
697
+ smart-open==6.4.0
698
+ smmap==5.0.1
699
+ sniffio==1.3.0
700
+ snowballstemmer==2.2.0
701
+ snuggs==1.4.7
702
+ sortedcontainers==2.4.0
703
+ soundfile==0.12.1
704
+ soupsieve==2.5
705
+ soxr==0.3.7
706
+ spacy-legacy==3.0.12
707
+ spacy-loggers==1.0.5
708
+ spacy==3.7.2
709
+ spaghetti==1.7.5.post1
710
+ spectral==0.23.1
711
+ spglm==1.1.0
712
+ sphinx-rtd-theme==0.2.4
713
+ spint==1.0.7
714
+ splot==1.1.5.post1
715
+ spopt==0.6.0
716
+ spreg==1.4.2
717
+ spvcm==0.3.0
718
+ sqlparse==0.4.4
719
+ squarify==0.4.3
720
+ srsly==2.4.8
721
+ stable-baselines3==2.1.0
722
+ stack-data==0.6.2
723
+ stack-data==0.6.3
724
+ stanio==0.5.0
725
+ starlette==0.32.0.post1
726
+ statsmodels==0.14.1
727
+ stemming==1.0.1
728
+ stop-words==2018.7.23
729
+ stopit==1.1.2
730
+ stumpy==1.12.0
731
+ sympy==1.12
732
+ tables==3.9.2
733
+ tabulate==0.9.0
734
+ tangled-up-in-unicode==0.2.0
735
+ tbb==2021.12.0
736
+ tblib==3.0.0
737
+ tenacity==8.2.3
738
+ tensorboard-data-server==0.7.2
739
+ tensorboard-plugin-profile==2.15.0
740
+ tensorboard==2.15.1
741
+ tensorboardX==2.6.2.2
742
+ tensorflow-cloud==0.1.16
743
+ tensorflow-datasets==4.9.4
744
+ tensorflow-decision-forests==1.8.1
745
+ tensorflow-estimator==2.15.0
746
+ tensorflow-hub==0.16.1
747
+ tensorflow-io-gcs-filesystem==0.35.0
748
+ tensorflow-io==0.35.0
749
+ tensorflow-metadata==0.14.0
750
+ tensorflow-probability==0.23.0
751
+ tensorflow-serving-api==2.14.1
752
+ tensorflow-text==2.15.0
753
+ tensorflow-transform==0.14.0
754
+ tensorflow==2.15.0
755
+ tensorstore==0.1.56
756
+ termcolor==2.4.0
757
+ terminado==0.18.0
758
+ testpath==0.6.0
759
+ text-unidecode==1.3
760
+ textblob==0.18.0.post0
761
+ texttable==1.7.0
762
+ tf_keras==2.15.1
763
+ tfp-nightly==0.24.0.dev0
764
+ thinc==8.2.2
765
+ threadpoolctl==3.2.0
766
+ tifffile==2023.12.9
767
+ timm==0.9.16
768
+ tinycss2==1.2.1
769
+ tobler==0.11.2
770
+ tokenizers==0.15.2
771
+ toml==0.10.2
772
+ tomli==2.0.1
773
+ tomlkit==0.12.4
774
+ toolz==0.12.1
775
+ torch==2.1.2
776
+ torchaudio==2.1.2
777
+ torchdata==0.7.1
778
+ torchinfo==1.8.0
779
+ torchmetrics==1.3.2
780
+ torchtext==0.16.2
781
+ torchvision==0.16.2
782
+ tornado==6.3.3
783
+ tqdm==4.66.1
784
+ traceml==1.0.8
785
+ traitlets==5.9.0
786
+ traittypes==0.2.1
787
+ transformers==4.39.3
788
+ treelite-runtime==3.2.0
789
+ treelite==3.2.0
790
+ trl==0.8.1
791
+ truststore==0.8.0
792
+ trx-python==0.2.9
793
+ tsfresh==0.20.2
794
+ typeguard==4.1.5
795
+ typer==0.9.0
796
+ typer==0.9.4
797
+ types-python-dateutil==2.8.19.20240106
798
+ typing-inspect==0.9.0
799
+ typing-utils==0.1.0
800
+ typing_extensions==4.9.0
801
+ tyro==0.8.3
802
+ tzdata==2023.4
803
+ uc-micro-py==1.0.3
804
+ ucx-py==0.33.0
805
+ ujson==5.9.0
806
+ umap-learn==0.5.5
807
+ unicodedata2==15.1.0
808
+ update-checker==0.18.0
809
+ uri-template==1.3.0
810
+ uritemplate==3.0.1
811
+ urllib3==1.26.18
812
+ urllib3==2.1.0
813
+ urwid==2.6.10
814
+ urwid_readline==0.14
815
+ uvicorn==0.25.0
816
+ uvloop==0.19.0
817
+ vaex-astro==0.9.3
818
+ vaex-core==4.17.1
819
+ vaex-hdf5==0.14.1
820
+ vaex-jupyter==0.8.2
821
+ vaex-ml==0.18.3
822
+ vaex-server==0.9.0
823
+ vaex-viz==0.5.4
824
+ vaex==4.17.0
825
+ vec_noise==1.1.4
826
+ vecstack==0.4.0
827
+ virtualenv==20.21.0
828
+ visions==0.7.5
829
+ vowpalwabbit==9.9.0
830
+ vtk==9.3.0
831
+ wandb==0.16.5
832
+ wasabi==1.1.2
833
+ watchfiles==0.21.0
834
+ wavio==0.0.8
835
+ wcwidth==0.2.13
836
+ weasel==0.3.4
837
+ webcolors==1.13
838
+ webencodings==0.5.1
839
+ websocket-client==1.7.0
840
+ websockets==12.0
841
+ wfdb==4.1.2
842
+ whatthepatch==1.0.5
843
+ wheel==0.42.0
844
+ widgetsnbextension==3.6.6
845
+ witwidget==1.8.1
846
+ woodwork==0.29.0
847
+ wordcloud==1.9.3
848
+ wordsegment==1.3.1
849
+ wrapt==1.14.1
850
+ xarray-einstats==0.7.0
851
+ xarray==2024.3.0
852
+ xgboost==2.0.3
853
+ xvfbwrapper==0.2.9
854
+ xxhash==3.4.1
855
+ xyzservices==2023.10.1
856
+ y-py==0.6.2
857
+ yapf==0.40.2
858
+ yarl==1.9.3
859
+ yarl==1.9.4
860
+ ydata-profiling==4.6.4
861
+ yellowbrick==1.5
862
+ ypy-websocket==0.8.4
863
+ zict==3.0.0
864
+ zipp==3.17.0
865
+ zstandard==0.22.0
wandb/run-20240410_215101-ta1pahei/files/wandb-metadata.json ADDED
@@ -0,0 +1,66 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.15.133+-x86_64-with-glibc2.31",
3
+ "python": "3.10.13",
4
+ "heartbeatAt": "2024-04-10T21:51:01.802955",
5
+ "startedAt": "2024-04-10T21:51:01.066772",
6
+ "docker": null,
7
+ "cuda": null,
8
+ "args": [],
9
+ "state": "running",
10
+ "program": "kaggle.ipynb",
11
+ "codePathLocal": null,
12
+ "root": "/kaggle/working",
13
+ "host": "4bf7c58fba03",
14
+ "username": "root",
15
+ "executable": "/opt/conda/bin/python3.10",
16
+ "cpu_count": 2,
17
+ "cpu_count_logical": 4,
18
+ "cpu_freq": {
19
+ "current": 2000.192,
20
+ "min": 0.0,
21
+ "max": 0.0
22
+ },
23
+ "cpu_freq_per_core": [
24
+ {
25
+ "current": 2000.192,
26
+ "min": 0.0,
27
+ "max": 0.0
28
+ },
29
+ {
30
+ "current": 2000.192,
31
+ "min": 0.0,
32
+ "max": 0.0
33
+ },
34
+ {
35
+ "current": 2000.192,
36
+ "min": 0.0,
37
+ "max": 0.0
38
+ },
39
+ {
40
+ "current": 2000.192,
41
+ "min": 0.0,
42
+ "max": 0.0
43
+ }
44
+ ],
45
+ "disk": {
46
+ "/": {
47
+ "total": 8062.387607574463,
48
+ "used": 5578.723789215088
49
+ }
50
+ },
51
+ "gpu": "Tesla T4",
52
+ "gpu_count": 2,
53
+ "gpu_devices": [
54
+ {
55
+ "name": "Tesla T4",
56
+ "memory_total": 16106127360
57
+ },
58
+ {
59
+ "name": "Tesla T4",
60
+ "memory_total": 16106127360
61
+ }
62
+ ],
63
+ "memory": {
64
+ "total": 31.357559204101562
65
+ }
66
+ }
wandb/run-20240410_215101-ta1pahei/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"train/loss": 10.1111, "train/grad_norm": 140.29942321777344, "train/learning_rate": 0.0, "train/epoch": 0.0, "train/global_step": 20, "_timestamp": 1712786517.3377423, "_runtime": 656.2637143135071, "_step": 1, "train_runtime": 50.9578, "train_samples_per_second": 1.57, "train_steps_per_second": 0.392, "total_flos": 122901052833792.0, "train_loss": 10.111084747314454}
wandb/run-20240410_215101-ta1pahei/logs/debug-internal.log ADDED
@@ -0,0 +1,406 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-04-10 21:51:01,073 INFO StreamThr :274 [internal.py:wandb_internal():86] W&B internal server running at pid: 274, started at: 2024-04-10 21:51:01.072932
2
+ 2024-04-10 21:51:01,075 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status
3
+ 2024-04-10 21:51:01,447 INFO WriterThread:274 [datastore.py:open_for_write():87] open: /kaggle/working/wandb/run-20240410_215101-ta1pahei/run-ta1pahei.wandb
4
+ 2024-04-10 21:51:01,448 DEBUG SenderThread:274 [sender.py:send():379] send: header
5
+ 2024-04-10 21:51:01,459 DEBUG SenderThread:274 [sender.py:send():379] send: run
6
+ 2024-04-10 21:51:01,670 INFO SenderThread:274 [dir_watcher.py:__init__():211] watching files in: /kaggle/working/wandb/run-20240410_215101-ta1pahei/files
7
+ 2024-04-10 21:51:01,670 INFO SenderThread:274 [sender.py:_start_run_threads():1124] run started: ta1pahei with start time 1712785861.074028
8
+ 2024-04-10 21:51:01,678 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: check_version
9
+ 2024-04-10 21:51:01,678 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: check_version
10
+ 2024-04-10 21:51:01,777 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: run_start
11
+ 2024-04-10 21:51:01,788 DEBUG HandlerThread:274 [system_info.py:__init__():26] System info init
12
+ 2024-04-10 21:51:01,789 DEBUG HandlerThread:274 [system_info.py:__init__():41] System info init done
13
+ 2024-04-10 21:51:01,789 INFO HandlerThread:274 [system_monitor.py:start():194] Starting system monitor
14
+ 2024-04-10 21:51:01,789 INFO SystemMonitor:274 [system_monitor.py:_start():158] Starting system asset monitoring threads
15
+ 2024-04-10 21:51:01,789 INFO HandlerThread:274 [system_monitor.py:probe():214] Collecting system info
16
+ 2024-04-10 21:51:01,789 INFO SystemMonitor:274 [interfaces.py:start():190] Started cpu monitoring
17
+ 2024-04-10 21:51:01,790 INFO SystemMonitor:274 [interfaces.py:start():190] Started disk monitoring
18
+ 2024-04-10 21:51:01,791 INFO SystemMonitor:274 [interfaces.py:start():190] Started gpu monitoring
19
+ 2024-04-10 21:51:01,792 INFO SystemMonitor:274 [interfaces.py:start():190] Started memory monitoring
20
+ 2024-04-10 21:51:01,792 INFO SystemMonitor:274 [interfaces.py:start():190] Started network monitoring
21
+ 2024-04-10 21:51:01,802 DEBUG HandlerThread:274 [system_info.py:probe():150] Probing system
22
+ 2024-04-10 21:51:01,805 DEBUG HandlerThread:274 [gitlib.py:_init_repo():56] git repository is invalid
23
+ 2024-04-10 21:51:01,805 DEBUG HandlerThread:274 [system_info.py:probe():198] Probing system done
24
+ 2024-04-10 21:51:01,805 DEBUG HandlerThread:274 [system_monitor.py:probe():223] {'os': 'Linux-5.15.133+-x86_64-with-glibc2.31', 'python': '3.10.13', 'heartbeatAt': '2024-04-10T21:51:01.802955', 'startedAt': '2024-04-10T21:51:01.066772', 'docker': None, 'cuda': None, 'args': (), 'state': 'running', 'program': 'kaggle.ipynb', 'codePathLocal': None, 'root': '/kaggle/working', 'host': '4bf7c58fba03', 'username': 'root', 'executable': '/opt/conda/bin/python3.10', 'cpu_count': 2, 'cpu_count_logical': 4, 'cpu_freq': {'current': 2000.192, 'min': 0.0, 'max': 0.0}, 'cpu_freq_per_core': [{'current': 2000.192, 'min': 0.0, 'max': 0.0}, {'current': 2000.192, 'min': 0.0, 'max': 0.0}, {'current': 2000.192, 'min': 0.0, 'max': 0.0}, {'current': 2000.192, 'min': 0.0, 'max': 0.0}], 'disk': {'/': {'total': 8062.387607574463, 'used': 5578.723789215088}}, 'gpu': 'Tesla T4', 'gpu_count': 2, 'gpu_devices': [{'name': 'Tesla T4', 'memory_total': 16106127360}, {'name': 'Tesla T4', 'memory_total': 16106127360}], 'memory': {'total': 31.357559204101562}}
25
+ 2024-04-10 21:51:01,805 INFO HandlerThread:274 [system_monitor.py:probe():224] Finished collecting system info
26
+ 2024-04-10 21:51:01,805 INFO HandlerThread:274 [system_monitor.py:probe():227] Publishing system info
27
+ 2024-04-10 21:51:01,805 DEBUG HandlerThread:274 [system_info.py:_save_conda():207] Saving list of conda packages installed into the current environment
28
+ 2024-04-10 21:51:02,672 INFO Thread-12 :274 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240410_215101-ta1pahei/files/conda-environment.yaml
29
+ 2024-04-10 21:51:16,821 ERROR HandlerThread:274 [system_info.py:_save_conda():221] Error saving conda packages: Command '['conda', 'env', 'export']' timed out after 15 seconds
30
+ Traceback (most recent call last):
31
+ File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/internal/system/system_info.py", line 214, in _save_conda
32
+ subprocess.call(
33
+ File "/opt/conda/lib/python3.10/subprocess.py", line 347, in call
34
+ return p.wait(timeout=timeout)
35
+ File "/opt/conda/lib/python3.10/subprocess.py", line 1209, in wait
36
+ return self._wait(timeout=timeout)
37
+ File "/opt/conda/lib/python3.10/subprocess.py", line 1951, in _wait
38
+ raise TimeoutExpired(self.args, timeout)
39
+ subprocess.TimeoutExpired: Command '['conda', 'env', 'export']' timed out after 15 seconds
40
+ 2024-04-10 21:51:16,824 DEBUG HandlerThread:274 [system_info.py:_save_conda():222] Saving conda packages done
41
+ 2024-04-10 21:51:16,825 INFO HandlerThread:274 [system_monitor.py:probe():229] Finished publishing system info
42
+ 2024-04-10 21:51:16,831 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
43
+ 2024-04-10 21:51:16,832 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: keepalive
44
+ 2024-04-10 21:51:16,832 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
45
+ 2024-04-10 21:51:16,832 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: keepalive
46
+ 2024-04-10 21:51:16,832 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
47
+ 2024-04-10 21:51:16,832 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: keepalive
48
+ 2024-04-10 21:51:16,833 DEBUG SenderThread:274 [sender.py:send():379] send: files
49
+ 2024-04-10 21:51:16,833 INFO SenderThread:274 [sender.py:_save_file():1390] saving file wandb-metadata.json with policy now
50
+ 2024-04-10 21:51:17,270 INFO wandb-upload_0:274 [upload_job.py:push():131] Uploaded file /tmp/tmphet9wukmwandb/tfhoobcd-wandb-metadata.json
51
+ 2024-04-10 21:51:17,675 INFO Thread-12 :274 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240410_215101-ta1pahei/files/wandb-metadata.json
52
+ 2024-04-10 21:51:17,776 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: python_packages
53
+ 2024-04-10 21:51:17,776 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: python_packages
54
+ 2024-04-10 21:51:17,779 DEBUG SenderThread:274 [sender.py:send():379] send: telemetry
55
+ 2024-04-10 21:51:17,789 DEBUG SenderThread:274 [sender.py:send():379] send: config
56
+ 2024-04-10 21:51:17,791 DEBUG SenderThread:274 [sender.py:send():379] send: metric
57
+ 2024-04-10 21:51:17,792 DEBUG SenderThread:274 [sender.py:send():379] send: telemetry
58
+ 2024-04-10 21:51:17,792 DEBUG SenderThread:274 [sender.py:send():379] send: metric
59
+ 2024-04-10 21:51:17,792 WARNING SenderThread:274 [sender.py:send_metric():1341] Seen metric with glob (shouldn't happen)
60
+ 2024-04-10 21:51:17,792 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
61
+ 2024-04-10 21:51:17,793 DEBUG SenderThread:274 [sender.py:send():379] send: telemetry
62
+ 2024-04-10 21:51:17,793 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
63
+ 2024-04-10 21:51:17,793 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
64
+ 2024-04-10 21:51:18,676 INFO Thread-12 :274 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240410_215101-ta1pahei/files/requirements.txt
65
+ 2024-04-10 21:51:18,676 INFO Thread-12 :274 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240410_215101-ta1pahei/files/output.log
66
+ 2024-04-10 21:51:20,677 INFO Thread-12 :274 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240410_215101-ta1pahei/files/output.log
67
+ 2024-04-10 21:51:22,163 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
68
+ 2024-04-10 21:51:22,849 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: pause
69
+ 2024-04-10 21:51:22,849 INFO HandlerThread:274 [handler.py:handle_request_pause():708] stopping system metrics thread
70
+ 2024-04-10 21:51:22,849 INFO HandlerThread:274 [system_monitor.py:finish():203] Stopping system monitor
71
+ 2024-04-10 21:51:22,849 DEBUG SystemMonitor:274 [system_monitor.py:_start():172] Starting system metrics aggregation loop
72
+ 2024-04-10 21:51:22,849 DEBUG SystemMonitor:274 [system_monitor.py:_start():179] Finished system metrics aggregation loop
73
+ 2024-04-10 21:51:22,849 DEBUG SystemMonitor:274 [system_monitor.py:_start():183] Publishing last batch of metrics
74
+ 2024-04-10 21:51:22,850 INFO HandlerThread:274 [interfaces.py:finish():202] Joined cpu monitor
75
+ 2024-04-10 21:51:22,850 INFO HandlerThread:274 [interfaces.py:finish():202] Joined disk monitor
76
+ 2024-04-10 21:51:22,861 INFO HandlerThread:274 [interfaces.py:finish():202] Joined gpu monitor
77
+ 2024-04-10 21:51:22,861 INFO HandlerThread:274 [interfaces.py:finish():202] Joined memory monitor
78
+ 2024-04-10 21:51:22,861 INFO HandlerThread:274 [interfaces.py:finish():202] Joined network monitor
79
+ 2024-04-10 21:51:22,861 DEBUG SenderThread:274 [sender.py:send():379] send: stats
80
+ 2024-04-10 21:51:27,862 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
81
+ 2024-04-10 21:51:32,778 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
82
+ 2024-04-10 21:51:32,779 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
83
+ 2024-04-10 21:51:32,779 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
84
+ 2024-04-10 21:51:33,855 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
85
+ 2024-04-10 21:51:34,682 INFO Thread-12 :274 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240410_215101-ta1pahei/files/config.yaml
86
+ 2024-04-10 21:51:38,978 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
87
+ 2024-04-10 21:51:43,979 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
88
+ 2024-04-10 21:51:47,777 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
89
+ 2024-04-10 21:51:47,778 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
90
+ 2024-04-10 21:51:47,817 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
91
+ 2024-04-10 21:51:49,947 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
92
+ 2024-04-10 21:51:54,947 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
93
+ 2024-04-10 21:51:59,948 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
94
+ 2024-04-10 21:52:02,777 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
95
+ 2024-04-10 21:52:02,778 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
96
+ 2024-04-10 21:52:02,818 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
97
+ 2024-04-10 21:52:05,863 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
98
+ 2024-04-10 21:52:10,864 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
99
+ 2024-04-10 21:52:15,865 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
100
+ 2024-04-10 21:52:17,777 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
101
+ 2024-04-10 21:52:17,778 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
102
+ 2024-04-10 21:52:17,818 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
103
+ 2024-04-10 21:52:20,900 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
104
+ 2024-04-10 21:52:25,901 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
105
+ 2024-04-10 21:52:30,902 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
106
+ 2024-04-10 21:52:32,777 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
107
+ 2024-04-10 21:52:32,778 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
108
+ 2024-04-10 21:52:32,818 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
109
+ 2024-04-10 21:52:36,843 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
110
+ 2024-04-10 21:52:40,759 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: resume
111
+ 2024-04-10 21:52:40,759 INFO HandlerThread:274 [handler.py:handle_request_resume():699] starting system metrics thread
112
+ 2024-04-10 21:52:40,759 INFO HandlerThread:274 [system_monitor.py:start():194] Starting system monitor
113
+ 2024-04-10 21:52:40,759 INFO SystemMonitor:274 [system_monitor.py:_start():158] Starting system asset monitoring threads
114
+ 2024-04-10 21:52:40,760 INFO SystemMonitor:274 [interfaces.py:start():190] Started cpu monitoring
115
+ 2024-04-10 21:52:40,761 INFO SystemMonitor:274 [interfaces.py:start():190] Started disk monitoring
116
+ 2024-04-10 21:52:40,762 INFO SystemMonitor:274 [interfaces.py:start():190] Started gpu monitoring
117
+ 2024-04-10 21:52:40,763 INFO SystemMonitor:274 [interfaces.py:start():190] Started memory monitoring
118
+ 2024-04-10 21:52:40,764 INFO SystemMonitor:274 [interfaces.py:start():190] Started network monitoring
119
+ 2024-04-10 21:52:41,844 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
120
+ 2024-04-10 21:52:46,708 INFO Thread-12 :274 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240410_215101-ta1pahei/files/output.log
121
+ 2024-04-10 21:52:47,665 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
122
+ 2024-04-10 21:52:47,789 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
123
+ 2024-04-10 21:52:47,789 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
124
+ 2024-04-10 21:52:47,829 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
125
+ 2024-04-10 21:52:48,709 INFO Thread-12 :274 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240410_215101-ta1pahei/files/output.log
126
+ 2024-04-10 21:52:52,915 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
127
+ 2024-04-10 21:52:57,916 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
128
+ 2024-04-10 21:53:02,783 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
129
+ 2024-04-10 21:53:02,784 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
130
+ 2024-04-10 21:53:02,784 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
131
+ 2024-04-10 21:53:04,789 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
132
+ 2024-04-10 21:53:09,789 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
133
+ 2024-04-10 21:53:14,790 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
134
+ 2024-04-10 21:53:17,800 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
135
+ 2024-04-10 21:53:17,802 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
136
+ 2024-04-10 21:53:17,802 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
137
+ 2024-04-10 21:53:19,871 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
138
+ 2024-04-10 21:53:24,874 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
139
+ 2024-04-10 21:53:29,875 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
140
+ 2024-04-10 21:53:32,801 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
141
+ 2024-04-10 21:53:32,801 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
142
+ 2024-04-10 21:53:32,802 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
143
+ 2024-04-10 21:53:34,952 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
144
+ 2024-04-10 21:53:39,953 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
145
+ 2024-04-10 21:53:40,767 DEBUG SystemMonitor:274 [system_monitor.py:_start():172] Starting system metrics aggregation loop
146
+ 2024-04-10 21:53:40,768 DEBUG SenderThread:274 [sender.py:send():379] send: stats
147
+ 2024-04-10 21:53:45,770 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
148
+ 2024-04-10 21:53:48,058 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
149
+ 2024-04-10 21:53:48,058 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
150
+ 2024-04-10 21:53:48,106 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
151
+ 2024-04-10 21:53:51,112 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
152
+ 2024-04-10 21:53:56,113 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
153
+ 2024-04-10 21:54:01,113 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
154
+ 2024-04-10 21:54:03,034 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
155
+ 2024-04-10 21:54:03,035 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
156
+ 2024-04-10 21:54:03,075 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
157
+ 2024-04-10 21:54:06,135 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
158
+ 2024-04-10 21:54:10,769 DEBUG SenderThread:274 [sender.py:send():379] send: stats
159
+ 2024-04-10 21:54:11,770 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
160
+ 2024-04-10 21:54:16,771 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
161
+ 2024-04-10 21:54:18,045 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
162
+ 2024-04-10 21:54:18,045 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
163
+ 2024-04-10 21:54:18,087 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
164
+ 2024-04-10 21:54:22,200 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
165
+ 2024-04-10 21:54:27,201 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
166
+ 2024-04-10 21:54:32,202 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
167
+ 2024-04-10 21:54:33,045 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
168
+ 2024-04-10 21:54:33,045 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
169
+ 2024-04-10 21:54:33,085 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
170
+ 2024-04-10 21:54:37,214 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
171
+ 2024-04-10 21:54:40,770 DEBUG SenderThread:274 [sender.py:send():379] send: stats
172
+ 2024-04-10 21:54:42,771 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
173
+ 2024-04-10 21:54:47,772 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
174
+ 2024-04-10 21:54:48,045 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
175
+ 2024-04-10 21:54:48,046 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
176
+ 2024-04-10 21:54:48,087 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
177
+ 2024-04-10 21:54:53,167 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
178
+ 2024-04-10 21:54:58,167 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
179
+ 2024-04-10 21:55:03,168 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
180
+ 2024-04-10 21:55:04,522 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
181
+ 2024-04-10 21:55:04,523 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
182
+ 2024-04-10 21:55:04,524 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
183
+ 2024-04-10 21:55:08,693 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
184
+ 2024-04-10 21:55:10,771 DEBUG SenderThread:274 [sender.py:send():379] send: stats
185
+ 2024-04-10 21:55:13,772 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
186
+ 2024-04-10 21:55:18,288 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
187
+ 2024-04-10 21:55:18,288 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
188
+ 2024-04-10 21:55:18,399 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
189
+ 2024-04-10 21:55:19,387 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
190
+ 2024-04-10 21:55:24,388 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
191
+ 2024-04-10 21:55:29,389 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
192
+ 2024-04-10 21:55:33,895 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
193
+ 2024-04-10 21:55:33,895 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
194
+ 2024-04-10 21:55:33,896 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
195
+ 2024-04-10 21:55:34,984 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
196
+ 2024-04-10 21:55:39,985 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
197
+ 2024-04-10 21:55:40,772 DEBUG SenderThread:274 [sender.py:send():379] send: stats
198
+ 2024-04-10 21:55:45,774 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
199
+ 2024-04-10 21:55:48,531 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
200
+ 2024-04-10 21:55:48,581 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
201
+ 2024-04-10 21:55:48,582 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
202
+ 2024-04-10 21:55:50,895 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
203
+ 2024-04-10 21:55:55,897 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
204
+ 2024-04-10 21:56:00,897 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
205
+ 2024-04-10 21:56:03,531 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
206
+ 2024-04-10 21:56:03,581 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
207
+ 2024-04-10 21:56:03,582 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
208
+ 2024-04-10 21:56:06,895 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
209
+ 2024-04-10 21:56:10,774 DEBUG SenderThread:274 [sender.py:send():379] send: stats
210
+ 2024-04-10 21:56:12,775 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
211
+ 2024-04-10 21:56:17,775 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
212
+ 2024-04-10 21:56:18,531 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
213
+ 2024-04-10 21:56:18,581 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
214
+ 2024-04-10 21:56:18,582 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
215
+ 2024-04-10 21:56:22,895 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
216
+ 2024-04-10 21:56:27,896 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
217
+ 2024-04-10 21:56:32,898 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
218
+ 2024-04-10 21:56:33,722 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
219
+ 2024-04-10 21:56:33,729 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
220
+ 2024-04-10 21:56:33,729 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
221
+ 2024-04-10 21:56:38,159 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
222
+ 2024-04-10 21:56:40,775 DEBUG SenderThread:274 [sender.py:send():379] send: stats
223
+ 2024-04-10 21:56:43,776 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
224
+ 2024-04-10 21:56:48,546 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
225
+ 2024-04-10 21:56:48,585 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
226
+ 2024-04-10 21:56:48,586 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
227
+ 2024-04-10 21:56:49,717 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
228
+ 2024-04-10 21:56:55,577 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
229
+ 2024-04-10 21:57:00,578 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
230
+ 2024-04-10 21:57:03,546 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
231
+ 2024-04-10 21:57:03,586 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
232
+ 2024-04-10 21:57:03,586 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
233
+ 2024-04-10 21:57:05,636 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
234
+ 2024-04-10 21:57:10,637 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
235
+ 2024-04-10 21:57:10,776 DEBUG SenderThread:274 [sender.py:send():379] send: stats
236
+ 2024-04-10 21:57:15,777 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
237
+ 2024-04-10 21:57:18,547 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
238
+ 2024-04-10 21:57:18,586 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
239
+ 2024-04-10 21:57:18,586 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
240
+ 2024-04-10 21:57:21,585 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
241
+ 2024-04-10 21:57:26,586 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
242
+ 2024-04-10 21:57:31,587 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
243
+ 2024-04-10 21:57:33,567 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
244
+ 2024-04-10 21:57:33,588 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
245
+ 2024-04-10 21:57:33,589 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
246
+ 2024-04-10 21:57:36,605 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
247
+ 2024-04-10 21:57:40,777 DEBUG SenderThread:274 [sender.py:send():379] send: stats
248
+ 2024-04-10 21:57:41,778 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
249
+ 2024-04-10 21:57:46,779 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
250
+ 2024-04-10 21:57:48,567 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
251
+ 2024-04-10 21:57:48,588 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
252
+ 2024-04-10 21:57:48,589 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
253
+ 2024-04-10 21:57:52,586 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
254
+ 2024-04-10 21:57:57,587 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
255
+ 2024-04-10 21:58:02,588 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
256
+ 2024-04-10 21:58:03,567 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
257
+ 2024-04-10 21:58:03,588 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
258
+ 2024-04-10 21:58:03,589 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
259
+ 2024-04-10 21:58:08,587 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
260
+ 2024-04-10 21:58:10,778 DEBUG SenderThread:274 [sender.py:send():379] send: stats
261
+ 2024-04-10 21:58:13,779 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
262
+ 2024-04-10 21:58:18,567 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
263
+ 2024-04-10 21:58:18,588 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
264
+ 2024-04-10 21:58:18,589 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
265
+ 2024-04-10 21:58:19,703 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
266
+ 2024-04-10 21:58:25,586 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
267
+ 2024-04-10 21:58:30,587 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
268
+ 2024-04-10 21:58:33,567 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
269
+ 2024-04-10 21:58:33,589 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
270
+ 2024-04-10 21:58:33,589 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
271
+ 2024-04-10 21:58:36,586 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
272
+ 2024-04-10 21:58:40,779 DEBUG SenderThread:274 [sender.py:send():379] send: stats
273
+ 2024-04-10 21:58:41,780 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
274
+ 2024-04-10 21:58:46,780 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
275
+ 2024-04-10 21:58:48,567 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
276
+ 2024-04-10 21:58:48,602 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
277
+ 2024-04-10 21:58:48,602 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
278
+ 2024-04-10 21:58:52,596 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
279
+ 2024-04-10 21:58:57,597 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
280
+ 2024-04-10 21:59:02,598 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
281
+ 2024-04-10 21:59:03,568 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
282
+ 2024-04-10 21:59:03,589 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
283
+ 2024-04-10 21:59:03,589 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
284
+ 2024-04-10 21:59:08,596 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
285
+ 2024-04-10 21:59:10,780 DEBUG SenderThread:274 [sender.py:send():379] send: stats
286
+ 2024-04-10 21:59:13,781 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
287
+ 2024-04-10 21:59:18,568 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
288
+ 2024-04-10 21:59:18,589 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
289
+ 2024-04-10 21:59:18,590 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
290
+ 2024-04-10 21:59:19,663 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
291
+ 2024-04-10 21:59:25,597 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
292
+ 2024-04-10 21:59:30,598 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
293
+ 2024-04-10 21:59:33,568 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
294
+ 2024-04-10 21:59:33,589 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
295
+ 2024-04-10 21:59:33,590 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
296
+ 2024-04-10 21:59:35,711 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
297
+ 2024-04-10 21:59:40,781 DEBUG SenderThread:274 [sender.py:send():379] send: stats
298
+ 2024-04-10 21:59:40,783 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
299
+ 2024-04-10 21:59:45,783 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
300
+ 2024-04-10 21:59:48,568 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
301
+ 2024-04-10 21:59:48,589 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
302
+ 2024-04-10 21:59:48,590 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
303
+ 2024-04-10 21:59:51,602 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
304
+ 2024-04-10 21:59:56,604 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
305
+ 2024-04-10 22:00:01,605 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
306
+ 2024-04-10 22:00:03,568 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
307
+ 2024-04-10 22:00:03,589 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
308
+ 2024-04-10 22:00:03,590 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
309
+ 2024-04-10 22:00:06,608 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
310
+ 2024-04-10 22:00:10,782 DEBUG SenderThread:274 [sender.py:send():379] send: stats
311
+ 2024-04-10 22:00:11,783 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
312
+ 2024-04-10 22:00:16,784 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
313
+ 2024-04-10 22:00:18,568 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
314
+ 2024-04-10 22:00:18,590 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
315
+ 2024-04-10 22:00:18,590 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
316
+ 2024-04-10 22:00:22,603 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
317
+ 2024-04-10 22:00:27,604 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
318
+ 2024-04-10 22:00:32,605 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
319
+ 2024-04-10 22:00:33,584 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
320
+ 2024-04-10 22:00:33,614 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
321
+ 2024-04-10 22:00:33,615 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
322
+ 2024-04-10 22:00:38,603 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
323
+ 2024-04-10 22:00:40,783 DEBUG SenderThread:274 [sender.py:send():379] send: stats
324
+ 2024-04-10 22:00:44,504 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
325
+ 2024-04-10 22:00:44,558 DEBUG SenderThread:274 [sender.py:send():379] send: config
326
+ 2024-04-10 22:00:44,560 DEBUG SenderThread:274 [sender.py:send():379] send: metric
327
+ 2024-04-10 22:00:44,560 DEBUG SenderThread:274 [sender.py:send():379] send: metric
328
+ 2024-04-10 22:00:44,560 WARNING SenderThread:274 [sender.py:send_metric():1341] Seen metric with glob (shouldn't happen)
329
+ 2024-04-10 22:00:44,919 INFO Thread-12 :274 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240410_215101-ta1pahei/files/output.log
330
+ 2024-04-10 22:00:46,920 INFO Thread-12 :274 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240410_215101-ta1pahei/files/output.log
331
+ 2024-04-10 22:00:48,585 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
332
+ 2024-04-10 22:00:48,615 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
333
+ 2024-04-10 22:00:48,615 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
334
+ 2024-04-10 22:00:48,921 INFO Thread-12 :274 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240410_215101-ta1pahei/files/output.log
335
+ 2024-04-10 22:00:49,760 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
336
+ 2024-04-10 22:00:55,604 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
337
+ 2024-04-10 22:00:56,923 INFO Thread-12 :274 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240410_215101-ta1pahei/files/output.log
338
+ 2024-04-10 22:00:58,924 INFO Thread-12 :274 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240410_215101-ta1pahei/files/output.log
339
+ 2024-04-10 22:01:00,664 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
340
+ 2024-04-10 22:01:03,585 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
341
+ 2024-04-10 22:01:03,615 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
342
+ 2024-04-10 22:01:03,615 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
343
+ 2024-04-10 22:01:05,691 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
344
+ 2024-04-10 22:01:05,927 INFO Thread-12 :274 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240410_215101-ta1pahei/files/config.yaml
345
+ 2024-04-10 22:01:08,928 INFO Thread-12 :274 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240410_215101-ta1pahei/files/output.log
346
+ 2024-04-10 22:01:10,784 DEBUG SenderThread:274 [sender.py:send():379] send: stats
347
+ 2024-04-10 22:01:11,785 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
348
+ 2024-04-10 22:01:16,786 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
349
+ 2024-04-10 22:01:18,585 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
350
+ 2024-04-10 22:01:18,615 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
351
+ 2024-04-10 22:01:18,616 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
352
+ 2024-04-10 22:01:22,656 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
353
+ 2024-04-10 22:01:27,657 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
354
+ 2024-04-10 22:01:32,658 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
355
+ 2024-04-10 22:01:33,585 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
356
+ 2024-04-10 22:01:33,615 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
357
+ 2024-04-10 22:01:33,615 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
358
+ 2024-04-10 22:01:38,656 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
359
+ 2024-04-10 22:01:40,785 DEBUG SenderThread:274 [sender.py:send():379] send: stats
360
+ 2024-04-10 22:01:43,786 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
361
+ 2024-04-10 22:01:48,588 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
362
+ 2024-04-10 22:01:48,615 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
363
+ 2024-04-10 22:01:48,615 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
364
+ 2024-04-10 22:01:49,779 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
365
+ 2024-04-10 22:01:54,904 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: partial_history
366
+ 2024-04-10 22:01:54,907 DEBUG SenderThread:274 [sender.py:send():379] send: metric
367
+ 2024-04-10 22:01:54,907 DEBUG SenderThread:274 [sender.py:send():379] send: metric
368
+ 2024-04-10 22:01:54,907 DEBUG SenderThread:274 [sender.py:send():379] send: metric
369
+ 2024-04-10 22:01:54,907 DEBUG SenderThread:274 [sender.py:send():379] send: metric
370
+ 2024-04-10 22:01:54,907 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
371
+ 2024-04-10 22:01:54,907 DEBUG SenderThread:274 [sender.py:send():379] send: history
372
+ 2024-04-10 22:01:54,908 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: summary_record
373
+ 2024-04-10 22:01:54,908 INFO SenderThread:274 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
374
+ 2024-04-10 22:01:54,945 INFO Thread-12 :274 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240410_215101-ta1pahei/files/wandb-summary.json
375
+ 2024-04-10 22:01:56,946 INFO Thread-12 :274 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240410_215101-ta1pahei/files/output.log
376
+ 2024-04-10 22:01:57,338 DEBUG SenderThread:274 [sender.py:send():379] send: telemetry
377
+ 2024-04-10 22:01:57,338 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: summary_record
378
+ 2024-04-10 22:01:57,340 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: partial_history
379
+ 2024-04-10 22:01:57,341 INFO SenderThread:274 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
380
+ 2024-04-10 22:01:57,342 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: summary_record
381
+ 2024-04-10 22:01:57,342 INFO SenderThread:274 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
382
+ 2024-04-10 22:01:57,342 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: summary_record
383
+ 2024-04-10 22:01:57,343 INFO SenderThread:274 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
384
+ 2024-04-10 22:01:57,343 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: summary_record
385
+ 2024-04-10 22:01:57,343 INFO SenderThread:274 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
386
+ 2024-04-10 22:01:57,344 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: summary_record
387
+ 2024-04-10 22:01:57,344 INFO SenderThread:274 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
388
+ 2024-04-10 22:01:57,344 DEBUG SenderThread:274 [sender.py:send():379] send: history
389
+ 2024-04-10 22:01:57,345 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: summary_record
390
+ 2024-04-10 22:01:57,345 INFO SenderThread:274 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
391
+ 2024-04-10 22:01:57,946 INFO Thread-12 :274 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240410_215101-ta1pahei/files/wandb-summary.json
392
+ 2024-04-10 22:01:58,946 INFO Thread-12 :274 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240410_215101-ta1pahei/files/output.log
393
+ 2024-04-10 22:02:00,489 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
394
+ 2024-04-10 22:02:03,585 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
395
+ 2024-04-10 22:02:03,615 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
396
+ 2024-04-10 22:02:03,616 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
397
+ 2024-04-10 22:02:05,656 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
398
+ 2024-04-10 22:02:10,662 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
399
+ 2024-04-10 22:02:10,786 DEBUG SenderThread:274 [sender.py:send():379] send: stats
400
+ 2024-04-10 22:02:10,950 INFO Thread-12 :274 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240410_215101-ta1pahei/files/config.yaml
401
+ 2024-04-10 22:02:15,787 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
402
+ 2024-04-10 22:02:18,586 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: internal_messages
403
+ 2024-04-10 22:02:18,615 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: stop_status
404
+ 2024-04-10 22:02:18,616 DEBUG SenderThread:274 [sender.py:send_request():406] send_request: stop_status
405
+ 2024-04-10 22:02:21,656 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
406
+ 2024-04-10 22:02:26,657 DEBUG HandlerThread:274 [handler.py:handle_request():146] handle_request: status_report
wandb/run-20240410_215101-ta1pahei/logs/debug.log ADDED
@@ -0,0 +1,36 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-04-10 21:51:01,068 INFO MainThread:227 [wandb_setup.py:_flush():76] Current SDK version is 0.16.5
2
+ 2024-04-10 21:51:01,068 INFO MainThread:227 [wandb_setup.py:_flush():76] Configure stats pid to 227
3
+ 2024-04-10 21:51:01,068 INFO MainThread:227 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
4
+ 2024-04-10 21:51:01,068 INFO MainThread:227 [wandb_setup.py:_flush():76] Loading settings from /kaggle/working/wandb/settings
5
+ 2024-04-10 21:51:01,068 INFO MainThread:227 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
6
+ 2024-04-10 21:51:01,068 INFO MainThread:227 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
7
+ 2024-04-10 21:51:01,068 INFO MainThread:227 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program': '<python with no main file>'}
8
+ 2024-04-10 21:51:01,068 INFO MainThread:227 [wandb_setup.py:_flush():76] Applying login settings: {}
9
+ 2024-04-10 21:51:01,068 INFO MainThread:227 [wandb_setup.py:_flush():76] Applying login settings: {'api_key': '***REDACTED***'}
10
+ 2024-04-10 21:51:01,068 INFO MainThread:227 [wandb_init.py:_log_setup():527] Logging user logs to /kaggle/working/wandb/run-20240410_215101-ta1pahei/logs/debug.log
11
+ 2024-04-10 21:51:01,069 INFO MainThread:227 [wandb_init.py:_log_setup():528] Logging internal logs to /kaggle/working/wandb/run-20240410_215101-ta1pahei/logs/debug-internal.log
12
+ 2024-04-10 21:51:01,069 INFO MainThread:227 [wandb_init.py:_jupyter_setup():473] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x7b92ac4655a0>
13
+ 2024-04-10 21:51:01,069 INFO MainThread:227 [wandb_init.py:init():567] calling init triggers
14
+ 2024-04-10 21:51:01,069 INFO MainThread:227 [wandb_init.py:init():574] wandb.init called with sweep_config: {}
15
+ config: {}
16
+ 2024-04-10 21:51:01,069 INFO MainThread:227 [wandb_init.py:init():617] starting backend
17
+ 2024-04-10 21:51:01,069 INFO MainThread:227 [wandb_init.py:init():621] setting up manager
18
+ 2024-04-10 21:51:01,071 INFO MainThread:227 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
19
+ 2024-04-10 21:51:01,073 INFO MainThread:227 [wandb_init.py:init():629] backend started and connected
20
+ 2024-04-10 21:51:01,085 INFO MainThread:227 [wandb_run.py:_label_probe_notebook():1299] probe notebook
21
+ 2024-04-10 21:51:01,447 INFO MainThread:227 [wandb_init.py:init():721] updated telemetry
22
+ 2024-04-10 21:51:01,458 INFO MainThread:227 [wandb_init.py:init():754] communicating run to backend with 90.0 second timeout
23
+ 2024-04-10 21:51:01,677 INFO MainThread:227 [wandb_run.py:_on_init():2344] communicating current version
24
+ 2024-04-10 21:51:01,771 INFO MainThread:227 [wandb_run.py:_on_init():2353] got version response upgrade_message: "wandb version 0.16.6 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
25
+
26
+ 2024-04-10 21:51:01,771 INFO MainThread:227 [wandb_init.py:init():805] starting run threads in backend
27
+ 2024-04-10 21:51:17,777 INFO MainThread:227 [wandb_run.py:_console_start():2323] atexit reg
28
+ 2024-04-10 21:51:17,777 INFO MainThread:227 [wandb_run.py:_redirect():2178] redirect: wrap_raw
29
+ 2024-04-10 21:51:17,777 INFO MainThread:227 [wandb_run.py:_redirect():2243] Wrapping output streams.
30
+ 2024-04-10 21:51:17,778 INFO MainThread:227 [wandb_run.py:_redirect():2268] Redirects installed.
31
+ 2024-04-10 21:51:17,779 INFO MainThread:227 [wandb_init.py:init():848] run started, returning control to user process
32
+ 2024-04-10 21:51:17,784 INFO MainThread:227 [wandb_run.py:_config_callback():1347] config_cb None None {'vocab_size': 50304, 'hidden_size': 2048, 'num_hidden_layers': 24, 'num_attention_heads': 32, 'layer_norm_epsilon': 1e-05, 'initializer_range': 0.02, 'use_cache': False, 'hidden_dropout': 0.0, 'attention_dropout': 0.0, 'bos_token_id': 1, 'eos_token_id': 2, 'num_kv_heads': 32, 'alibi': True, 'new_decoder_architecture': False, 'multi_query': False, 'parallel_attn': False, 'bias': True, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'bfloat16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['FalconForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'pad_token_id': None, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'tiiuae/falcon-rw-1b', 'transformers_version': '4.39.3', 'apply_residual_connection_post_layernorm': False, 'auto_map': {'AutoConfig': 'tiiuae/falcon-rw-1b--configuration_falcon.FalconConfig', 'AutoModel': 'tiiuae/falcon-rw-1b--modeling_falcon.FalconModel', 'AutoModelForSequenceClassification': 'tiiuae/falcon-rw-1b--modeling_falcon.FalconForSequenceClassification', 'AutoModelForTokenClassification': 'tiiuae/falcon-rw-1b--modeling_falcon.FalconForTokenClassification', 'AutoModelForQuestionAnswering': 'tiiuae/falcon-rw-1b--modeling_falcon.FalconForQuestionAnswering', 'AutoModelForCausalLM': 'tiiuae/falcon-rw-1b--modeling_falcon.FalconForCausalLM'}, 'model_type': 'falcon', 'quantization_config': {'quant_method': 'QuantizationMethod.BITS_AND_BYTES', '_load_in_8bit': False, '_load_in_4bit': True, 'llm_int8_threshold': 6.0, 'llm_int8_skip_modules': None, 'llm_int8_enable_fp32_cpu_offload': False, 'llm_int8_has_fp16_weight': False, 'bnb_4bit_quant_type': 'nf4', 'bnb_4bit_use_double_quant': False, 'bnb_4bit_compute_dtype': 'float16', 'bnb_4bit_quant_storage': 'uint8', 'load_in_4bit': True, 'load_in_8bit': False}, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': False, 'do_predict': False, 'evaluation_strategy': 'no', 'prediction_loss_only': False, 'per_device_train_batch_size': 8, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 3e-05, 'weight_decay': 0.0001, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 0.3, 'num_train_epochs': 5, 'max_steps': 20, 'lr_scheduler_type': 'cosine', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.03, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Apr10_21-42-21_4bf7c58fba03', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 20, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 20, 'save_total_limit': 1, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 8, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'paged_adamw_8bit', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': True, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': True, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None}
33
+ 2024-04-10 21:51:22,848 INFO MainThread:227 [jupyter.py:save_ipynb():373] not saving jupyter notebook
34
+ 2024-04-10 21:51:22,848 INFO MainThread:227 [wandb_init.py:_pause_backend():438] pausing backend
35
+ 2024-04-10 21:52:40,758 INFO MainThread:227 [wandb_init.py:_resume_backend():443] resuming backend
36
+ 2024-04-10 22:00:44,550 INFO MainThread:227 [wandb_run.py:_config_callback():1347] config_cb None None {'vocab_size': 50304, 'hidden_size': 2048, 'num_hidden_layers': 24, 'num_attention_heads': 32, 'layer_norm_epsilon': 1e-05, 'initializer_range': 0.02, 'use_cache': False, 'hidden_dropout': 0.0, 'attention_dropout': 0.0, 'bos_token_id': 1, 'eos_token_id': 2, 'num_kv_heads': 32, 'alibi': True, 'new_decoder_architecture': False, 'multi_query': False, 'parallel_attn': False, 'bias': True, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'bfloat16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['FalconForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'pad_token_id': None, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'tiiuae/falcon-rw-1b', 'transformers_version': '4.39.3', 'apply_residual_connection_post_layernorm': False, 'auto_map': {'AutoConfig': 'tiiuae/falcon-rw-1b--configuration_falcon.FalconConfig', 'AutoModel': 'tiiuae/falcon-rw-1b--modeling_falcon.FalconModel', 'AutoModelForSequenceClassification': 'tiiuae/falcon-rw-1b--modeling_falcon.FalconForSequenceClassification', 'AutoModelForTokenClassification': 'tiiuae/falcon-rw-1b--modeling_falcon.FalconForTokenClassification', 'AutoModelForQuestionAnswering': 'tiiuae/falcon-rw-1b--modeling_falcon.FalconForQuestionAnswering', 'AutoModelForCausalLM': 'tiiuae/falcon-rw-1b--modeling_falcon.FalconForCausalLM'}, 'model_type': 'falcon', 'quantization_config': {'quant_method': 'QuantizationMethod.BITS_AND_BYTES', '_load_in_8bit': False, '_load_in_4bit': True, 'llm_int8_threshold': 6.0, 'llm_int8_skip_modules': None, 'llm_int8_enable_fp32_cpu_offload': False, 'llm_int8_has_fp16_weight': False, 'bnb_4bit_quant_type': 'nf4', 'bnb_4bit_use_double_quant': False, 'bnb_4bit_compute_dtype': 'float16', 'bnb_4bit_quant_storage': 'uint8', 'load_in_4bit': True, 'load_in_8bit': False}, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': False, 'do_predict': False, 'evaluation_strategy': 'no', 'prediction_loss_only': False, 'per_device_train_batch_size': 8, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 3e-05, 'weight_decay': 0.0001, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 0.3, 'num_train_epochs': 5, 'max_steps': 20, 'lr_scheduler_type': 'cosine', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.03, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Apr10_21-52-40_4bf7c58fba03', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 20, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 20, 'save_total_limit': 1, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 8, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'paged_adamw_8bit', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': True, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': True, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None}
wandb/run-20240410_215101-ta1pahei/run-ta1pahei.wandb ADDED
Binary file (44.7 kB). View file