Femboyuwu2000 commited on
Commit
0f88c91
1 Parent(s): 5062d14

Training in progress, step 20

Browse files
adapter_config.json CHANGED
@@ -20,8 +20,8 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "query_key_valuelm_head",
24
- "word_embeddings"
25
  ],
26
  "task_type": "CAUSAL_LM",
27
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
+ "word_embeddings",
24
+ "query_key_valuelm_head"
25
  ],
26
  "task_type": "CAUSAL_LM",
27
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:de52612f18d94eb85267cdbaa87cfac914b5a3415370220f048ff6627b5dc973
3
  size 8077608
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4f5b4211357ffac39570b1f6e86556f478b9274661061048dc057539e048d76a
3
  size 8077608
runs/Apr12_07-35-52_e5a48bec8248/events.out.tfevents.1712907354.e5a48bec8248.289.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a2a24bc3f89fa8d99361dc24608208c2de729e4866df68e8d394ffeac11a1e43
3
+ size 5489
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:83256b55088d45d0ebeda1319a91e9b247c70b5f4d82cb04a018aa4a684554bc
3
  size 4984
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:32e994f25267341e613f4d352977cc7a2847de358db5ece7a60fcf21be944170
3
  size 4984
wandb/debug-internal.log CHANGED
@@ -1,32 +1,32 @@
1
- 2024-04-12 07:27:36,633 INFO StreamThr :216 [internal.py:wandb_internal():86] W&B internal server running at pid: 216, started at: 2024-04-12 07:27:36.632321
2
- 2024-04-12 07:27:36,634 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status
3
- 2024-04-12 07:27:36,923 INFO WriterThread:216 [datastore.py:open_for_write():87] open: /kaggle/working/wandb/run-20240412_072736-gyl3aena/run-gyl3aena.wandb
4
- 2024-04-12 07:27:36,923 DEBUG SenderThread:216 [sender.py:send():379] send: header
5
- 2024-04-12 07:27:36,926 DEBUG SenderThread:216 [sender.py:send():379] send: run
6
- 2024-04-12 07:27:37,028 INFO SenderThread:216 [dir_watcher.py:__init__():211] watching files in: /kaggle/working/wandb/run-20240412_072736-gyl3aena/files
7
- 2024-04-12 07:27:37,029 INFO SenderThread:216 [sender.py:_start_run_threads():1124] run started: gyl3aena with start time 1712906856.632079
8
- 2024-04-12 07:27:37,038 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: check_version
9
- 2024-04-12 07:27:37,038 DEBUG SenderThread:216 [sender.py:send_request():406] send_request: check_version
10
- 2024-04-12 07:27:37,127 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: run_start
11
- 2024-04-12 07:27:37,138 DEBUG HandlerThread:216 [system_info.py:__init__():26] System info init
12
- 2024-04-12 07:27:37,139 DEBUG HandlerThread:216 [system_info.py:__init__():41] System info init done
13
- 2024-04-12 07:27:37,139 INFO HandlerThread:216 [system_monitor.py:start():194] Starting system monitor
14
- 2024-04-12 07:27:37,139 INFO SystemMonitor:216 [system_monitor.py:_start():158] Starting system asset monitoring threads
15
- 2024-04-12 07:27:37,139 INFO HandlerThread:216 [system_monitor.py:probe():214] Collecting system info
16
- 2024-04-12 07:27:37,140 INFO SystemMonitor:216 [interfaces.py:start():190] Started cpu monitoring
17
- 2024-04-12 07:27:37,140 INFO SystemMonitor:216 [interfaces.py:start():190] Started disk monitoring
18
- 2024-04-12 07:27:37,141 INFO SystemMonitor:216 [interfaces.py:start():190] Started gpu monitoring
19
- 2024-04-12 07:27:37,142 INFO SystemMonitor:216 [interfaces.py:start():190] Started memory monitoring
20
- 2024-04-12 07:27:37,144 INFO SystemMonitor:216 [interfaces.py:start():190] Started network monitoring
21
- 2024-04-12 07:27:37,154 DEBUG HandlerThread:216 [system_info.py:probe():150] Probing system
22
- 2024-04-12 07:27:37,156 DEBUG HandlerThread:216 [gitlib.py:_init_repo():56] git repository is invalid
23
- 2024-04-12 07:27:37,156 DEBUG HandlerThread:216 [system_info.py:probe():198] Probing system done
24
- 2024-04-12 07:27:37,156 DEBUG HandlerThread:216 [system_monitor.py:probe():223] {'os': 'Linux-5.15.133+-x86_64-with-glibc2.31', 'python': '3.10.13', 'heartbeatAt': '2024-04-12T07:27:37.154498', 'startedAt': '2024-04-12T07:27:36.625772', 'docker': None, 'cuda': None, 'args': (), 'state': 'running', 'program': 'kaggle.ipynb', 'codePathLocal': None, 'root': '/kaggle/working', 'host': 'e5a48bec8248', 'username': 'root', 'executable': '/opt/conda/bin/python3.10', 'cpu_count': 2, 'cpu_count_logical': 4, 'cpu_freq': {'current': 2000.138, 'min': 0.0, 'max': 0.0}, 'cpu_freq_per_core': [{'current': 2000.138, 'min': 0.0, 'max': 0.0}, {'current': 2000.138, 'min': 0.0, 'max': 0.0}, {'current': 2000.138, 'min': 0.0, 'max': 0.0}, {'current': 2000.138, 'min': 0.0, 'max': 0.0}], 'disk': {'/': {'total': 8062.387607574463, 'used': 5565.782409667969}}, 'gpu': 'Tesla T4', 'gpu_count': 2, 'gpu_devices': [{'name': 'Tesla T4', 'memory_total': 16106127360}, {'name': 'Tesla T4', 'memory_total': 16106127360}], 'memory': {'total': 31.357559204101562}}
25
- 2024-04-12 07:27:37,156 INFO HandlerThread:216 [system_monitor.py:probe():224] Finished collecting system info
26
- 2024-04-12 07:27:37,156 INFO HandlerThread:216 [system_monitor.py:probe():227] Publishing system info
27
- 2024-04-12 07:27:37,156 DEBUG HandlerThread:216 [system_info.py:_save_conda():207] Saving list of conda packages installed into the current environment
28
- 2024-04-12 07:27:38,030 INFO Thread-12 :216 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240412_072736-gyl3aena/files/conda-environment.yaml
29
- 2024-04-12 07:27:52,171 ERROR HandlerThread:216 [system_info.py:_save_conda():221] Error saving conda packages: Command '['conda', 'env', 'export']' timed out after 15 seconds
30
  Traceback (most recent call last):
31
  File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/internal/system/system_info.py", line 214, in _save_conda
32
  subprocess.call(
@@ -37,276 +37,41 @@ Traceback (most recent call last):
37
  File "/opt/conda/lib/python3.10/subprocess.py", line 1951, in _wait
38
  raise TimeoutExpired(self.args, timeout)
39
  subprocess.TimeoutExpired: Command '['conda', 'env', 'export']' timed out after 15 seconds
40
- 2024-04-12 07:27:52,172 DEBUG HandlerThread:216 [system_info.py:_save_conda():222] Saving conda packages done
41
- 2024-04-12 07:27:52,173 INFO HandlerThread:216 [system_monitor.py:probe():229] Finished publishing system info
42
- 2024-04-12 07:27:52,178 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
43
- 2024-04-12 07:27:52,178 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: keepalive
44
- 2024-04-12 07:27:52,178 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
45
- 2024-04-12 07:27:52,178 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: keepalive
46
- 2024-04-12 07:27:52,178 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
47
- 2024-04-12 07:27:52,178 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: keepalive
48
- 2024-04-12 07:27:52,179 DEBUG SenderThread:216 [sender.py:send():379] send: files
49
- 2024-04-12 07:27:52,179 INFO SenderThread:216 [sender.py:_save_file():1390] saving file wandb-metadata.json with policy now
50
- 2024-04-12 07:27:52,388 INFO wandb-upload_0:216 [upload_job.py:push():131] Uploaded file /tmp/tmpysomol4ewandb/y0nsu16m-wandb-metadata.json
51
- 2024-04-12 07:27:53,034 INFO Thread-12 :216 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240412_072736-gyl3aena/files/wandb-metadata.json
52
- 2024-04-12 07:27:53,192 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: python_packages
53
- 2024-04-12 07:27:53,193 DEBUG SenderThread:216 [sender.py:send_request():406] send_request: python_packages
54
- 2024-04-12 07:27:53,196 DEBUG SenderThread:216 [sender.py:send():379] send: telemetry
55
- 2024-04-12 07:27:53,207 DEBUG SenderThread:216 [sender.py:send():379] send: config
56
- 2024-04-12 07:27:53,209 DEBUG SenderThread:216 [sender.py:send():379] send: metric
57
- 2024-04-12 07:27:53,210 DEBUG SenderThread:216 [sender.py:send():379] send: telemetry
58
- 2024-04-12 07:27:53,210 DEBUG SenderThread:216 [sender.py:send():379] send: metric
59
- 2024-04-12 07:27:53,210 WARNING SenderThread:216 [sender.py:send_metric():1341] Seen metric with glob (shouldn't happen)
60
- 2024-04-12 07:27:53,211 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: internal_messages
61
- 2024-04-12 07:27:53,211 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: stop_status
62
- 2024-04-12 07:27:53,211 DEBUG SenderThread:216 [sender.py:send():379] send: telemetry
63
- 2024-04-12 07:27:53,212 DEBUG SenderThread:216 [sender.py:send_request():406] send_request: stop_status
64
- 2024-04-12 07:27:54,034 INFO Thread-12 :216 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240412_072736-gyl3aena/files/requirements.txt
65
- 2024-04-12 07:27:54,035 INFO Thread-12 :216 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240412_072736-gyl3aena/files/output.log
66
- 2024-04-12 07:27:56,035 INFO Thread-12 :216 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240412_072736-gyl3aena/files/output.log
67
- 2024-04-12 07:27:57,730 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
68
- 2024-04-12 07:28:00,302 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: partial_history
69
- 2024-04-12 07:28:00,305 DEBUG SenderThread:216 [sender.py:send():379] send: metric
70
- 2024-04-12 07:28:00,305 DEBUG SenderThread:216 [sender.py:send():379] send: metric
71
- 2024-04-12 07:28:00,306 DEBUG SenderThread:216 [sender.py:send():379] send: metric
72
- 2024-04-12 07:28:00,306 DEBUG SenderThread:216 [sender.py:send():379] send: metric
73
- 2024-04-12 07:28:00,306 DEBUG SenderThread:216 [sender.py:send():379] send: history
74
- 2024-04-12 07:28:00,306 DEBUG SenderThread:216 [sender.py:send_request():406] send_request: summary_record
75
- 2024-04-12 07:28:00,308 INFO SenderThread:216 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
76
- 2024-04-12 07:28:01,037 INFO Thread-12 :216 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240412_072736-gyl3aena/files/wandb-summary.json
77
- 2024-04-12 07:28:02,038 INFO Thread-12 :216 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240412_072736-gyl3aena/files/output.log
78
- 2024-04-12 07:28:03,008 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
79
- 2024-04-12 07:28:06,847 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: partial_history
80
- 2024-04-12 07:28:06,848 DEBUG SenderThread:216 [sender.py:send():379] send: history
81
- 2024-04-12 07:28:06,849 DEBUG SenderThread:216 [sender.py:send_request():406] send_request: summary_record
82
- 2024-04-12 07:28:06,851 INFO SenderThread:216 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
83
- 2024-04-12 07:28:07,040 INFO Thread-12 :216 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240412_072736-gyl3aena/files/wandb-summary.json
84
- 2024-04-12 07:28:08,201 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
85
- 2024-04-12 07:28:08,201 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: stop_status
86
- 2024-04-12 07:28:08,202 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: internal_messages
87
- 2024-04-12 07:28:08,314 DEBUG SenderThread:216 [sender.py:send_request():406] send_request: stop_status
88
- 2024-04-12 07:28:09,041 INFO Thread-12 :216 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240412_072736-gyl3aena/files/config.yaml
89
- 2024-04-12 07:28:10,041 INFO Thread-12 :216 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240412_072736-gyl3aena/files/output.log
90
- 2024-04-12 07:28:13,125 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: partial_history
91
- 2024-04-12 07:28:13,126 DEBUG SenderThread:216 [sender.py:send():379] send: history
92
- 2024-04-12 07:28:13,126 DEBUG SenderThread:216 [sender.py:send_request():406] send_request: summary_record
93
- 2024-04-12 07:28:13,128 INFO SenderThread:216 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
94
- 2024-04-12 07:28:13,806 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: partial_history
95
- 2024-04-12 07:28:13,807 DEBUG SenderThread:216 [sender.py:send():379] send: telemetry
96
- 2024-04-12 07:28:13,807 DEBUG SenderThread:216 [sender.py:send_request():406] send_request: summary_record
97
- 2024-04-12 07:28:13,807 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
98
- 2024-04-12 07:28:13,808 INFO SenderThread:216 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
99
- 2024-04-12 07:28:13,809 DEBUG SenderThread:216 [sender.py:send_request():406] send_request: summary_record
100
- 2024-04-12 07:28:13,809 INFO SenderThread:216 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
101
- 2024-04-12 07:28:13,809 DEBUG SenderThread:216 [sender.py:send_request():406] send_request: summary_record
102
- 2024-04-12 07:28:13,810 INFO SenderThread:216 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
103
- 2024-04-12 07:28:13,810 DEBUG SenderThread:216 [sender.py:send_request():406] send_request: summary_record
104
- 2024-04-12 07:28:13,810 INFO SenderThread:216 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
105
- 2024-04-12 07:28:13,811 DEBUG SenderThread:216 [sender.py:send_request():406] send_request: summary_record
106
- 2024-04-12 07:28:13,811 INFO SenderThread:216 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
107
- 2024-04-12 07:28:13,812 DEBUG SenderThread:216 [sender.py:send():379] send: history
108
- 2024-04-12 07:28:13,812 DEBUG SenderThread:216 [sender.py:send_request():406] send_request: summary_record
109
- 2024-04-12 07:28:13,812 INFO SenderThread:216 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
110
- 2024-04-12 07:28:14,043 INFO Thread-12 :216 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240412_072736-gyl3aena/files/wandb-summary.json
111
- 2024-04-12 07:28:18,814 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
112
- 2024-04-12 07:28:23,193 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: stop_status
113
- 2024-04-12 07:28:23,194 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: internal_messages
114
- 2024-04-12 07:28:23,194 DEBUG SenderThread:216 [sender.py:send_request():406] send_request: stop_status
115
- 2024-04-12 07:28:24,273 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
116
- 2024-04-12 07:28:26,760 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: pause
117
- 2024-04-12 07:28:26,760 INFO HandlerThread:216 [handler.py:handle_request_pause():708] stopping system metrics thread
118
- 2024-04-12 07:28:26,760 INFO HandlerThread:216 [system_monitor.py:finish():203] Stopping system monitor
119
- 2024-04-12 07:28:26,761 INFO HandlerThread:216 [interfaces.py:finish():202] Joined cpu monitor
120
- 2024-04-12 07:28:26,761 INFO HandlerThread:216 [interfaces.py:finish():202] Joined disk monitor
121
- 2024-04-12 07:28:26,761 DEBUG SystemMonitor:216 [system_monitor.py:_start():172] Starting system metrics aggregation loop
122
- 2024-04-12 07:28:26,761 DEBUG SystemMonitor:216 [system_monitor.py:_start():179] Finished system metrics aggregation loop
123
- 2024-04-12 07:28:26,762 DEBUG SystemMonitor:216 [system_monitor.py:_start():183] Publishing last batch of metrics
124
- 2024-04-12 07:28:26,774 INFO HandlerThread:216 [interfaces.py:finish():202] Joined gpu monitor
125
- 2024-04-12 07:28:26,774 INFO HandlerThread:216 [interfaces.py:finish():202] Joined memory monitor
126
- 2024-04-12 07:28:26,774 INFO HandlerThread:216 [interfaces.py:finish():202] Joined network monitor
127
- 2024-04-12 07:28:26,775 DEBUG SenderThread:216 [sender.py:send():379] send: stats
128
- 2024-04-12 07:28:29,776 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
129
- 2024-04-12 07:28:34,777 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
130
- 2024-04-12 07:28:38,193 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: stop_status
131
- 2024-04-12 07:28:38,194 DEBUG SenderThread:216 [sender.py:send_request():406] send_request: stop_status
132
- 2024-04-12 07:28:38,197 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: internal_messages
133
- 2024-04-12 07:28:40,261 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
134
- 2024-04-12 07:28:41,053 INFO Thread-12 :216 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240412_072736-gyl3aena/files/config.yaml
135
- 2024-04-12 07:28:45,363 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
136
- 2024-04-12 07:28:50,364 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
137
- 2024-04-12 07:28:53,193 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: stop_status
138
- 2024-04-12 07:28:53,194 DEBUG SenderThread:216 [sender.py:send_request():406] send_request: stop_status
139
- 2024-04-12 07:28:53,234 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: internal_messages
140
- 2024-04-12 07:28:56,263 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
141
- 2024-04-12 07:29:01,264 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
142
- 2024-04-12 07:29:06,266 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
143
- 2024-04-12 07:29:08,194 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: stop_status
144
- 2024-04-12 07:29:08,194 DEBUG SenderThread:216 [sender.py:send_request():406] send_request: stop_status
145
- 2024-04-12 07:29:08,235 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: internal_messages
146
- 2024-04-12 07:29:12,239 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
147
- 2024-04-12 07:29:17,240 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
148
- 2024-04-12 07:29:22,241 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
149
- 2024-04-12 07:29:23,194 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: stop_status
150
- 2024-04-12 07:29:23,194 DEBUG SenderThread:216 [sender.py:send_request():406] send_request: stop_status
151
- 2024-04-12 07:29:23,235 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: internal_messages
152
- 2024-04-12 07:29:27,289 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
153
- 2024-04-12 07:29:32,290 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
154
- 2024-04-12 07:29:37,291 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
155
- 2024-04-12 07:29:38,194 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: stop_status
156
- 2024-04-12 07:29:38,195 DEBUG SenderThread:216 [sender.py:send_request():406] send_request: stop_status
157
- 2024-04-12 07:29:38,235 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: internal_messages
158
- 2024-04-12 07:29:42,325 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
159
- 2024-04-12 07:29:47,326 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
160
- 2024-04-12 07:29:52,327 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
161
- 2024-04-12 07:29:53,194 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: stop_status
162
- 2024-04-12 07:29:53,195 DEBUG SenderThread:216 [sender.py:send_request():406] send_request: stop_status
163
- 2024-04-12 07:29:53,235 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: internal_messages
164
- 2024-04-12 07:29:58,280 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
165
- 2024-04-12 07:30:03,281 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
166
- 2024-04-12 07:30:08,194 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: stop_status
167
- 2024-04-12 07:30:08,195 DEBUG SenderThread:216 [sender.py:send_request():406] send_request: stop_status
168
- 2024-04-12 07:30:08,235 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: internal_messages
169
- 2024-04-12 07:30:09,269 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
170
- 2024-04-12 07:30:14,270 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
171
- 2024-04-12 07:30:19,271 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
172
- 2024-04-12 07:30:23,195 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: stop_status
173
- 2024-04-12 07:30:23,195 DEBUG SenderThread:216 [sender.py:send_request():406] send_request: stop_status
174
- 2024-04-12 07:30:23,235 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: internal_messages
175
- 2024-04-12 07:30:24,322 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
176
- 2024-04-12 07:30:29,323 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
177
- 2024-04-12 07:30:34,324 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
178
- 2024-04-12 07:30:38,195 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: stop_status
179
- 2024-04-12 07:30:38,195 DEBUG SenderThread:216 [sender.py:send_request():406] send_request: stop_status
180
- 2024-04-12 07:30:38,236 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: internal_messages
181
- 2024-04-12 07:30:39,325 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
182
- 2024-04-12 07:30:44,326 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
183
- 2024-04-12 07:30:49,327 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
184
- 2024-04-12 07:30:53,195 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: stop_status
185
- 2024-04-12 07:30:53,196 DEBUG SenderThread:216 [sender.py:send_request():406] send_request: stop_status
186
- 2024-04-12 07:30:53,236 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: internal_messages
187
- 2024-04-12 07:30:55,310 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
188
- 2024-04-12 07:31:00,311 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
189
- 2024-04-12 07:31:05,312 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
190
- 2024-04-12 07:31:08,195 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: stop_status
191
- 2024-04-12 07:31:08,196 DEBUG SenderThread:216 [sender.py:send_request():406] send_request: stop_status
192
- 2024-04-12 07:31:08,236 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: internal_messages
193
- 2024-04-12 07:31:11,254 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
194
- 2024-04-12 07:31:16,255 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
195
- 2024-04-12 07:31:21,256 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
196
- 2024-04-12 07:31:23,195 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: stop_status
197
- 2024-04-12 07:31:23,196 DEBUG SenderThread:216 [sender.py:send_request():406] send_request: stop_status
198
- 2024-04-12 07:31:23,236 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: internal_messages
199
- 2024-04-12 07:31:26,273 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
200
- 2024-04-12 07:31:31,274 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
201
- 2024-04-12 07:31:36,275 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
202
- 2024-04-12 07:31:38,195 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: stop_status
203
- 2024-04-12 07:31:38,196 DEBUG SenderThread:216 [sender.py:send_request():406] send_request: stop_status
204
- 2024-04-12 07:31:38,236 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: internal_messages
205
- 2024-04-12 07:31:41,323 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
206
- 2024-04-12 07:31:46,325 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
207
- 2024-04-12 07:31:51,326 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
208
- 2024-04-12 07:31:53,196 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: stop_status
209
- 2024-04-12 07:31:53,196 DEBUG SenderThread:216 [sender.py:send_request():406] send_request: stop_status
210
- 2024-04-12 07:31:53,236 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: internal_messages
211
- 2024-04-12 07:31:57,277 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
212
- 2024-04-12 07:32:02,278 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
213
- 2024-04-12 07:32:07,279 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
214
- 2024-04-12 07:32:08,196 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: stop_status
215
- 2024-04-12 07:32:08,196 DEBUG SenderThread:216 [sender.py:send_request():406] send_request: stop_status
216
- 2024-04-12 07:32:08,237 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: internal_messages
217
- 2024-04-12 07:32:12,341 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
218
- 2024-04-12 07:32:17,342 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
219
- 2024-04-12 07:32:22,343 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
220
- 2024-04-12 07:32:23,196 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: stop_status
221
- 2024-04-12 07:32:23,196 DEBUG SenderThread:216 [sender.py:send_request():406] send_request: stop_status
222
- 2024-04-12 07:32:23,237 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: internal_messages
223
- 2024-04-12 07:32:28,314 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
224
- 2024-04-12 07:32:33,315 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
225
- 2024-04-12 07:32:38,196 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: stop_status
226
- 2024-04-12 07:32:38,197 DEBUG SenderThread:216 [sender.py:send_request():406] send_request: stop_status
227
- 2024-04-12 07:32:38,237 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: internal_messages
228
- 2024-04-12 07:32:39,296 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
229
- 2024-04-12 07:32:44,297 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
230
- 2024-04-12 07:32:49,299 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
231
- 2024-04-12 07:32:53,196 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: stop_status
232
- 2024-04-12 07:32:53,197 DEBUG SenderThread:216 [sender.py:send_request():406] send_request: stop_status
233
- 2024-04-12 07:32:53,237 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: internal_messages
234
- 2024-04-12 07:32:55,259 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
235
- 2024-04-12 07:33:00,260 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
236
- 2024-04-12 07:33:05,261 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
237
- 2024-04-12 07:33:08,196 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: stop_status
238
- 2024-04-12 07:33:08,197 DEBUG SenderThread:216 [sender.py:send_request():406] send_request: stop_status
239
- 2024-04-12 07:33:08,237 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: internal_messages
240
- 2024-04-12 07:33:10,292 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
241
- 2024-04-12 07:33:15,293 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
242
- 2024-04-12 07:33:20,294 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
243
- 2024-04-12 07:33:23,197 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: stop_status
244
- 2024-04-12 07:33:23,197 DEBUG SenderThread:216 [sender.py:send_request():406] send_request: stop_status
245
- 2024-04-12 07:33:23,237 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: internal_messages
246
- 2024-04-12 07:33:26,254 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
247
- 2024-04-12 07:33:31,255 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
248
- 2024-04-12 07:33:36,256 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
249
- 2024-04-12 07:33:38,197 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: stop_status
250
- 2024-04-12 07:33:38,197 DEBUG SenderThread:216 [sender.py:send_request():406] send_request: stop_status
251
- 2024-04-12 07:33:38,238 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: internal_messages
252
- 2024-04-12 07:33:41,304 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
253
- 2024-04-12 07:33:46,269 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: resume
254
- 2024-04-12 07:33:46,269 INFO HandlerThread:216 [handler.py:handle_request_resume():699] starting system metrics thread
255
- 2024-04-12 07:33:46,269 INFO HandlerThread:216 [system_monitor.py:start():194] Starting system monitor
256
- 2024-04-12 07:33:46,269 INFO SystemMonitor:216 [system_monitor.py:_start():158] Starting system asset monitoring threads
257
- 2024-04-12 07:33:46,270 INFO SystemMonitor:216 [interfaces.py:start():190] Started cpu monitoring
258
- 2024-04-12 07:33:46,271 INFO SystemMonitor:216 [interfaces.py:start():190] Started disk monitoring
259
- 2024-04-12 07:33:46,271 INFO SystemMonitor:216 [interfaces.py:start():190] Started gpu monitoring
260
- 2024-04-12 07:33:46,272 INFO SystemMonitor:216 [interfaces.py:start():190] Started memory monitoring
261
- 2024-04-12 07:33:46,273 INFO SystemMonitor:216 [interfaces.py:start():190] Started network monitoring
262
- 2024-04-12 07:33:46,304 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
263
- 2024-04-12 07:33:50,616 DEBUG SenderThread:216 [sender.py:send():379] send: config
264
- 2024-04-12 07:33:50,617 DEBUG SenderThread:216 [sender.py:send():379] send: metric
265
- 2024-04-12 07:33:50,617 DEBUG SenderThread:216 [sender.py:send():379] send: metric
266
- 2024-04-12 07:33:50,617 WARNING SenderThread:216 [sender.py:send_metric():1341] Seen metric with glob (shouldn't happen)
267
- 2024-04-12 07:33:51,634 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
268
- 2024-04-12 07:33:52,173 INFO Thread-12 :216 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240412_072736-gyl3aena/files/config.yaml
269
- 2024-04-12 07:33:52,173 INFO Thread-12 :216 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240412_072736-gyl3aena/files/output.log
270
- 2024-04-12 07:33:53,197 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: stop_status
271
- 2024-04-12 07:33:53,197 DEBUG SenderThread:216 [sender.py:send_request():406] send_request: stop_status
272
- 2024-04-12 07:33:53,200 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: internal_messages
273
- 2024-04-12 07:33:57,222 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: partial_history
274
- 2024-04-12 07:33:57,224 DEBUG SenderThread:216 [sender.py:send():379] send: history
275
- 2024-04-12 07:33:57,224 DEBUG SenderThread:216 [sender.py:send_request():406] send_request: summary_record
276
- 2024-04-12 07:33:57,224 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
277
- 2024-04-12 07:33:57,225 INFO SenderThread:216 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
278
- 2024-04-12 07:33:58,175 INFO Thread-12 :216 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240412_072736-gyl3aena/files/wandb-summary.json
279
- 2024-04-12 07:34:00,176 INFO Thread-12 :216 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240412_072736-gyl3aena/files/output.log
280
- 2024-04-12 07:34:03,010 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
281
- 2024-04-12 07:34:04,112 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: partial_history
282
- 2024-04-12 07:34:04,114 DEBUG SenderThread:216 [sender.py:send():379] send: history
283
- 2024-04-12 07:34:04,114 DEBUG SenderThread:216 [sender.py:send_request():406] send_request: summary_record
284
- 2024-04-12 07:34:04,114 INFO SenderThread:216 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
285
- 2024-04-12 07:34:04,178 INFO Thread-12 :216 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240412_072736-gyl3aena/files/wandb-summary.json
286
- 2024-04-12 07:34:06,179 INFO Thread-12 :216 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240412_072736-gyl3aena/files/output.log
287
- 2024-04-12 07:34:08,197 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: stop_status
288
- 2024-04-12 07:34:08,198 DEBUG SenderThread:216 [sender.py:send_request():406] send_request: stop_status
289
- 2024-04-12 07:34:08,201 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: internal_messages
290
- 2024-04-12 07:34:08,317 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
291
- 2024-04-12 07:34:10,454 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: partial_history
292
- 2024-04-12 07:34:10,455 DEBUG SenderThread:216 [sender.py:send():379] send: history
293
- 2024-04-12 07:34:10,455 DEBUG SenderThread:216 [sender.py:send_request():406] send_request: summary_record
294
- 2024-04-12 07:34:10,456 INFO SenderThread:216 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
295
- 2024-04-12 07:34:11,182 DEBUG SenderThread:216 [sender.py:send_request():406] send_request: summary_record
296
- 2024-04-12 07:34:11,184 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: partial_history
297
- 2024-04-12 07:34:11,185 INFO SenderThread:216 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
298
- 2024-04-12 07:34:11,185 DEBUG SenderThread:216 [sender.py:send_request():406] send_request: summary_record
299
- 2024-04-12 07:34:11,186 INFO SenderThread:216 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
300
- 2024-04-12 07:34:11,187 DEBUG SenderThread:216 [sender.py:send_request():406] send_request: summary_record
301
- 2024-04-12 07:34:11,187 INFO Thread-12 :216 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240412_072736-gyl3aena/files/wandb-summary.json
302
- 2024-04-12 07:34:11,188 INFO SenderThread:216 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
303
- 2024-04-12 07:34:11,188 DEBUG SenderThread:216 [sender.py:send_request():406] send_request: summary_record
304
- 2024-04-12 07:34:11,188 INFO SenderThread:216 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
305
- 2024-04-12 07:34:11,189 DEBUG SenderThread:216 [sender.py:send_request():406] send_request: summary_record
306
- 2024-04-12 07:34:11,189 INFO SenderThread:216 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
307
- 2024-04-12 07:34:11,189 DEBUG SenderThread:216 [sender.py:send():379] send: history
308
- 2024-04-12 07:34:11,189 DEBUG SenderThread:216 [sender.py:send_request():406] send_request: summary_record
309
- 2024-04-12 07:34:11,190 INFO SenderThread:216 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
310
- 2024-04-12 07:34:12,187 INFO Thread-12 :216 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240412_072736-gyl3aena/files/wandb-summary.json
311
- 2024-04-12 07:34:14,191 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
312
- 2024-04-12 07:34:19,191 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
 
1
+ 2024-04-12 07:35:55,749 INFO StreamThr :334 [internal.py:wandb_internal():86] W&B internal server running at pid: 334, started at: 2024-04-12 07:35:55.748821
2
+ 2024-04-12 07:35:55,751 DEBUG HandlerThread:334 [handler.py:handle_request():146] handle_request: status
3
+ 2024-04-12 07:35:56,159 INFO WriterThread:334 [datastore.py:open_for_write():87] open: /kaggle/working/wandb/run-20240412_073555-bw7oy9ix/run-bw7oy9ix.wandb
4
+ 2024-04-12 07:35:56,160 DEBUG SenderThread:334 [sender.py:send():379] send: header
5
+ 2024-04-12 07:35:56,163 DEBUG SenderThread:334 [sender.py:send():379] send: run
6
+ 2024-04-12 07:35:56,307 INFO SenderThread:334 [dir_watcher.py:__init__():211] watching files in: /kaggle/working/wandb/run-20240412_073555-bw7oy9ix/files
7
+ 2024-04-12 07:35:56,307 INFO SenderThread:334 [sender.py:_start_run_threads():1124] run started: bw7oy9ix with start time 1712907355.748886
8
+ 2024-04-12 07:35:56,317 DEBUG HandlerThread:334 [handler.py:handle_request():146] handle_request: check_version
9
+ 2024-04-12 07:35:56,317 DEBUG SenderThread:334 [sender.py:send_request():406] send_request: check_version
10
+ 2024-04-12 07:35:56,409 DEBUG HandlerThread:334 [handler.py:handle_request():146] handle_request: run_start
11
+ 2024-04-12 07:35:56,421 DEBUG HandlerThread:334 [system_info.py:__init__():26] System info init
12
+ 2024-04-12 07:35:56,421 DEBUG HandlerThread:334 [system_info.py:__init__():41] System info init done
13
+ 2024-04-12 07:35:56,422 INFO HandlerThread:334 [system_monitor.py:start():194] Starting system monitor
14
+ 2024-04-12 07:35:56,422 INFO SystemMonitor:334 [system_monitor.py:_start():158] Starting system asset monitoring threads
15
+ 2024-04-12 07:35:56,422 INFO HandlerThread:334 [system_monitor.py:probe():214] Collecting system info
16
+ 2024-04-12 07:35:56,423 INFO SystemMonitor:334 [interfaces.py:start():190] Started cpu monitoring
17
+ 2024-04-12 07:35:56,423 INFO SystemMonitor:334 [interfaces.py:start():190] Started disk monitoring
18
+ 2024-04-12 07:35:56,424 INFO SystemMonitor:334 [interfaces.py:start():190] Started gpu monitoring
19
+ 2024-04-12 07:35:56,425 INFO SystemMonitor:334 [interfaces.py:start():190] Started memory monitoring
20
+ 2024-04-12 07:35:56,426 INFO SystemMonitor:334 [interfaces.py:start():190] Started network monitoring
21
+ 2024-04-12 07:35:56,437 DEBUG HandlerThread:334 [system_info.py:probe():150] Probing system
22
+ 2024-04-12 07:35:56,439 DEBUG HandlerThread:334 [gitlib.py:_init_repo():56] git repository is invalid
23
+ 2024-04-12 07:35:56,439 DEBUG HandlerThread:334 [system_info.py:probe():198] Probing system done
24
+ 2024-04-12 07:35:56,439 DEBUG HandlerThread:334 [system_monitor.py:probe():223] {'os': 'Linux-5.15.133+-x86_64-with-glibc2.31', 'python': '3.10.13', 'heartbeatAt': '2024-04-12T07:35:56.437403', 'startedAt': '2024-04-12T07:35:55.741618', 'docker': None, 'cuda': None, 'args': (), 'state': 'running', 'program': 'kaggle.ipynb', 'codePathLocal': None, 'root': '/kaggle/working', 'host': 'e5a48bec8248', 'username': 'root', 'executable': '/opt/conda/bin/python3.10', 'cpu_count': 2, 'cpu_count_logical': 4, 'cpu_freq': {'current': 2000.138, 'min': 0.0, 'max': 0.0}, 'cpu_freq_per_core': [{'current': 2000.138, 'min': 0.0, 'max': 0.0}, {'current': 2000.138, 'min': 0.0, 'max': 0.0}, {'current': 2000.138, 'min': 0.0, 'max': 0.0}, {'current': 2000.138, 'min': 0.0, 'max': 0.0}], 'disk': {'/': {'total': 8062.387607574463, 'used': 5565.782459259033}}, 'gpu': 'Tesla T4', 'gpu_count': 2, 'gpu_devices': [{'name': 'Tesla T4', 'memory_total': 16106127360}, {'name': 'Tesla T4', 'memory_total': 16106127360}], 'memory': {'total': 31.357559204101562}}
25
+ 2024-04-12 07:35:56,440 INFO HandlerThread:334 [system_monitor.py:probe():224] Finished collecting system info
26
+ 2024-04-12 07:35:56,440 INFO HandlerThread:334 [system_monitor.py:probe():227] Publishing system info
27
+ 2024-04-12 07:35:56,440 DEBUG HandlerThread:334 [system_info.py:_save_conda():207] Saving list of conda packages installed into the current environment
28
+ 2024-04-12 07:35:57,309 INFO Thread-12 :334 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240412_073555-bw7oy9ix/files/conda-environment.yaml
29
+ 2024-04-12 07:36:11,455 ERROR HandlerThread:334 [system_info.py:_save_conda():221] Error saving conda packages: Command '['conda', 'env', 'export']' timed out after 15 seconds
30
  Traceback (most recent call last):
31
  File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/internal/system/system_info.py", line 214, in _save_conda
32
  subprocess.call(
 
37
  File "/opt/conda/lib/python3.10/subprocess.py", line 1951, in _wait
38
  raise TimeoutExpired(self.args, timeout)
39
  subprocess.TimeoutExpired: Command '['conda', 'env', 'export']' timed out after 15 seconds
40
+ 2024-04-12 07:36:11,456 DEBUG HandlerThread:334 [system_info.py:_save_conda():222] Saving conda packages done
41
+ 2024-04-12 07:36:11,456 INFO HandlerThread:334 [system_monitor.py:probe():229] Finished publishing system info
42
+ 2024-04-12 07:36:11,462 DEBUG HandlerThread:334 [handler.py:handle_request():146] handle_request: status_report
43
+ 2024-04-12 07:36:11,462 DEBUG HandlerThread:334 [handler.py:handle_request():146] handle_request: keepalive
44
+ 2024-04-12 07:36:11,462 DEBUG HandlerThread:334 [handler.py:handle_request():146] handle_request: status_report
45
+ 2024-04-12 07:36:11,462 DEBUG HandlerThread:334 [handler.py:handle_request():146] handle_request: keepalive
46
+ 2024-04-12 07:36:11,462 DEBUG HandlerThread:334 [handler.py:handle_request():146] handle_request: status_report
47
+ 2024-04-12 07:36:11,462 DEBUG HandlerThread:334 [handler.py:handle_request():146] handle_request: keepalive
48
+ 2024-04-12 07:36:11,463 DEBUG SenderThread:334 [sender.py:send():379] send: files
49
+ 2024-04-12 07:36:11,463 INFO SenderThread:334 [sender.py:_save_file():1390] saving file wandb-metadata.json with policy now
50
+ 2024-04-12 07:36:11,664 INFO wandb-upload_0:334 [upload_job.py:push():131] Uploaded file /tmp/tmpjph8qv3dwandb/f1up76ir-wandb-metadata.json
51
+ 2024-04-12 07:36:12,312 INFO Thread-12 :334 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240412_073555-bw7oy9ix/files/wandb-metadata.json
52
+ 2024-04-12 07:36:12,509 DEBUG HandlerThread:334 [handler.py:handle_request():146] handle_request: python_packages
53
+ 2024-04-12 07:36:12,509 DEBUG SenderThread:334 [sender.py:send_request():406] send_request: python_packages
54
+ 2024-04-12 07:36:12,512 DEBUG SenderThread:334 [sender.py:send():379] send: telemetry
55
+ 2024-04-12 07:36:12,523 DEBUG SenderThread:334 [sender.py:send():379] send: config
56
+ 2024-04-12 07:36:12,525 DEBUG SenderThread:334 [sender.py:send():379] send: metric
57
+ 2024-04-12 07:36:12,534 DEBUG HandlerThread:334 [handler.py:handle_request():146] handle_request: stop_status
58
+ 2024-04-12 07:36:12,535 DEBUG SenderThread:334 [sender.py:send():379] send: telemetry
59
+ 2024-04-12 07:36:12,535 DEBUG HandlerThread:334 [handler.py:handle_request():146] handle_request: internal_messages
60
+ 2024-04-12 07:36:12,535 DEBUG SenderThread:334 [sender.py:send():379] send: metric
61
+ 2024-04-12 07:36:12,536 WARNING SenderThread:334 [sender.py:send_metric():1341] Seen metric with glob (shouldn't happen)
62
+ 2024-04-12 07:36:12,536 DEBUG SenderThread:334 [sender.py:send():379] send: telemetry
63
+ 2024-04-12 07:36:12,536 DEBUG SenderThread:334 [sender.py:send_request():406] send_request: stop_status
64
+ 2024-04-12 07:36:13,313 INFO Thread-12 :334 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240412_073555-bw7oy9ix/files/requirements.txt
65
+ 2024-04-12 07:36:13,313 INFO Thread-12 :334 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240412_073555-bw7oy9ix/files/output.log
66
+ 2024-04-12 07:36:15,314 INFO Thread-12 :334 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240412_073555-bw7oy9ix/files/output.log
67
+ 2024-04-12 07:36:17,014 DEBUG HandlerThread:334 [handler.py:handle_request():146] handle_request: status_report
68
+ 2024-04-12 07:36:19,589 DEBUG HandlerThread:334 [handler.py:handle_request():146] handle_request: partial_history
69
+ 2024-04-12 07:36:19,592 DEBUG SenderThread:334 [sender.py:send():379] send: metric
70
+ 2024-04-12 07:36:19,593 DEBUG SenderThread:334 [sender.py:send():379] send: metric
71
+ 2024-04-12 07:36:19,593 DEBUG SenderThread:334 [sender.py:send():379] send: metric
72
+ 2024-04-12 07:36:19,593 DEBUG SenderThread:334 [sender.py:send():379] send: metric
73
+ 2024-04-12 07:36:19,593 DEBUG SenderThread:334 [sender.py:send():379] send: history
74
+ 2024-04-12 07:36:19,593 DEBUG SenderThread:334 [sender.py:send_request():406] send_request: summary_record
75
+ 2024-04-12 07:36:19,595 INFO SenderThread:334 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
76
+ 2024-04-12 07:36:20,316 INFO Thread-12 :334 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240412_073555-bw7oy9ix/files/wandb-summary.json
77
+ 2024-04-12 07:36:21,316 INFO Thread-12 :334 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240412_073555-bw7oy9ix/files/output.log
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
wandb/debug.log CHANGED
@@ -1,35 +1,31 @@
1
- 2024-04-12 07:27:36,627 INFO MainThread:174 [wandb_setup.py:_flush():76] Current SDK version is 0.16.5
2
- 2024-04-12 07:27:36,627 INFO MainThread:174 [wandb_setup.py:_flush():76] Configure stats pid to 174
3
- 2024-04-12 07:27:36,627 INFO MainThread:174 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
4
- 2024-04-12 07:27:36,627 INFO MainThread:174 [wandb_setup.py:_flush():76] Loading settings from /kaggle/working/wandb/settings
5
- 2024-04-12 07:27:36,627 INFO MainThread:174 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
6
- 2024-04-12 07:27:36,627 INFO MainThread:174 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
7
- 2024-04-12 07:27:36,628 INFO MainThread:174 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program': '<python with no main file>'}
8
- 2024-04-12 07:27:36,628 INFO MainThread:174 [wandb_setup.py:_flush():76] Applying login settings: {}
9
- 2024-04-12 07:27:36,628 INFO MainThread:174 [wandb_init.py:_log_setup():527] Logging user logs to /kaggle/working/wandb/run-20240412_072736-gyl3aena/logs/debug.log
10
- 2024-04-12 07:27:36,628 INFO MainThread:174 [wandb_init.py:_log_setup():528] Logging internal logs to /kaggle/working/wandb/run-20240412_072736-gyl3aena/logs/debug-internal.log
11
- 2024-04-12 07:27:36,628 INFO MainThread:174 [wandb_init.py:_jupyter_setup():473] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x7db4ec9acdf0>
12
- 2024-04-12 07:27:36,628 INFO MainThread:174 [wandb_init.py:init():567] calling init triggers
13
- 2024-04-12 07:27:36,628 INFO MainThread:174 [wandb_init.py:init():574] wandb.init called with sweep_config: {}
14
  config: {}
15
- 2024-04-12 07:27:36,628 INFO MainThread:174 [wandb_init.py:init():617] starting backend
16
- 2024-04-12 07:27:36,628 INFO MainThread:174 [wandb_init.py:init():621] setting up manager
17
- 2024-04-12 07:27:36,630 INFO MainThread:174 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
18
- 2024-04-12 07:27:36,631 INFO MainThread:174 [wandb_init.py:init():629] backend started and connected
19
- 2024-04-12 07:27:36,644 INFO MainThread:174 [wandb_run.py:_label_probe_notebook():1299] probe notebook
20
- 2024-04-12 07:27:36,922 INFO MainThread:174 [wandb_init.py:init():721] updated telemetry
21
- 2024-04-12 07:27:36,925 INFO MainThread:174 [wandb_init.py:init():754] communicating run to backend with 90.0 second timeout
22
- 2024-04-12 07:27:37,037 INFO MainThread:174 [wandb_run.py:_on_init():2344] communicating current version
23
- 2024-04-12 07:27:37,121 INFO MainThread:174 [wandb_run.py:_on_init():2353] got version response upgrade_message: "wandb version 0.16.6 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
24
 
25
- 2024-04-12 07:27:37,121 INFO MainThread:174 [wandb_init.py:init():805] starting run threads in backend
26
- 2024-04-12 07:27:53,193 INFO MainThread:174 [wandb_run.py:_console_start():2323] atexit reg
27
- 2024-04-12 07:27:53,194 INFO MainThread:174 [wandb_run.py:_redirect():2178] redirect: wrap_raw
28
- 2024-04-12 07:27:53,194 INFO MainThread:174 [wandb_run.py:_redirect():2243] Wrapping output streams.
29
- 2024-04-12 07:27:53,194 INFO MainThread:174 [wandb_run.py:_redirect():2268] Redirects installed.
30
- 2024-04-12 07:27:53,195 INFO MainThread:174 [wandb_init.py:init():848] run started, returning control to user process
31
- 2024-04-12 07:27:53,201 INFO MainThread:174 [wandb_run.py:_config_callback():1347] config_cb None None {'vocab_size': 250880, 'hidden_size': 1536, 'n_layer': 24, 'n_head': 16, 'layer_norm_epsilon': 1e-05, 'initializer_range': 0.02, 'use_cache': False, 'pretraining_tp': 1, 'apply_residual_connection_post_layernorm': False, 'hidden_dropout': 0.0, 'attention_dropout': 0.0, 'bos_token_id': 1, 'eos_token_id': 2, 'slow_but_exact': False, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'bfloat16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['BloomForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'pad_token_id': 3, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'bigscience/bloomz-1b1', 'transformers_version': '4.39.3', 'attention_softmax_in_fp32': True, 'bias_dropout_fusion': True, 'unk_token_id': 0, 'masked_softmax_fusion': True, 'model_type': 'bloom', 'n_inner': None, 'offset_alibi': 100, 'seq_length': 2048, 'skip_bias_add': True, 'skip_bias_add_qkv': False, 'quantization_config': {'quant_method': 'QuantizationMethod.BITS_AND_BYTES', '_load_in_8bit': False, '_load_in_4bit': True, 'llm_int8_threshold': 6.0, 'llm_int8_skip_modules': None, 'llm_int8_enable_fp32_cpu_offload': False, 'llm_int8_has_fp16_weight': False, 'bnb_4bit_quant_type': 'nf4', 'bnb_4bit_use_double_quant': True, 'bnb_4bit_compute_dtype': 'float16', 'bnb_4bit_quant_storage': 'uint8', 'load_in_4bit': True, 'load_in_8bit': False}, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': False, 'do_predict': False, 'evaluation_strategy': 'no', 'prediction_loss_only': False, 'per_device_train_batch_size': 2, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 3e-05, 'weight_decay': 0.0001, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 0.3, 'num_train_epochs': 5, 'max_steps': 60, 'lr_scheduler_type': 'cosine', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.03, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Apr12_07-27-33_e5a48bec8248', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 20, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 20, 'save_total_limit': 1, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'paged_adamw_8bit', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': 'Femboyuwu2000/bloomz-1b1-vn-chat', 'hub_strategy': 'checkpoint', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': True, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None}
32
- 2024-04-12 07:28:26,759 INFO MainThread:174 [jupyter.py:save_ipynb():373] not saving jupyter notebook
33
- 2024-04-12 07:28:26,759 INFO MainThread:174 [wandb_init.py:_pause_backend():438] pausing backend
34
- 2024-04-12 07:33:46,268 INFO MainThread:174 [wandb_init.py:_resume_backend():443] resuming backend
35
- 2024-04-12 07:33:50,610 INFO MainThread:174 [wandb_run.py:_config_callback():1347] config_cb None None {'vocab_size': 250880, 'hidden_size': 1536, 'n_layer': 24, 'n_head': 16, 'layer_norm_epsilon': 1e-05, 'initializer_range': 0.02, 'use_cache': False, 'pretraining_tp': 1, 'apply_residual_connection_post_layernorm': False, 'hidden_dropout': 0.0, 'attention_dropout': 0.0, 'bos_token_id': 1, 'eos_token_id': 2, 'slow_but_exact': False, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'bfloat16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['BloomForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'pad_token_id': 3, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'bigscience/bloomz-1b1', 'transformers_version': '4.39.3', 'attention_softmax_in_fp32': True, 'bias_dropout_fusion': True, 'unk_token_id': 0, 'masked_softmax_fusion': True, 'model_type': 'bloom', 'n_inner': None, 'offset_alibi': 100, 'seq_length': 2048, 'skip_bias_add': True, 'skip_bias_add_qkv': False, 'quantization_config': {'quant_method': 'QuantizationMethod.BITS_AND_BYTES', '_load_in_8bit': False, '_load_in_4bit': True, 'llm_int8_threshold': 6.0, 'llm_int8_skip_modules': None, 'llm_int8_enable_fp32_cpu_offload': False, 'llm_int8_has_fp16_weight': False, 'bnb_4bit_quant_type': 'nf4', 'bnb_4bit_use_double_quant': True, 'bnb_4bit_compute_dtype': 'float16', 'bnb_4bit_quant_storage': 'uint8', 'load_in_4bit': True, 'load_in_8bit': False}, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': False, 'do_predict': False, 'evaluation_strategy': 'no', 'prediction_loss_only': False, 'per_device_train_batch_size': 2, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 3e-05, 'weight_decay': 0.0001, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 0.3, 'num_train_epochs': 5, 'max_steps': 60, 'lr_scheduler_type': 'cosine', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.03, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Apr12_07-33-49_e5a48bec8248', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 20, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 20, 'save_total_limit': 1, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'paged_adamw_8bit', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': 'Femboyuwu2000/bloomz-1b1-vn-chat', 'hub_strategy': 'checkpoint', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': True, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None}
 
1
+ 2024-04-12 07:35:55,744 INFO MainThread:289 [wandb_setup.py:_flush():76] Current SDK version is 0.16.5
2
+ 2024-04-12 07:35:55,744 INFO MainThread:289 [wandb_setup.py:_flush():76] Configure stats pid to 289
3
+ 2024-04-12 07:35:55,744 INFO MainThread:289 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
4
+ 2024-04-12 07:35:55,744 INFO MainThread:289 [wandb_setup.py:_flush():76] Loading settings from /kaggle/working/wandb/settings
5
+ 2024-04-12 07:35:55,744 INFO MainThread:289 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
6
+ 2024-04-12 07:35:55,744 INFO MainThread:289 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
7
+ 2024-04-12 07:35:55,744 INFO MainThread:289 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program': '<python with no main file>'}
8
+ 2024-04-12 07:35:55,744 INFO MainThread:289 [wandb_setup.py:_flush():76] Applying login settings: {}
9
+ 2024-04-12 07:35:55,744 INFO MainThread:289 [wandb_init.py:_log_setup():527] Logging user logs to /kaggle/working/wandb/run-20240412_073555-bw7oy9ix/logs/debug.log
10
+ 2024-04-12 07:35:55,744 INFO MainThread:289 [wandb_init.py:_log_setup():528] Logging internal logs to /kaggle/working/wandb/run-20240412_073555-bw7oy9ix/logs/debug-internal.log
11
+ 2024-04-12 07:35:55,744 INFO MainThread:289 [wandb_init.py:_jupyter_setup():473] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x7c2cd1fccd60>
12
+ 2024-04-12 07:35:55,745 INFO MainThread:289 [wandb_init.py:init():567] calling init triggers
13
+ 2024-04-12 07:35:55,745 INFO MainThread:289 [wandb_init.py:init():574] wandb.init called with sweep_config: {}
14
  config: {}
15
+ 2024-04-12 07:35:55,745 INFO MainThread:289 [wandb_init.py:init():617] starting backend
16
+ 2024-04-12 07:35:55,745 INFO MainThread:289 [wandb_init.py:init():621] setting up manager
17
+ 2024-04-12 07:35:55,747 INFO MainThread:289 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
18
+ 2024-04-12 07:35:55,748 INFO MainThread:289 [wandb_init.py:init():629] backend started and connected
19
+ 2024-04-12 07:35:55,761 INFO MainThread:289 [wandb_run.py:_label_probe_notebook():1299] probe notebook
20
+ 2024-04-12 07:35:56,159 INFO MainThread:289 [wandb_init.py:init():721] updated telemetry
21
+ 2024-04-12 07:35:56,162 INFO MainThread:289 [wandb_init.py:init():754] communicating run to backend with 90.0 second timeout
22
+ 2024-04-12 07:35:56,316 INFO MainThread:289 [wandb_run.py:_on_init():2344] communicating current version
23
+ 2024-04-12 07:35:56,402 INFO MainThread:289 [wandb_run.py:_on_init():2353] got version response upgrade_message: "wandb version 0.16.6 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
24
 
25
+ 2024-04-12 07:35:56,403 INFO MainThread:289 [wandb_init.py:init():805] starting run threads in backend
26
+ 2024-04-12 07:36:12,510 INFO MainThread:289 [wandb_run.py:_console_start():2323] atexit reg
27
+ 2024-04-12 07:36:12,510 INFO MainThread:289 [wandb_run.py:_redirect():2178] redirect: wrap_raw
28
+ 2024-04-12 07:36:12,511 INFO MainThread:289 [wandb_run.py:_redirect():2243] Wrapping output streams.
29
+ 2024-04-12 07:36:12,511 INFO MainThread:289 [wandb_run.py:_redirect():2268] Redirects installed.
30
+ 2024-04-12 07:36:12,512 INFO MainThread:289 [wandb_init.py:init():848] run started, returning control to user process
31
+ 2024-04-12 07:36:12,518 INFO MainThread:289 [wandb_run.py:_config_callback():1347] config_cb None None {'vocab_size': 250880, 'hidden_size': 1536, 'n_layer': 24, 'n_head': 16, 'layer_norm_epsilon': 1e-05, 'initializer_range': 0.02, 'use_cache': False, 'pretraining_tp': 1, 'apply_residual_connection_post_layernorm': False, 'hidden_dropout': 0.0, 'attention_dropout': 0.0, 'bos_token_id': 1, 'eos_token_id': 2, 'slow_but_exact': False, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'bfloat16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['BloomForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'pad_token_id': 3, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'bigscience/bloomz-1b1', 'transformers_version': '4.39.3', 'attention_softmax_in_fp32': True, 'bias_dropout_fusion': True, 'unk_token_id': 0, 'masked_softmax_fusion': True, 'model_type': 'bloom', 'n_inner': None, 'offset_alibi': 100, 'seq_length': 2048, 'skip_bias_add': True, 'skip_bias_add_qkv': False, 'quantization_config': {'quant_method': 'QuantizationMethod.BITS_AND_BYTES', '_load_in_8bit': False, '_load_in_4bit': True, 'llm_int8_threshold': 6.0, 'llm_int8_skip_modules': None, 'llm_int8_enable_fp32_cpu_offload': False, 'llm_int8_has_fp16_weight': False, 'bnb_4bit_quant_type': 'nf4', 'bnb_4bit_use_double_quant': True, 'bnb_4bit_compute_dtype': 'float16', 'bnb_4bit_quant_storage': 'uint8', 'load_in_4bit': True, 'load_in_8bit': False}, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': False, 'do_predict': False, 'evaluation_strategy': 'no', 'prediction_loss_only': False, 'per_device_train_batch_size': 2, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 3e-05, 'weight_decay': 0.0001, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 0.3, 'num_train_epochs': 5, 'max_steps': 20000, 'lr_scheduler_type': 'cosine', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.03, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Apr12_07-35-52_e5a48bec8248', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 20, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 20, 'save_total_limit': 1, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'paged_adamw_8bit', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': 'Femboyuwu2000/bloomz-1b1-vn-chat', 'hub_strategy': 'checkpoint', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': True, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None}
 
 
 
 
wandb/run-20240412_072736-gyl3aena/files/output.log CHANGED
@@ -13,3 +13,4 @@ dataloader_config = DataLoaderConfiguration(dispatch_batches=None, split_batches
13
  /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
14
  warnings.warn(
15
  /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
 
 
13
  /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
14
  warnings.warn(
15
  /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
16
+ warnings.warn(
wandb/run-20240412_072736-gyl3aena/files/wandb-summary.json CHANGED
@@ -1 +1 @@
1
- {"train/loss": 4.8834, "train/grad_norm": 86.45769500732422, "train/learning_rate": 0.0, "train/epoch": 0.0, "train/global_step": 60, "_timestamp": 1712907251.181722, "_runtime": 394.54964303970337, "_step": 7, "train_runtime": 20.5872, "train_samples_per_second": 5.829, "train_steps_per_second": 2.914, "total_flos": 19812939669504.0, "train_loss": 4.715130106608073}
 
1
+ {"train/loss": 4.8834, "train/grad_norm": 86.45769500732422, "train/learning_rate": 0.0, "train/epoch": 0.0, "train/global_step": 60, "_timestamp": 1712907251.181722, "_runtime": 394.54964303970337, "_step": 7, "train_runtime": 20.5872, "train_samples_per_second": 5.829, "train_steps_per_second": 2.914, "total_flos": 19812939669504.0, "train_loss": 4.715130106608073, "_wandb": {"runtime": 86}}
wandb/run-20240412_072736-gyl3aena/logs/debug-internal.log CHANGED
@@ -310,3 +310,124 @@ subprocess.TimeoutExpired: Command '['conda', 'env', 'export']' timed out after
310
  2024-04-12 07:34:12,187 INFO Thread-12 :216 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240412_072736-gyl3aena/files/wandb-summary.json
311
  2024-04-12 07:34:14,191 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
312
  2024-04-12 07:34:19,191 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
310
  2024-04-12 07:34:12,187 INFO Thread-12 :216 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240412_072736-gyl3aena/files/wandb-summary.json
311
  2024-04-12 07:34:14,191 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
312
  2024-04-12 07:34:19,191 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
313
+ 2024-04-12 07:34:22,622 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: pause
314
+ 2024-04-12 07:34:22,622 INFO HandlerThread:216 [handler.py:handle_request_pause():708] stopping system metrics thread
315
+ 2024-04-12 07:34:22,622 INFO HandlerThread:216 [system_monitor.py:finish():203] Stopping system monitor
316
+ 2024-04-12 07:34:22,623 DEBUG SystemMonitor:216 [system_monitor.py:_start():172] Starting system metrics aggregation loop
317
+ 2024-04-12 07:34:22,623 INFO HandlerThread:216 [interfaces.py:finish():202] Joined cpu monitor
318
+ 2024-04-12 07:34:22,623 DEBUG SystemMonitor:216 [system_monitor.py:_start():179] Finished system metrics aggregation loop
319
+ 2024-04-12 07:34:22,624 INFO HandlerThread:216 [interfaces.py:finish():202] Joined disk monitor
320
+ 2024-04-12 07:34:22,624 DEBUG SystemMonitor:216 [system_monitor.py:_start():183] Publishing last batch of metrics
321
+ 2024-04-12 07:34:22,634 INFO HandlerThread:216 [interfaces.py:finish():202] Joined gpu monitor
322
+ 2024-04-12 07:34:22,635 INFO HandlerThread:216 [interfaces.py:finish():202] Joined memory monitor
323
+ 2024-04-12 07:34:22,635 INFO HandlerThread:216 [interfaces.py:finish():202] Joined network monitor
324
+ 2024-04-12 07:34:22,635 DEBUG SenderThread:216 [sender.py:send():379] send: stats
325
+ 2024-04-12 07:34:23,197 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: stop_status
326
+ 2024-04-12 07:34:23,198 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: internal_messages
327
+ 2024-04-12 07:34:23,198 DEBUG SenderThread:216 [sender.py:send_request():406] send_request: stop_status
328
+ 2024-04-12 07:34:24,309 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
329
+ 2024-04-12 07:34:29,310 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
330
+ 2024-04-12 07:34:34,311 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
331
+ 2024-04-12 07:34:38,197 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: stop_status
332
+ 2024-04-12 07:34:38,198 DEBUG SenderThread:216 [sender.py:send_request():406] send_request: stop_status
333
+ 2024-04-12 07:34:38,238 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: internal_messages
334
+ 2024-04-12 07:34:40,296 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
335
+ 2024-04-12 07:34:45,297 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
336
+ 2024-04-12 07:34:50,298 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
337
+ 2024-04-12 07:34:53,198 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: stop_status
338
+ 2024-04-12 07:34:53,198 DEBUG SenderThread:216 [sender.py:send_request():406] send_request: stop_status
339
+ 2024-04-12 07:34:53,238 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: internal_messages
340
+ 2024-04-12 07:34:56,282 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
341
+ 2024-04-12 07:35:01,283 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
342
+ 2024-04-12 07:35:06,284 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
343
+ 2024-04-12 07:35:08,198 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: stop_status
344
+ 2024-04-12 07:35:08,198 DEBUG SenderThread:216 [sender.py:send_request():406] send_request: stop_status
345
+ 2024-04-12 07:35:08,238 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: internal_messages
346
+ 2024-04-12 07:35:11,071 DEBUG SenderThread:216 [sender.py:send():379] send: exit
347
+ 2024-04-12 07:35:11,071 INFO SenderThread:216 [sender.py:send_exit():586] handling exit code: 0
348
+ 2024-04-12 07:35:11,071 INFO SenderThread:216 [sender.py:send_exit():588] handling runtime: 86
349
+ 2024-04-12 07:35:11,073 INFO SenderThread:216 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
350
+ 2024-04-12 07:35:11,074 INFO SenderThread:216 [sender.py:send_exit():594] send defer
351
+ 2024-04-12 07:35:11,074 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: defer
352
+ 2024-04-12 07:35:11,074 INFO HandlerThread:216 [handler.py:handle_request_defer():172] handle defer: 0
353
+ 2024-04-12 07:35:11,074 DEBUG SenderThread:216 [sender.py:send_request():406] send_request: defer
354
+ 2024-04-12 07:35:11,074 INFO SenderThread:216 [sender.py:send_request_defer():610] handle sender defer: 0
355
+ 2024-04-12 07:35:11,074 INFO SenderThread:216 [sender.py:transition_state():614] send defer: 1
356
+ 2024-04-12 07:35:11,074 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: defer
357
+ 2024-04-12 07:35:11,074 INFO HandlerThread:216 [handler.py:handle_request_defer():172] handle defer: 1
358
+ 2024-04-12 07:35:11,075 DEBUG SenderThread:216 [sender.py:send_request():406] send_request: defer
359
+ 2024-04-12 07:35:11,075 INFO SenderThread:216 [sender.py:send_request_defer():610] handle sender defer: 1
360
+ 2024-04-12 07:35:11,075 INFO SenderThread:216 [sender.py:transition_state():614] send defer: 2
361
+ 2024-04-12 07:35:11,075 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: defer
362
+ 2024-04-12 07:35:11,075 INFO HandlerThread:216 [handler.py:handle_request_defer():172] handle defer: 2
363
+ 2024-04-12 07:35:11,075 DEBUG SenderThread:216 [sender.py:send_request():406] send_request: defer
364
+ 2024-04-12 07:35:11,075 INFO SenderThread:216 [sender.py:send_request_defer():610] handle sender defer: 2
365
+ 2024-04-12 07:35:11,075 INFO SenderThread:216 [sender.py:transition_state():614] send defer: 3
366
+ 2024-04-12 07:35:11,075 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: defer
367
+ 2024-04-12 07:35:11,075 INFO HandlerThread:216 [handler.py:handle_request_defer():172] handle defer: 3
368
+ 2024-04-12 07:35:11,076 DEBUG SenderThread:216 [sender.py:send_request():406] send_request: defer
369
+ 2024-04-12 07:35:11,076 INFO SenderThread:216 [sender.py:send_request_defer():610] handle sender defer: 3
370
+ 2024-04-12 07:35:11,076 INFO SenderThread:216 [sender.py:transition_state():614] send defer: 4
371
+ 2024-04-12 07:35:11,076 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: defer
372
+ 2024-04-12 07:35:11,076 INFO HandlerThread:216 [handler.py:handle_request_defer():172] handle defer: 4
373
+ 2024-04-12 07:35:11,076 DEBUG SenderThread:216 [sender.py:send_request():406] send_request: defer
374
+ 2024-04-12 07:35:11,076 INFO SenderThread:216 [sender.py:send_request_defer():610] handle sender defer: 4
375
+ 2024-04-12 07:35:11,076 INFO SenderThread:216 [sender.py:transition_state():614] send defer: 5
376
+ 2024-04-12 07:35:11,076 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: defer
377
+ 2024-04-12 07:35:11,077 INFO HandlerThread:216 [handler.py:handle_request_defer():172] handle defer: 5
378
+ 2024-04-12 07:35:11,077 DEBUG SenderThread:216 [sender.py:send():379] send: summary
379
+ 2024-04-12 07:35:11,077 INFO SenderThread:216 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
380
+ 2024-04-12 07:35:11,078 DEBUG SenderThread:216 [sender.py:send_request():406] send_request: defer
381
+ 2024-04-12 07:35:11,078 INFO SenderThread:216 [sender.py:send_request_defer():610] handle sender defer: 5
382
+ 2024-04-12 07:35:11,078 INFO SenderThread:216 [sender.py:transition_state():614] send defer: 6
383
+ 2024-04-12 07:35:11,078 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: defer
384
+ 2024-04-12 07:35:11,078 INFO HandlerThread:216 [handler.py:handle_request_defer():172] handle defer: 6
385
+ 2024-04-12 07:35:11,078 DEBUG SenderThread:216 [sender.py:send_request():406] send_request: defer
386
+ 2024-04-12 07:35:11,078 INFO SenderThread:216 [sender.py:send_request_defer():610] handle sender defer: 6
387
+ 2024-04-12 07:35:11,078 INFO SenderThread:216 [sender.py:transition_state():614] send defer: 7
388
+ 2024-04-12 07:35:11,078 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: status_report
389
+ 2024-04-12 07:35:11,079 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: defer
390
+ 2024-04-12 07:35:11,079 INFO HandlerThread:216 [handler.py:handle_request_defer():172] handle defer: 7
391
+ 2024-04-12 07:35:11,079 DEBUG SenderThread:216 [sender.py:send_request():406] send_request: defer
392
+ 2024-04-12 07:35:11,079 INFO SenderThread:216 [sender.py:send_request_defer():610] handle sender defer: 7
393
+ 2024-04-12 07:35:11,209 INFO Thread-12 :216 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240412_072736-gyl3aena/files/wandb-summary.json
394
+ 2024-04-12 07:35:12,071 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: poll_exit
395
+ 2024-04-12 07:35:13,826 INFO SenderThread:216 [sender.py:transition_state():614] send defer: 8
396
+ 2024-04-12 07:35:13,826 DEBUG SenderThread:216 [sender.py:send_request():406] send_request: poll_exit
397
+ 2024-04-12 07:35:13,827 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: defer
398
+ 2024-04-12 07:35:13,827 INFO HandlerThread:216 [handler.py:handle_request_defer():172] handle defer: 8
399
+ 2024-04-12 07:35:13,827 DEBUG SenderThread:216 [sender.py:send_request():406] send_request: defer
400
+ 2024-04-12 07:35:13,827 INFO SenderThread:216 [sender.py:send_request_defer():610] handle sender defer: 8
401
+ 2024-04-12 07:35:13,827 INFO SenderThread:216 [job_builder.py:build():318] Attempting to build job artifact
402
+ 2024-04-12 07:35:13,830 INFO SenderThread:216 [job_builder.py:_get_source_type():466] no source found
403
+ 2024-04-12 07:35:13,830 INFO SenderThread:216 [sender.py:transition_state():614] send defer: 9
404
+ 2024-04-12 07:35:13,830 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: defer
405
+ 2024-04-12 07:35:13,830 INFO HandlerThread:216 [handler.py:handle_request_defer():172] handle defer: 9
406
+ 2024-04-12 07:35:13,830 DEBUG SenderThread:216 [sender.py:send_request():406] send_request: defer
407
+ 2024-04-12 07:35:13,830 INFO SenderThread:216 [sender.py:send_request_defer():610] handle sender defer: 9
408
+ 2024-04-12 07:35:13,830 INFO SenderThread:216 [dir_watcher.py:finish():358] shutting down directory watcher
409
+ 2024-04-12 07:35:14,073 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: poll_exit
410
+ 2024-04-12 07:35:14,124 WARNING StreamThr :216 [internal.py:is_dead():414] Internal process exiting, parent pid 174 disappeared
411
+ 2024-04-12 07:35:14,125 ERROR StreamThr :216 [internal.py:wandb_internal():152] Internal process shutdown.
412
+ 2024-04-12 07:35:14,210 INFO SenderThread:216 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240412_072736-gyl3aena/files/output.log
413
+ 2024-04-12 07:35:14,211 INFO SenderThread:216 [dir_watcher.py:finish():388] scan: /kaggle/working/wandb/run-20240412_072736-gyl3aena/files
414
+ 2024-04-12 07:35:14,211 INFO SenderThread:216 [dir_watcher.py:finish():402] scan save: /kaggle/working/wandb/run-20240412_072736-gyl3aena/files/conda-environment.yaml conda-environment.yaml
415
+ 2024-04-12 07:35:14,212 INFO SenderThread:216 [dir_watcher.py:finish():402] scan save: /kaggle/working/wandb/run-20240412_072736-gyl3aena/files/requirements.txt requirements.txt
416
+ 2024-04-12 07:35:14,212 INFO SenderThread:216 [dir_watcher.py:finish():402] scan save: /kaggle/working/wandb/run-20240412_072736-gyl3aena/files/output.log output.log
417
+ 2024-04-12 07:35:14,213 INFO SenderThread:216 [dir_watcher.py:finish():402] scan save: /kaggle/working/wandb/run-20240412_072736-gyl3aena/files/wandb-summary.json wandb-summary.json
418
+ 2024-04-12 07:35:14,219 INFO SenderThread:216 [dir_watcher.py:finish():402] scan save: /kaggle/working/wandb/run-20240412_072736-gyl3aena/files/config.yaml config.yaml
419
+ 2024-04-12 07:35:14,225 INFO SenderThread:216 [dir_watcher.py:finish():402] scan save: /kaggle/working/wandb/run-20240412_072736-gyl3aena/files/wandb-metadata.json wandb-metadata.json
420
+ 2024-04-12 07:35:14,225 INFO SenderThread:216 [sender.py:transition_state():614] send defer: 10
421
+ 2024-04-12 07:35:14,226 INFO SenderThread:216 [sender.py:finish():1546] shutting down sender
422
+ 2024-04-12 07:35:14,226 INFO SenderThread:216 [file_pusher.py:finish():172] shutting down file pusher
423
+ 2024-04-12 07:35:14,226 INFO SenderThread:216 [file_pusher.py:join():178] waiting for file pusher
424
+ 2024-04-12 07:35:14,226 DEBUG HandlerThread:216 [handler.py:handle_request():146] handle_request: defer
425
+ 2024-04-12 07:35:14,226 INFO HandlerThread:216 [handler.py:handle_request_defer():172] handle defer: 10
426
+ 2024-04-12 07:35:14,226 INFO HandlerThread:216 [handler.py:finish():866] shutting down handler
427
+ 2024-04-12 07:35:14,227 INFO WriterThread:216 [datastore.py:close():296] close: /kaggle/working/wandb/run-20240412_072736-gyl3aena/run-gyl3aena.wandb
428
+ 2024-04-12 07:35:14,441 INFO wandb-upload_1:216 [upload_job.py:push():131] Uploaded file /kaggle/working/wandb/run-20240412_072736-gyl3aena/files/output.log
429
+ 2024-04-12 07:35:14,519 INFO wandb-upload_2:216 [upload_job.py:push():131] Uploaded file /kaggle/working/wandb/run-20240412_072736-gyl3aena/files/wandb-summary.json
430
+ 2024-04-12 07:35:14,520 INFO wandb-upload_3:216 [upload_job.py:push():131] Uploaded file /kaggle/working/wandb/run-20240412_072736-gyl3aena/files/config.yaml
431
+ 2024-04-12 07:35:14,521 INFO wandb-upload_0:216 [upload_job.py:push():131] Uploaded file /kaggle/working/wandb/run-20240412_072736-gyl3aena/files/requirements.txt
432
+ 2024-04-12 07:35:14,727 INFO SenderThread:216 [file_stream.py:finish():614] file stream finish called
433
+ 2024-04-12 07:35:14,790 INFO SenderThread:216 [file_stream.py:finish():618] file stream finish is done
wandb/run-20240412_072736-gyl3aena/logs/debug.log CHANGED
@@ -33,3 +33,5 @@ config: {}
33
  2024-04-12 07:28:26,759 INFO MainThread:174 [wandb_init.py:_pause_backend():438] pausing backend
34
  2024-04-12 07:33:46,268 INFO MainThread:174 [wandb_init.py:_resume_backend():443] resuming backend
35
  2024-04-12 07:33:50,610 INFO MainThread:174 [wandb_run.py:_config_callback():1347] config_cb None None {'vocab_size': 250880, 'hidden_size': 1536, 'n_layer': 24, 'n_head': 16, 'layer_norm_epsilon': 1e-05, 'initializer_range': 0.02, 'use_cache': False, 'pretraining_tp': 1, 'apply_residual_connection_post_layernorm': False, 'hidden_dropout': 0.0, 'attention_dropout': 0.0, 'bos_token_id': 1, 'eos_token_id': 2, 'slow_but_exact': False, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'bfloat16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['BloomForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'pad_token_id': 3, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'bigscience/bloomz-1b1', 'transformers_version': '4.39.3', 'attention_softmax_in_fp32': True, 'bias_dropout_fusion': True, 'unk_token_id': 0, 'masked_softmax_fusion': True, 'model_type': 'bloom', 'n_inner': None, 'offset_alibi': 100, 'seq_length': 2048, 'skip_bias_add': True, 'skip_bias_add_qkv': False, 'quantization_config': {'quant_method': 'QuantizationMethod.BITS_AND_BYTES', '_load_in_8bit': False, '_load_in_4bit': True, 'llm_int8_threshold': 6.0, 'llm_int8_skip_modules': None, 'llm_int8_enable_fp32_cpu_offload': False, 'llm_int8_has_fp16_weight': False, 'bnb_4bit_quant_type': 'nf4', 'bnb_4bit_use_double_quant': True, 'bnb_4bit_compute_dtype': 'float16', 'bnb_4bit_quant_storage': 'uint8', 'load_in_4bit': True, 'load_in_8bit': False}, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': False, 'do_predict': False, 'evaluation_strategy': 'no', 'prediction_loss_only': False, 'per_device_train_batch_size': 2, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 3e-05, 'weight_decay': 0.0001, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 0.3, 'num_train_epochs': 5, 'max_steps': 60, 'lr_scheduler_type': 'cosine', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.03, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Apr12_07-33-49_e5a48bec8248', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 20, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 20, 'save_total_limit': 1, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'paged_adamw_8bit', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': 'Femboyuwu2000/bloomz-1b1-vn-chat', 'hub_strategy': 'checkpoint', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': True, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None}
 
 
 
33
  2024-04-12 07:28:26,759 INFO MainThread:174 [wandb_init.py:_pause_backend():438] pausing backend
34
  2024-04-12 07:33:46,268 INFO MainThread:174 [wandb_init.py:_resume_backend():443] resuming backend
35
  2024-04-12 07:33:50,610 INFO MainThread:174 [wandb_run.py:_config_callback():1347] config_cb None None {'vocab_size': 250880, 'hidden_size': 1536, 'n_layer': 24, 'n_head': 16, 'layer_norm_epsilon': 1e-05, 'initializer_range': 0.02, 'use_cache': False, 'pretraining_tp': 1, 'apply_residual_connection_post_layernorm': False, 'hidden_dropout': 0.0, 'attention_dropout': 0.0, 'bos_token_id': 1, 'eos_token_id': 2, 'slow_but_exact': False, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'bfloat16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['BloomForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'pad_token_id': 3, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'bigscience/bloomz-1b1', 'transformers_version': '4.39.3', 'attention_softmax_in_fp32': True, 'bias_dropout_fusion': True, 'unk_token_id': 0, 'masked_softmax_fusion': True, 'model_type': 'bloom', 'n_inner': None, 'offset_alibi': 100, 'seq_length': 2048, 'skip_bias_add': True, 'skip_bias_add_qkv': False, 'quantization_config': {'quant_method': 'QuantizationMethod.BITS_AND_BYTES', '_load_in_8bit': False, '_load_in_4bit': True, 'llm_int8_threshold': 6.0, 'llm_int8_skip_modules': None, 'llm_int8_enable_fp32_cpu_offload': False, 'llm_int8_has_fp16_weight': False, 'bnb_4bit_quant_type': 'nf4', 'bnb_4bit_use_double_quant': True, 'bnb_4bit_compute_dtype': 'float16', 'bnb_4bit_quant_storage': 'uint8', 'load_in_4bit': True, 'load_in_8bit': False}, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': False, 'do_predict': False, 'evaluation_strategy': 'no', 'prediction_loss_only': False, 'per_device_train_batch_size': 2, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 3e-05, 'weight_decay': 0.0001, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 0.3, 'num_train_epochs': 5, 'max_steps': 60, 'lr_scheduler_type': 'cosine', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.03, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Apr12_07-33-49_e5a48bec8248', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 20, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 20, 'save_total_limit': 1, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'paged_adamw_8bit', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': 'Femboyuwu2000/bloomz-1b1-vn-chat', 'hub_strategy': 'checkpoint', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': True, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None}
36
+ 2024-04-12 07:34:22,621 INFO MainThread:174 [jupyter.py:save_ipynb():373] not saving jupyter notebook
37
+ 2024-04-12 07:34:22,621 INFO MainThread:174 [wandb_init.py:_pause_backend():438] pausing backend
wandb/run-20240412_072736-gyl3aena/run-gyl3aena.wandb CHANGED
Binary files a/wandb/run-20240412_072736-gyl3aena/run-gyl3aena.wandb and b/wandb/run-20240412_072736-gyl3aena/run-gyl3aena.wandb differ
 
wandb/run-20240412_073555-bw7oy9ix/files/conda-environment.yaml ADDED
File without changes
wandb/run-20240412_073555-bw7oy9ix/files/config.yaml ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wandb_version: 1
2
+
3
+ _wandb:
4
+ desc: null
5
+ value:
6
+ python_version: 3.10.13
7
+ cli_version: 0.16.5
8
+ framework: huggingface
9
+ huggingface_version: 4.39.3
10
+ is_jupyter_run: true
11
+ is_kaggle_kernel: true
12
+ start_time: 1712907355.0
13
+ t:
14
+ 1:
15
+ - 1
16
+ - 2
17
+ - 3
18
+ - 5
19
+ - 11
20
+ - 12
21
+ - 49
22
+ - 51
23
+ - 53
24
+ - 55
25
+ - 71
26
+ - 84
27
+ - 98
28
+ - 105
29
+ 3:
30
+ - 23
31
+ 4: 3.10.13
32
+ 5: 0.16.5
33
+ 6: 4.39.3
34
+ 8:
35
+ - 1
36
+ - 2
37
+ - 5
38
+ 13: linux-x86_64
wandb/run-20240412_073555-bw7oy9ix/files/output.log ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
2
+ warnings.warn(
3
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
wandb/run-20240412_073555-bw7oy9ix/files/requirements.txt ADDED
@@ -0,0 +1,864 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Babel==2.14.0
2
+ Boruta==0.3
3
+ Brotli==1.0.9
4
+ CVXcanon==0.1.2
5
+ Cartopy==0.22.0
6
+ Cython==3.0.8
7
+ Deprecated==1.2.14
8
+ Farama-Notifications==0.0.4
9
+ Flask==3.0.2
10
+ Geohash==1.0
11
+ GitPython==3.1.41
12
+ ImageHash==4.3.1
13
+ Janome==0.5.0
14
+ Jinja2==3.1.2
15
+ LunarCalendar==0.0.9
16
+ Mako==1.3.2
17
+ Markdown==3.5.2
18
+ MarkupSafe==2.1.3
19
+ MarkupSafe==2.1.5
20
+ Pillow==9.5.0
21
+ PuLP==2.8.0
22
+ PyArabic==0.6.15
23
+ PyJWT==2.8.0
24
+ PyMeeus==0.5.12
25
+ PySocks==1.7.1
26
+ PyUpSet==0.1.1.post7
27
+ PyWavelets==1.5.0
28
+ PyYAML==6.0.1
29
+ Pygments==2.17.2
30
+ Pympler==1.0.1
31
+ QtPy==2.4.1
32
+ Rtree==1.2.0
33
+ SQLAlchemy==2.0.25
34
+ SecretStorage==3.3.3
35
+ Send2Trash==1.8.2
36
+ Shapely==1.8.5.post1
37
+ Shimmy==1.3.0
38
+ SimpleITK==2.3.1
39
+ TPOT==0.12.1
40
+ Theano-PyMC==1.1.2
41
+ Theano==1.0.5
42
+ Wand==0.6.13
43
+ Werkzeug==3.0.2
44
+ absl-py==1.4.0
45
+ accelerate==0.28.0
46
+ access==1.1.9
47
+ affine==2.4.0
48
+ aiobotocore==2.12.2
49
+ aiofiles==22.1.0
50
+ aiohttp-cors==0.7.0
51
+ aiohttp==3.9.1
52
+ aioitertools==0.11.0
53
+ aiorwlock==1.3.0
54
+ aiosignal==1.3.1
55
+ aiosqlite==0.19.0
56
+ albumentations==1.4.0
57
+ alembic==1.13.1
58
+ altair==5.3.0
59
+ annotated-types==0.6.0
60
+ annoy==1.17.3
61
+ anyio==4.2.0
62
+ apache-beam==2.46.0
63
+ aplus==0.11.0
64
+ appdirs==1.4.4
65
+ archspec==0.2.3
66
+ argon2-cffi-bindings==21.2.0
67
+ argon2-cffi==23.1.0
68
+ array-record==0.5.0
69
+ arrow==1.3.0
70
+ arviz==0.17.1
71
+ astroid==3.1.0
72
+ astropy-iers-data==0.2024.4.1.0.33.14
73
+ astropy==6.0.1
74
+ asttokens==2.4.1
75
+ astunparse==1.6.3
76
+ async-lru==2.0.4
77
+ async-timeout==4.0.3
78
+ attrs==23.2.0
79
+ audioread==3.0.1
80
+ autopep8==2.0.4
81
+ backoff==2.2.1
82
+ bayesian-optimization==1.4.3
83
+ beatrix_jupyterlab==2023.128.151533
84
+ beautifulsoup4==4.12.2
85
+ bitsandbytes==0.43.1
86
+ blake3==0.2.1
87
+ bleach==6.1.0
88
+ blessed==1.20.0
89
+ blinker==1.7.0
90
+ blis==0.7.10
91
+ blosc2==2.6.0
92
+ bokeh==3.3.4
93
+ boltons==23.1.1
94
+ boto3==1.26.100
95
+ botocore==1.34.51
96
+ bq_helper==0.4.1
97
+ bqplot==0.12.43
98
+ branca==0.7.1
99
+ brewer2mpl==1.4.1
100
+ brotlipy==0.7.0
101
+ cached-property==1.5.2
102
+ cachetools==4.2.4
103
+ cachetools==5.3.2
104
+ catalogue==2.0.10
105
+ catalyst==22.4
106
+ catboost==1.2.3
107
+ category-encoders==2.6.3
108
+ certifi==2024.2.2
109
+ cesium==0.12.1
110
+ cffi==1.16.0
111
+ charset-normalizer==3.3.2
112
+ chex==0.1.86
113
+ cleverhans==4.0.0
114
+ click-plugins==1.1.1
115
+ click==8.1.7
116
+ cligj==0.7.2
117
+ cloud-tpu-client==0.10
118
+ cloud-tpu-profiler==2.4.0
119
+ cloudpathlib==0.16.0
120
+ cloudpickle==2.2.1
121
+ cloudpickle==3.0.0
122
+ cmdstanpy==1.2.2
123
+ colorama==0.4.6
124
+ colorcet==3.1.0
125
+ colorful==0.5.6
126
+ colorlog==6.8.2
127
+ colorlover==0.3.0
128
+ comm==0.2.1
129
+ conda-libmamba-solver==23.7.0
130
+ conda-package-handling==2.2.0
131
+ conda==23.7.4
132
+ conda_package_streaming==0.9.0
133
+ confection==0.1.4
134
+ contextily==1.6.0
135
+ contourpy==1.2.0
136
+ convertdate==2.4.0
137
+ crcmod==1.7
138
+ cryptography==41.0.7
139
+ cuda-python==12.4.0
140
+ cudf==23.8.0
141
+ cufflinks==0.17.3
142
+ cuml==23.8.0
143
+ cupy==13.0.0
144
+ cycler==0.12.1
145
+ cymem==2.0.8
146
+ cytoolz==0.12.3
147
+ daal4py==2024.2.0
148
+ daal==2024.2.0
149
+ dacite==1.8.1
150
+ dask-cuda==23.8.0
151
+ dask-cudf==23.8.0
152
+ dask-expr==1.0.9
153
+ dask==2024.4.0
154
+ dataclasses-json==0.6.4
155
+ dataproc_jupyter_plugin==0.1.66
156
+ datasets==2.16.0
157
+ datashader==0.16.0
158
+ datatile==1.0.3
159
+ db-dtypes==1.2.0
160
+ deap==1.4.1
161
+ debugpy==1.8.0
162
+ decorator==5.1.1
163
+ deepdiff==6.7.1
164
+ defusedxml==0.7.1
165
+ deprecation==2.1.0
166
+ descartes==1.1.0
167
+ dill==0.3.7
168
+ dipy==1.9.0
169
+ distlib==0.3.8
170
+ distributed==2023.7.1
171
+ distro==1.9.0
172
+ dm-tree==0.1.8
173
+ docker-pycreds==0.4.0
174
+ docker==7.0.0
175
+ docopt==0.6.2
176
+ docstring-parser==0.15
177
+ docstring-to-markdown==0.15
178
+ docutils==0.20.1
179
+ earthengine-api==0.1.395
180
+ easydict==1.13
181
+ easyocr==1.7.1
182
+ ecos==2.0.13
183
+ eli5==0.13.0
184
+ emoji==2.11.0
185
+ en-core-web-lg==3.7.1
186
+ en-core-web-sm==3.7.1
187
+ entrypoints==0.4
188
+ ephem==4.1.5
189
+ esda==2.5.1
190
+ essentia==2.1b6.dev1110
191
+ et-xmlfile==1.1.0
192
+ etils==1.6.0
193
+ exceptiongroup==1.2.0
194
+ executing==2.0.1
195
+ explainable-ai-sdk==1.3.3
196
+ fastai==2.7.14
197
+ fastapi==0.108.0
198
+ fastavro==1.9.3
199
+ fastcore==1.5.29
200
+ fastdownload==0.0.7
201
+ fasteners==0.19
202
+ fastjsonschema==2.19.1
203
+ fastprogress==1.0.3
204
+ fastrlock==0.8.2
205
+ fasttext==0.9.2
206
+ feather-format==0.4.1
207
+ featuretools==1.30.0
208
+ filelock==3.13.1
209
+ fiona==1.9.6
210
+ fitter==1.7.0
211
+ flake8==7.0.0
212
+ flashtext==2.7
213
+ flatbuffers==23.5.26
214
+ flax==0.8.2
215
+ folium==0.16.0
216
+ fonttools==4.47.0
217
+ fonttools==4.50.0
218
+ fqdn==1.5.1
219
+ frozendict==2.4.1
220
+ frozenlist==1.4.1
221
+ fsspec==2023.10.0
222
+ fsspec==2024.3.1
223
+ funcy==2.0
224
+ fury==0.10.0
225
+ future==1.0.0
226
+ fuzzywuzzy==0.18.0
227
+ gast==0.5.4
228
+ gatspy==0.3
229
+ gcsfs==2024.2.0
230
+ gensim==4.3.2
231
+ geographiclib==2.0
232
+ geojson==3.1.0
233
+ geopandas==0.14.3
234
+ geoplot==0.5.1
235
+ geopy==2.4.1
236
+ geoviews==1.11.1
237
+ ggplot==0.11.5
238
+ giddy==2.3.5
239
+ gitdb==4.0.11
240
+ google-ai-generativelanguage==0.4.0
241
+ google-api-core==2.11.1
242
+ google-api-core==2.18.0
243
+ google-api-python-client==2.125.0
244
+ google-apitools==0.5.31
245
+ google-auth-httplib2==0.2.0
246
+ google-auth-oauthlib==1.2.0
247
+ google-auth==2.26.1
248
+ google-cloud-aiplatform==0.6.0a1
249
+ google-cloud-artifact-registry==1.10.0
250
+ google-cloud-automl==1.0.1
251
+ google-cloud-bigquery==2.34.4
252
+ google-cloud-bigtable==1.7.3
253
+ google-cloud-core==2.4.1
254
+ google-cloud-datastore==2.19.0
255
+ google-cloud-dlp==3.14.0
256
+ google-cloud-jupyter-config==0.0.5
257
+ google-cloud-language==2.13.3
258
+ google-cloud-monitoring==2.18.0
259
+ google-cloud-pubsub==2.19.0
260
+ google-cloud-pubsublite==1.9.0
261
+ google-cloud-recommendations-ai==0.7.1
262
+ google-cloud-resource-manager==1.11.0
263
+ google-cloud-spanner==3.40.1
264
+ google-cloud-storage==1.44.0
265
+ google-cloud-translate==3.12.1
266
+ google-cloud-videointelligence==2.13.3
267
+ google-cloud-vision==2.8.0
268
+ google-crc32c==1.5.0
269
+ google-generativeai==0.4.1
270
+ google-pasta==0.2.0
271
+ google-resumable-media==2.7.0
272
+ googleapis-common-protos==1.62.0
273
+ gplearn==0.4.2
274
+ gpustat==1.0.0
275
+ gpxpy==1.6.2
276
+ graphviz==0.20.3
277
+ greenlet==3.0.3
278
+ grpc-google-iam-v1==0.12.7
279
+ grpcio-status==1.48.1
280
+ grpcio-status==1.48.2
281
+ grpcio==1.51.1
282
+ grpcio==1.60.0
283
+ gviz-api==1.10.0
284
+ gym-notices==0.0.8
285
+ gym==0.26.2
286
+ gymnasium==0.29.0
287
+ h11==0.14.0
288
+ h2o==3.46.0.1
289
+ h5netcdf==1.3.0
290
+ h5py==3.10.0
291
+ haversine==2.8.1
292
+ hdfs==2.7.3
293
+ hep-ml==0.7.2
294
+ hijri-converter==2.3.1
295
+ hmmlearn==0.3.2
296
+ holidays==0.24
297
+ holoviews==1.18.3
298
+ hpsklearn==0.1.0
299
+ html5lib==1.1
300
+ htmlmin==0.1.12
301
+ httpcore==1.0.5
302
+ httplib2==0.21.0
303
+ httptools==0.6.1
304
+ httpx==0.27.0
305
+ huggingface-hub==0.22.2
306
+ hunspell==0.5.5
307
+ hydra-slayer==0.5.0
308
+ hyperopt==0.2.7
309
+ hypertools==0.8.0
310
+ idna==3.6
311
+ igraph==0.11.4
312
+ imagecodecs==2024.1.1
313
+ imageio==2.33.1
314
+ imbalanced-learn==0.12.2
315
+ imgaug==0.4.0
316
+ importlib-metadata==6.11.0
317
+ importlib-metadata==7.0.1
318
+ importlib-resources==6.1.1
319
+ inequality==1.0.1
320
+ iniconfig==2.0.0
321
+ ipydatawidgets==4.3.5
322
+ ipykernel==6.28.0
323
+ ipyleaflet==0.18.2
324
+ ipympl==0.7.0
325
+ ipython-genutils==0.2.0
326
+ ipython-genutils==0.2.0
327
+ ipython-sql==0.5.0
328
+ ipython==8.20.0
329
+ ipyvolume==0.6.3
330
+ ipyvue==1.10.2
331
+ ipyvuetify==1.9.3
332
+ ipywebrtc==0.6.0
333
+ ipywidgets==7.7.1
334
+ isoduration==20.11.0
335
+ isort==5.13.2
336
+ isoweek==1.3.3
337
+ itsdangerous==2.1.2
338
+ jaraco.classes==3.3.0
339
+ jax-jumpy==1.0.0
340
+ jax==0.4.23
341
+ jaxlib==0.4.23.dev20240116
342
+ jedi==0.19.1
343
+ jeepney==0.8.0
344
+ jieba==0.42.1
345
+ jmespath==1.0.1
346
+ joblib==1.3.2
347
+ json5==0.9.14
348
+ jsonpatch==1.33
349
+ jsonpointer==2.4
350
+ jsonschema-specifications==2023.12.1
351
+ jsonschema==4.20.0
352
+ jupyter-console==6.6.3
353
+ jupyter-events==0.9.0
354
+ jupyter-http-over-ws==0.0.8
355
+ jupyter-lsp==1.5.1
356
+ jupyter-server-mathjax==0.2.6
357
+ jupyter-ydoc==0.2.5
358
+ jupyter_client==7.4.9
359
+ jupyter_client==8.6.0
360
+ jupyter_core==5.7.1
361
+ jupyter_server==2.13.0
362
+ jupyter_server_fileid==0.9.1
363
+ jupyter_server_proxy==4.1.0
364
+ jupyter_server_terminals==0.5.1
365
+ jupyter_server_ydoc==0.8.0
366
+ jupyterlab-lsp==5.1.0
367
+ jupyterlab-widgets==3.0.9
368
+ jupyterlab==4.1.5
369
+ jupyterlab_git==0.44.0
370
+ jupyterlab_pygments==0.3.0
371
+ jupyterlab_server==2.25.2
372
+ jupytext==1.16.0
373
+ kaggle-environments==1.14.3
374
+ kaggle==1.6.8
375
+ kagglehub==0.2.2
376
+ keras-cv==0.8.2
377
+ keras-nlp==0.8.2
378
+ keras-tuner==1.4.6
379
+ keras==3.1.1
380
+ kernels-mixer==0.0.7
381
+ keyring==24.3.0
382
+ keyrings.google-artifactregistry-auth==1.1.2
383
+ kfp-pipeline-spec==0.2.2
384
+ kfp-server-api==2.0.5
385
+ kfp==2.5.0
386
+ kiwisolver==1.4.5
387
+ kmapper==2.0.1
388
+ kmodes==0.12.2
389
+ korean-lunar-calendar==0.3.1
390
+ kornia==0.7.2
391
+ kornia_rs==0.1.3
392
+ kt-legacy==1.0.5
393
+ kubernetes==26.1.0
394
+ langcodes==3.3.0
395
+ langid==1.1.6
396
+ lazy_loader==0.3
397
+ learntools==0.3.4
398
+ leven==1.0.4
399
+ libclang==16.0.6
400
+ libmambapy==1.5.0
401
+ libpysal==4.9.2
402
+ librosa==0.10.1
403
+ lightgbm==4.2.0
404
+ lightning-utilities==0.11.2
405
+ lime==0.2.0.1
406
+ line-profiler==4.1.2
407
+ linkify-it-py==2.0.3
408
+ llvmlite==0.41.1
409
+ llvmlite==0.42.0
410
+ lml==0.1.0
411
+ locket==1.0.0
412
+ loguru==0.7.2
413
+ lxml==5.2.1
414
+ lz4==4.3.3
415
+ mamba==1.5.0
416
+ mapclassify==2.6.1
417
+ markdown-it-py==3.0.0
418
+ marshmallow==3.21.1
419
+ matplotlib-inline==0.1.6
420
+ matplotlib-venn==0.11.10
421
+ matplotlib==3.7.5
422
+ matplotlib==3.8.3
423
+ mccabe==0.7.0
424
+ mdit-py-plugins==0.4.0
425
+ mdurl==0.1.2
426
+ memory-profiler==0.61.0
427
+ menuinst==2.0.1
428
+ mercantile==1.2.1
429
+ mgwr==2.2.1
430
+ missingno==0.5.2
431
+ mistune==0.8.4
432
+ mizani==0.11.1
433
+ ml-dtypes==0.2.0
434
+ mlcrate==0.2.0
435
+ mlens==0.2.3
436
+ mlxtend==0.23.1
437
+ mne==1.6.1
438
+ mnist==0.2.2
439
+ momepy==0.7.0
440
+ more-itertools==10.2.0
441
+ mpld3==0.5.10
442
+ mpmath==1.3.0
443
+ msgpack==1.0.7
444
+ multidict==6.0.4
445
+ multimethod==1.10
446
+ multipledispatch==1.0.0
447
+ multiprocess==0.70.15
448
+ munkres==1.1.4
449
+ murmurhash==1.0.10
450
+ mypy-extensions==1.0.0
451
+ namex==0.0.7
452
+ nb-conda-kernels==2.3.1
453
+ nb_conda==2.2.1
454
+ nbclassic==1.0.0
455
+ nbclient==0.5.13
456
+ nbconvert==6.4.5
457
+ nbdime==3.2.0
458
+ nbformat==5.9.2
459
+ ndindex==1.8
460
+ nest-asyncio==1.5.8
461
+ networkx==3.2.1
462
+ nibabel==5.2.1
463
+ nilearn==0.10.3
464
+ ninja==1.11.1.1
465
+ nltk==3.2.4
466
+ nose==1.3.7
467
+ notebook==6.5.4
468
+ notebook==6.5.6
469
+ notebook_executor==0.2
470
+ notebook_shim==0.2.3
471
+ numba==0.58.1
472
+ numba==0.59.1
473
+ numexpr==2.10.0
474
+ numpy==1.26.4
475
+ nvidia-ml-py==11.495.46
476
+ nvtx==0.2.10
477
+ oauth2client==4.1.3
478
+ oauthlib==3.2.2
479
+ objsize==0.6.1
480
+ odfpy==1.4.1
481
+ olefile==0.47
482
+ onnx==1.16.0
483
+ opencensus-context==0.1.3
484
+ opencensus==0.11.4
485
+ opencv-contrib-python==4.9.0.80
486
+ opencv-python-headless==4.9.0.80
487
+ opencv-python==4.9.0.80
488
+ openpyxl==3.1.2
489
+ openslide-python==1.3.1
490
+ opentelemetry-api==1.22.0
491
+ opentelemetry-exporter-otlp-proto-common==1.22.0
492
+ opentelemetry-exporter-otlp-proto-grpc==1.22.0
493
+ opentelemetry-exporter-otlp-proto-http==1.22.0
494
+ opentelemetry-exporter-otlp==1.22.0
495
+ opentelemetry-proto==1.22.0
496
+ opentelemetry-sdk==1.22.0
497
+ opentelemetry-semantic-conventions==0.43b0
498
+ opt-einsum==3.3.0
499
+ optax==0.2.2
500
+ optree==0.11.0
501
+ optuna==3.6.1
502
+ orbax-checkpoint==0.5.7
503
+ ordered-set==4.1.0
504
+ orjson==3.9.10
505
+ ortools==9.4.1874
506
+ osmnx==1.9.2
507
+ overrides==7.4.0
508
+ packaging==21.3
509
+ pandas-datareader==0.10.0
510
+ pandas-profiling==3.6.6
511
+ pandas-summary==0.2.0
512
+ pandas==2.1.4
513
+ pandas==2.2.1
514
+ pandasql==0.7.3
515
+ pandocfilters==1.5.0
516
+ panel==1.3.8
517
+ papermill==2.5.0
518
+ param==2.1.0
519
+ parso==0.8.3
520
+ partd==1.4.1
521
+ path.py==12.5.0
522
+ path==16.10.0
523
+ pathos==0.3.2
524
+ pathy==0.10.3
525
+ patsy==0.5.6
526
+ pdf2image==1.17.0
527
+ peft==0.10.0
528
+ pettingzoo==1.24.0
529
+ pexpect==4.8.0
530
+ pexpect==4.9.0
531
+ phik==0.12.4
532
+ pickleshare==0.7.5
533
+ pillow==10.3.0
534
+ pip==23.3.2
535
+ pkgutil_resolve_name==1.3.10
536
+ platformdirs==4.2.0
537
+ plotly-express==0.4.1
538
+ plotly==5.18.0
539
+ plotnine==0.13.4
540
+ pluggy==1.4.0
541
+ pointpats==2.4.0
542
+ polars==0.20.18
543
+ polyglot==16.7.4
544
+ pooch==1.8.1
545
+ pox==0.3.4
546
+ ppca==0.0.4
547
+ ppft==1.7.6.8
548
+ preprocessing==0.1.13
549
+ preshed==3.0.9
550
+ prettytable==3.9.0
551
+ progressbar2==4.4.2
552
+ prometheus-client==0.19.0
553
+ promise==2.3
554
+ prompt-toolkit==3.0.42
555
+ prompt-toolkit==3.0.43
556
+ prophet==1.1.1
557
+ proto-plus==1.23.0
558
+ protobuf==3.20.3
559
+ protobuf==4.21.12
560
+ psutil==5.9.3
561
+ psutil==5.9.7
562
+ ptyprocess==0.7.0
563
+ pudb==2024.1
564
+ pure-eval==0.2.2
565
+ py-cpuinfo==9.0.0
566
+ py-spy==0.3.14
567
+ py4j==0.10.9.7
568
+ pyLDAvis==3.4.1
569
+ pyOpenSSL==23.3.0
570
+ pyaml==23.12.0
571
+ pyarrow-hotfix==0.6
572
+ pyarrow==15.0.2
573
+ pyasn1-modules==0.3.0
574
+ pyasn1==0.5.1
575
+ pybind11==2.12.0
576
+ pyclipper==1.3.0.post5
577
+ pycodestyle==2.11.1
578
+ pycosat==0.6.6
579
+ pycparser==2.21
580
+ pycryptodome==3.20.0
581
+ pyct==0.5.0
582
+ pycuda==2024.1
583
+ pydantic==2.5.3
584
+ pydantic==2.6.4
585
+ pydantic_core==2.14.6
586
+ pydantic_core==2.16.3
587
+ pydegensac==0.1.2
588
+ pydicom==2.4.4
589
+ pydocstyle==6.3.0
590
+ pydot==1.4.2
591
+ pydub==0.25.1
592
+ pyemd==1.0.0
593
+ pyerfa==2.0.1.1
594
+ pyexcel-io==0.6.6
595
+ pyexcel-ods==0.6.0
596
+ pyflakes==3.2.0
597
+ pygltflib==1.16.2
598
+ pykalman==0.9.7
599
+ pylibraft==23.8.0
600
+ pylint==3.1.0
601
+ pymc3==3.11.4
602
+ pymongo==3.13.0
603
+ pynndescent==0.5.12
604
+ pynvml==11.4.1
605
+ pynvrtc==9.2
606
+ pyparsing==3.1.1
607
+ pyparsing==3.1.2
608
+ pypdf==4.1.0
609
+ pyproj==3.6.1
610
+ pysal==24.1
611
+ pyshp==2.3.1
612
+ pytesseract==0.3.10
613
+ pytest==8.1.1
614
+ python-bidi==0.4.2
615
+ python-dateutil==2.9.0.post0
616
+ python-dotenv==1.0.0
617
+ python-json-logger==2.0.7
618
+ python-louvain==0.16
619
+ python-lsp-jsonrpc==1.1.2
620
+ python-lsp-server==1.11.0
621
+ python-slugify==8.0.4
622
+ python-utils==3.8.2
623
+ pythreejs==2.4.2
624
+ pytoolconfig==1.3.1
625
+ pytools==2024.1.1
626
+ pytorch-ignite==0.5.0.post2
627
+ pytorch-lightning==2.2.1
628
+ pytz==2023.3.post1
629
+ pytz==2024.1
630
+ pyu2f==0.1.5
631
+ pyviz_comms==3.0.2
632
+ pyzmq==24.0.1
633
+ pyzmq==25.1.2
634
+ qgrid==1.3.1
635
+ qtconsole==5.5.1
636
+ quantecon==0.7.2
637
+ qudida==0.0.4
638
+ raft-dask==23.8.0
639
+ rasterio==1.3.9
640
+ rasterstats==0.19.0
641
+ ray-cpp==2.9.0
642
+ ray==2.9.0
643
+ referencing==0.32.1
644
+ regex==2023.12.25
645
+ requests-oauthlib==1.3.1
646
+ requests-toolbelt==0.10.1
647
+ requests==2.31.0
648
+ retrying==1.3.3
649
+ retrying==1.3.4
650
+ rfc3339-validator==0.1.4
651
+ rfc3986-validator==0.1.1
652
+ rgf-python==3.12.0
653
+ rich-click==1.7.4
654
+ rich==13.7.0
655
+ rich==13.7.1
656
+ rmm==23.8.0
657
+ rope==1.13.0
658
+ rpds-py==0.16.2
659
+ rsa==4.9
660
+ ruamel-yaml-conda==0.15.100
661
+ ruamel.yaml.clib==0.2.7
662
+ ruamel.yaml==0.17.40
663
+ s2sphere==0.2.5
664
+ s3fs==2024.2.0
665
+ s3transfer==0.6.2
666
+ safetensors==0.4.2
667
+ scattertext==0.1.19
668
+ scikit-image==0.22.0
669
+ scikit-learn-intelex==2024.2.0
670
+ scikit-learn==1.2.2
671
+ scikit-multilearn==0.2.0
672
+ scikit-optimize==0.10.1
673
+ scikit-plot==0.3.7
674
+ scikit-surprise==1.1.3
675
+ scipy==1.11.4
676
+ scipy==1.12.0
677
+ seaborn==0.12.2
678
+ segment_anything==1.0
679
+ segregation==2.5
680
+ semver==3.0.2
681
+ sentencepiece==0.2.0
682
+ sentry-sdk==1.44.1
683
+ setproctitle==1.3.3
684
+ setuptools-git==1.2
685
+ setuptools-scm==8.0.4
686
+ setuptools==69.0.3
687
+ shap==0.44.1
688
+ shapely==2.0.3
689
+ shellingham==1.5.4
690
+ shtab==1.7.1
691
+ simpervisor==1.0.0
692
+ simplejson==3.19.2
693
+ six==1.16.0
694
+ sklearn-pandas==2.2.0
695
+ slicer==0.0.7
696
+ smart-open==6.4.0
697
+ smmap==5.0.1
698
+ sniffio==1.3.0
699
+ snowballstemmer==2.2.0
700
+ snuggs==1.4.7
701
+ sortedcontainers==2.4.0
702
+ soundfile==0.12.1
703
+ soupsieve==2.5
704
+ soxr==0.3.7
705
+ spacy-legacy==3.0.12
706
+ spacy-loggers==1.0.5
707
+ spacy==3.7.2
708
+ spaghetti==1.7.5.post1
709
+ spectral==0.23.1
710
+ spglm==1.1.0
711
+ sphinx-rtd-theme==0.2.4
712
+ spint==1.0.7
713
+ splot==1.1.5.post1
714
+ spopt==0.6.0
715
+ spreg==1.4.2
716
+ spvcm==0.3.0
717
+ sqlparse==0.4.4
718
+ squarify==0.4.3
719
+ srsly==2.4.8
720
+ stable-baselines3==2.1.0
721
+ stack-data==0.6.2
722
+ stack-data==0.6.3
723
+ stanio==0.5.0
724
+ starlette==0.32.0.post1
725
+ statsmodels==0.14.1
726
+ stemming==1.0.1
727
+ stop-words==2018.7.23
728
+ stopit==1.1.2
729
+ stumpy==1.12.0
730
+ sympy==1.12
731
+ tables==3.9.2
732
+ tabulate==0.9.0
733
+ tangled-up-in-unicode==0.2.0
734
+ tbb==2021.12.0
735
+ tblib==3.0.0
736
+ tenacity==8.2.3
737
+ tensorboard-data-server==0.7.2
738
+ tensorboard-plugin-profile==2.15.0
739
+ tensorboard==2.15.1
740
+ tensorboardX==2.6.2.2
741
+ tensorflow-cloud==0.1.16
742
+ tensorflow-datasets==4.9.4
743
+ tensorflow-decision-forests==1.8.1
744
+ tensorflow-estimator==2.15.0
745
+ tensorflow-hub==0.16.1
746
+ tensorflow-io-gcs-filesystem==0.35.0
747
+ tensorflow-io==0.35.0
748
+ tensorflow-metadata==0.14.0
749
+ tensorflow-probability==0.23.0
750
+ tensorflow-serving-api==2.14.1
751
+ tensorflow-text==2.15.0
752
+ tensorflow-transform==0.14.0
753
+ tensorflow==2.15.0
754
+ tensorstore==0.1.56
755
+ termcolor==2.4.0
756
+ terminado==0.18.0
757
+ testpath==0.6.0
758
+ text-unidecode==1.3
759
+ textblob==0.18.0.post0
760
+ texttable==1.7.0
761
+ tf_keras==2.15.1
762
+ tfp-nightly==0.24.0.dev0
763
+ thinc==8.2.2
764
+ threadpoolctl==3.2.0
765
+ tifffile==2023.12.9
766
+ timm==0.9.16
767
+ tinycss2==1.2.1
768
+ tobler==0.11.2
769
+ tokenizers==0.15.2
770
+ toml==0.10.2
771
+ tomli==2.0.1
772
+ tomlkit==0.12.4
773
+ toolz==0.12.1
774
+ torch==2.1.2
775
+ torchaudio==2.1.2
776
+ torchdata==0.7.1
777
+ torchinfo==1.8.0
778
+ torchmetrics==1.3.2
779
+ torchtext==0.16.2
780
+ torchvision==0.16.2
781
+ tornado==6.3.3
782
+ tqdm==4.66.1
783
+ traceml==1.0.8
784
+ traitlets==5.9.0
785
+ traittypes==0.2.1
786
+ transformers==4.39.3
787
+ treelite-runtime==3.2.0
788
+ treelite==3.2.0
789
+ trl==0.8.2
790
+ truststore==0.8.0
791
+ trx-python==0.2.9
792
+ tsfresh==0.20.2
793
+ typeguard==4.1.5
794
+ typer==0.9.0
795
+ typer==0.9.4
796
+ types-python-dateutil==2.8.19.20240106
797
+ typing-inspect==0.9.0
798
+ typing-utils==0.1.0
799
+ typing_extensions==4.9.0
800
+ tyro==0.8.3
801
+ tzdata==2023.4
802
+ uc-micro-py==1.0.3
803
+ ucx-py==0.33.0
804
+ ujson==5.9.0
805
+ umap-learn==0.5.5
806
+ unicodedata2==15.1.0
807
+ update-checker==0.18.0
808
+ uri-template==1.3.0
809
+ uritemplate==3.0.1
810
+ urllib3==1.26.18
811
+ urllib3==2.1.0
812
+ urwid==2.6.10
813
+ urwid_readline==0.14
814
+ uvicorn==0.25.0
815
+ uvloop==0.19.0
816
+ vaex-astro==0.9.3
817
+ vaex-core==4.17.1
818
+ vaex-hdf5==0.14.1
819
+ vaex-jupyter==0.8.2
820
+ vaex-ml==0.18.3
821
+ vaex-server==0.9.0
822
+ vaex-viz==0.5.4
823
+ vaex==4.17.0
824
+ vec_noise==1.1.4
825
+ vecstack==0.4.0
826
+ virtualenv==20.21.0
827
+ visions==0.7.5
828
+ vowpalwabbit==9.9.0
829
+ vtk==9.3.0
830
+ wandb==0.16.5
831
+ wasabi==1.1.2
832
+ watchfiles==0.21.0
833
+ wavio==0.0.8
834
+ wcwidth==0.2.13
835
+ weasel==0.3.4
836
+ webcolors==1.13
837
+ webencodings==0.5.1
838
+ websocket-client==1.7.0
839
+ websockets==12.0
840
+ wfdb==4.1.2
841
+ whatthepatch==1.0.5
842
+ wheel==0.42.0
843
+ widgetsnbextension==3.6.6
844
+ witwidget==1.8.1
845
+ woodwork==0.29.0
846
+ wordcloud==1.9.3
847
+ wordsegment==1.3.1
848
+ wrapt==1.14.1
849
+ xarray-einstats==0.7.0
850
+ xarray==2024.3.0
851
+ xgboost==2.0.3
852
+ xvfbwrapper==0.2.9
853
+ xxhash==3.4.1
854
+ xyzservices==2023.10.1
855
+ y-py==0.6.2
856
+ yapf==0.40.2
857
+ yarl==1.9.3
858
+ yarl==1.9.4
859
+ ydata-profiling==4.6.4
860
+ yellowbrick==1.5
861
+ ypy-websocket==0.8.4
862
+ zict==3.0.0
863
+ zipp==3.17.0
864
+ zstandard==0.22.0
wandb/run-20240412_073555-bw7oy9ix/files/wandb-metadata.json ADDED
@@ -0,0 +1,66 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.15.133+-x86_64-with-glibc2.31",
3
+ "python": "3.10.13",
4
+ "heartbeatAt": "2024-04-12T07:35:56.437403",
5
+ "startedAt": "2024-04-12T07:35:55.741618",
6
+ "docker": null,
7
+ "cuda": null,
8
+ "args": [],
9
+ "state": "running",
10
+ "program": "kaggle.ipynb",
11
+ "codePathLocal": null,
12
+ "root": "/kaggle/working",
13
+ "host": "e5a48bec8248",
14
+ "username": "root",
15
+ "executable": "/opt/conda/bin/python3.10",
16
+ "cpu_count": 2,
17
+ "cpu_count_logical": 4,
18
+ "cpu_freq": {
19
+ "current": 2000.138,
20
+ "min": 0.0,
21
+ "max": 0.0
22
+ },
23
+ "cpu_freq_per_core": [
24
+ {
25
+ "current": 2000.138,
26
+ "min": 0.0,
27
+ "max": 0.0
28
+ },
29
+ {
30
+ "current": 2000.138,
31
+ "min": 0.0,
32
+ "max": 0.0
33
+ },
34
+ {
35
+ "current": 2000.138,
36
+ "min": 0.0,
37
+ "max": 0.0
38
+ },
39
+ {
40
+ "current": 2000.138,
41
+ "min": 0.0,
42
+ "max": 0.0
43
+ }
44
+ ],
45
+ "disk": {
46
+ "/": {
47
+ "total": 8062.387607574463,
48
+ "used": 5565.782459259033
49
+ }
50
+ },
51
+ "gpu": "Tesla T4",
52
+ "gpu_count": 2,
53
+ "gpu_devices": [
54
+ {
55
+ "name": "Tesla T4",
56
+ "memory_total": 16106127360
57
+ },
58
+ {
59
+ "name": "Tesla T4",
60
+ "memory_total": 16106127360
61
+ }
62
+ ],
63
+ "memory": {
64
+ "total": 31.357559204101562
65
+ }
66
+ }
wandb/run-20240412_073555-bw7oy9ix/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"train/loss": 4.6771, "train/grad_norm": 79.12922668457031, "train/learning_rate": 1e-06, "train/epoch": 0.0, "train/global_step": 20, "_timestamp": 1712907379.5892065, "_runtime": 23.840320348739624, "_step": 0}
wandb/run-20240412_073555-bw7oy9ix/logs/debug-internal.log ADDED
@@ -0,0 +1,77 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-04-12 07:35:55,749 INFO StreamThr :334 [internal.py:wandb_internal():86] W&B internal server running at pid: 334, started at: 2024-04-12 07:35:55.748821
2
+ 2024-04-12 07:35:55,751 DEBUG HandlerThread:334 [handler.py:handle_request():146] handle_request: status
3
+ 2024-04-12 07:35:56,159 INFO WriterThread:334 [datastore.py:open_for_write():87] open: /kaggle/working/wandb/run-20240412_073555-bw7oy9ix/run-bw7oy9ix.wandb
4
+ 2024-04-12 07:35:56,160 DEBUG SenderThread:334 [sender.py:send():379] send: header
5
+ 2024-04-12 07:35:56,163 DEBUG SenderThread:334 [sender.py:send():379] send: run
6
+ 2024-04-12 07:35:56,307 INFO SenderThread:334 [dir_watcher.py:__init__():211] watching files in: /kaggle/working/wandb/run-20240412_073555-bw7oy9ix/files
7
+ 2024-04-12 07:35:56,307 INFO SenderThread:334 [sender.py:_start_run_threads():1124] run started: bw7oy9ix with start time 1712907355.748886
8
+ 2024-04-12 07:35:56,317 DEBUG HandlerThread:334 [handler.py:handle_request():146] handle_request: check_version
9
+ 2024-04-12 07:35:56,317 DEBUG SenderThread:334 [sender.py:send_request():406] send_request: check_version
10
+ 2024-04-12 07:35:56,409 DEBUG HandlerThread:334 [handler.py:handle_request():146] handle_request: run_start
11
+ 2024-04-12 07:35:56,421 DEBUG HandlerThread:334 [system_info.py:__init__():26] System info init
12
+ 2024-04-12 07:35:56,421 DEBUG HandlerThread:334 [system_info.py:__init__():41] System info init done
13
+ 2024-04-12 07:35:56,422 INFO HandlerThread:334 [system_monitor.py:start():194] Starting system monitor
14
+ 2024-04-12 07:35:56,422 INFO SystemMonitor:334 [system_monitor.py:_start():158] Starting system asset monitoring threads
15
+ 2024-04-12 07:35:56,422 INFO HandlerThread:334 [system_monitor.py:probe():214] Collecting system info
16
+ 2024-04-12 07:35:56,423 INFO SystemMonitor:334 [interfaces.py:start():190] Started cpu monitoring
17
+ 2024-04-12 07:35:56,423 INFO SystemMonitor:334 [interfaces.py:start():190] Started disk monitoring
18
+ 2024-04-12 07:35:56,424 INFO SystemMonitor:334 [interfaces.py:start():190] Started gpu monitoring
19
+ 2024-04-12 07:35:56,425 INFO SystemMonitor:334 [interfaces.py:start():190] Started memory monitoring
20
+ 2024-04-12 07:35:56,426 INFO SystemMonitor:334 [interfaces.py:start():190] Started network monitoring
21
+ 2024-04-12 07:35:56,437 DEBUG HandlerThread:334 [system_info.py:probe():150] Probing system
22
+ 2024-04-12 07:35:56,439 DEBUG HandlerThread:334 [gitlib.py:_init_repo():56] git repository is invalid
23
+ 2024-04-12 07:35:56,439 DEBUG HandlerThread:334 [system_info.py:probe():198] Probing system done
24
+ 2024-04-12 07:35:56,439 DEBUG HandlerThread:334 [system_monitor.py:probe():223] {'os': 'Linux-5.15.133+-x86_64-with-glibc2.31', 'python': '3.10.13', 'heartbeatAt': '2024-04-12T07:35:56.437403', 'startedAt': '2024-04-12T07:35:55.741618', 'docker': None, 'cuda': None, 'args': (), 'state': 'running', 'program': 'kaggle.ipynb', 'codePathLocal': None, 'root': '/kaggle/working', 'host': 'e5a48bec8248', 'username': 'root', 'executable': '/opt/conda/bin/python3.10', 'cpu_count': 2, 'cpu_count_logical': 4, 'cpu_freq': {'current': 2000.138, 'min': 0.0, 'max': 0.0}, 'cpu_freq_per_core': [{'current': 2000.138, 'min': 0.0, 'max': 0.0}, {'current': 2000.138, 'min': 0.0, 'max': 0.0}, {'current': 2000.138, 'min': 0.0, 'max': 0.0}, {'current': 2000.138, 'min': 0.0, 'max': 0.0}], 'disk': {'/': {'total': 8062.387607574463, 'used': 5565.782459259033}}, 'gpu': 'Tesla T4', 'gpu_count': 2, 'gpu_devices': [{'name': 'Tesla T4', 'memory_total': 16106127360}, {'name': 'Tesla T4', 'memory_total': 16106127360}], 'memory': {'total': 31.357559204101562}}
25
+ 2024-04-12 07:35:56,440 INFO HandlerThread:334 [system_monitor.py:probe():224] Finished collecting system info
26
+ 2024-04-12 07:35:56,440 INFO HandlerThread:334 [system_monitor.py:probe():227] Publishing system info
27
+ 2024-04-12 07:35:56,440 DEBUG HandlerThread:334 [system_info.py:_save_conda():207] Saving list of conda packages installed into the current environment
28
+ 2024-04-12 07:35:57,309 INFO Thread-12 :334 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240412_073555-bw7oy9ix/files/conda-environment.yaml
29
+ 2024-04-12 07:36:11,455 ERROR HandlerThread:334 [system_info.py:_save_conda():221] Error saving conda packages: Command '['conda', 'env', 'export']' timed out after 15 seconds
30
+ Traceback (most recent call last):
31
+ File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/internal/system/system_info.py", line 214, in _save_conda
32
+ subprocess.call(
33
+ File "/opt/conda/lib/python3.10/subprocess.py", line 347, in call
34
+ return p.wait(timeout=timeout)
35
+ File "/opt/conda/lib/python3.10/subprocess.py", line 1209, in wait
36
+ return self._wait(timeout=timeout)
37
+ File "/opt/conda/lib/python3.10/subprocess.py", line 1951, in _wait
38
+ raise TimeoutExpired(self.args, timeout)
39
+ subprocess.TimeoutExpired: Command '['conda', 'env', 'export']' timed out after 15 seconds
40
+ 2024-04-12 07:36:11,456 DEBUG HandlerThread:334 [system_info.py:_save_conda():222] Saving conda packages done
41
+ 2024-04-12 07:36:11,456 INFO HandlerThread:334 [system_monitor.py:probe():229] Finished publishing system info
42
+ 2024-04-12 07:36:11,462 DEBUG HandlerThread:334 [handler.py:handle_request():146] handle_request: status_report
43
+ 2024-04-12 07:36:11,462 DEBUG HandlerThread:334 [handler.py:handle_request():146] handle_request: keepalive
44
+ 2024-04-12 07:36:11,462 DEBUG HandlerThread:334 [handler.py:handle_request():146] handle_request: status_report
45
+ 2024-04-12 07:36:11,462 DEBUG HandlerThread:334 [handler.py:handle_request():146] handle_request: keepalive
46
+ 2024-04-12 07:36:11,462 DEBUG HandlerThread:334 [handler.py:handle_request():146] handle_request: status_report
47
+ 2024-04-12 07:36:11,462 DEBUG HandlerThread:334 [handler.py:handle_request():146] handle_request: keepalive
48
+ 2024-04-12 07:36:11,463 DEBUG SenderThread:334 [sender.py:send():379] send: files
49
+ 2024-04-12 07:36:11,463 INFO SenderThread:334 [sender.py:_save_file():1390] saving file wandb-metadata.json with policy now
50
+ 2024-04-12 07:36:11,664 INFO wandb-upload_0:334 [upload_job.py:push():131] Uploaded file /tmp/tmpjph8qv3dwandb/f1up76ir-wandb-metadata.json
51
+ 2024-04-12 07:36:12,312 INFO Thread-12 :334 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240412_073555-bw7oy9ix/files/wandb-metadata.json
52
+ 2024-04-12 07:36:12,509 DEBUG HandlerThread:334 [handler.py:handle_request():146] handle_request: python_packages
53
+ 2024-04-12 07:36:12,509 DEBUG SenderThread:334 [sender.py:send_request():406] send_request: python_packages
54
+ 2024-04-12 07:36:12,512 DEBUG SenderThread:334 [sender.py:send():379] send: telemetry
55
+ 2024-04-12 07:36:12,523 DEBUG SenderThread:334 [sender.py:send():379] send: config
56
+ 2024-04-12 07:36:12,525 DEBUG SenderThread:334 [sender.py:send():379] send: metric
57
+ 2024-04-12 07:36:12,534 DEBUG HandlerThread:334 [handler.py:handle_request():146] handle_request: stop_status
58
+ 2024-04-12 07:36:12,535 DEBUG SenderThread:334 [sender.py:send():379] send: telemetry
59
+ 2024-04-12 07:36:12,535 DEBUG HandlerThread:334 [handler.py:handle_request():146] handle_request: internal_messages
60
+ 2024-04-12 07:36:12,535 DEBUG SenderThread:334 [sender.py:send():379] send: metric
61
+ 2024-04-12 07:36:12,536 WARNING SenderThread:334 [sender.py:send_metric():1341] Seen metric with glob (shouldn't happen)
62
+ 2024-04-12 07:36:12,536 DEBUG SenderThread:334 [sender.py:send():379] send: telemetry
63
+ 2024-04-12 07:36:12,536 DEBUG SenderThread:334 [sender.py:send_request():406] send_request: stop_status
64
+ 2024-04-12 07:36:13,313 INFO Thread-12 :334 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240412_073555-bw7oy9ix/files/requirements.txt
65
+ 2024-04-12 07:36:13,313 INFO Thread-12 :334 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240412_073555-bw7oy9ix/files/output.log
66
+ 2024-04-12 07:36:15,314 INFO Thread-12 :334 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240412_073555-bw7oy9ix/files/output.log
67
+ 2024-04-12 07:36:17,014 DEBUG HandlerThread:334 [handler.py:handle_request():146] handle_request: status_report
68
+ 2024-04-12 07:36:19,589 DEBUG HandlerThread:334 [handler.py:handle_request():146] handle_request: partial_history
69
+ 2024-04-12 07:36:19,592 DEBUG SenderThread:334 [sender.py:send():379] send: metric
70
+ 2024-04-12 07:36:19,593 DEBUG SenderThread:334 [sender.py:send():379] send: metric
71
+ 2024-04-12 07:36:19,593 DEBUG SenderThread:334 [sender.py:send():379] send: metric
72
+ 2024-04-12 07:36:19,593 DEBUG SenderThread:334 [sender.py:send():379] send: metric
73
+ 2024-04-12 07:36:19,593 DEBUG SenderThread:334 [sender.py:send():379] send: history
74
+ 2024-04-12 07:36:19,593 DEBUG SenderThread:334 [sender.py:send_request():406] send_request: summary_record
75
+ 2024-04-12 07:36:19,595 INFO SenderThread:334 [sender.py:_save_file():1390] saving file wandb-summary.json with policy end
76
+ 2024-04-12 07:36:20,316 INFO Thread-12 :334 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240412_073555-bw7oy9ix/files/wandb-summary.json
77
+ 2024-04-12 07:36:21,316 INFO Thread-12 :334 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240412_073555-bw7oy9ix/files/output.log
wandb/run-20240412_073555-bw7oy9ix/logs/debug.log ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-04-12 07:35:55,744 INFO MainThread:289 [wandb_setup.py:_flush():76] Current SDK version is 0.16.5
2
+ 2024-04-12 07:35:55,744 INFO MainThread:289 [wandb_setup.py:_flush():76] Configure stats pid to 289
3
+ 2024-04-12 07:35:55,744 INFO MainThread:289 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
4
+ 2024-04-12 07:35:55,744 INFO MainThread:289 [wandb_setup.py:_flush():76] Loading settings from /kaggle/working/wandb/settings
5
+ 2024-04-12 07:35:55,744 INFO MainThread:289 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
6
+ 2024-04-12 07:35:55,744 INFO MainThread:289 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
7
+ 2024-04-12 07:35:55,744 INFO MainThread:289 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program': '<python with no main file>'}
8
+ 2024-04-12 07:35:55,744 INFO MainThread:289 [wandb_setup.py:_flush():76] Applying login settings: {}
9
+ 2024-04-12 07:35:55,744 INFO MainThread:289 [wandb_init.py:_log_setup():527] Logging user logs to /kaggle/working/wandb/run-20240412_073555-bw7oy9ix/logs/debug.log
10
+ 2024-04-12 07:35:55,744 INFO MainThread:289 [wandb_init.py:_log_setup():528] Logging internal logs to /kaggle/working/wandb/run-20240412_073555-bw7oy9ix/logs/debug-internal.log
11
+ 2024-04-12 07:35:55,744 INFO MainThread:289 [wandb_init.py:_jupyter_setup():473] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x7c2cd1fccd60>
12
+ 2024-04-12 07:35:55,745 INFO MainThread:289 [wandb_init.py:init():567] calling init triggers
13
+ 2024-04-12 07:35:55,745 INFO MainThread:289 [wandb_init.py:init():574] wandb.init called with sweep_config: {}
14
+ config: {}
15
+ 2024-04-12 07:35:55,745 INFO MainThread:289 [wandb_init.py:init():617] starting backend
16
+ 2024-04-12 07:35:55,745 INFO MainThread:289 [wandb_init.py:init():621] setting up manager
17
+ 2024-04-12 07:35:55,747 INFO MainThread:289 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
18
+ 2024-04-12 07:35:55,748 INFO MainThread:289 [wandb_init.py:init():629] backend started and connected
19
+ 2024-04-12 07:35:55,761 INFO MainThread:289 [wandb_run.py:_label_probe_notebook():1299] probe notebook
20
+ 2024-04-12 07:35:56,159 INFO MainThread:289 [wandb_init.py:init():721] updated telemetry
21
+ 2024-04-12 07:35:56,162 INFO MainThread:289 [wandb_init.py:init():754] communicating run to backend with 90.0 second timeout
22
+ 2024-04-12 07:35:56,316 INFO MainThread:289 [wandb_run.py:_on_init():2344] communicating current version
23
+ 2024-04-12 07:35:56,402 INFO MainThread:289 [wandb_run.py:_on_init():2353] got version response upgrade_message: "wandb version 0.16.6 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
24
+
25
+ 2024-04-12 07:35:56,403 INFO MainThread:289 [wandb_init.py:init():805] starting run threads in backend
26
+ 2024-04-12 07:36:12,510 INFO MainThread:289 [wandb_run.py:_console_start():2323] atexit reg
27
+ 2024-04-12 07:36:12,510 INFO MainThread:289 [wandb_run.py:_redirect():2178] redirect: wrap_raw
28
+ 2024-04-12 07:36:12,511 INFO MainThread:289 [wandb_run.py:_redirect():2243] Wrapping output streams.
29
+ 2024-04-12 07:36:12,511 INFO MainThread:289 [wandb_run.py:_redirect():2268] Redirects installed.
30
+ 2024-04-12 07:36:12,512 INFO MainThread:289 [wandb_init.py:init():848] run started, returning control to user process
31
+ 2024-04-12 07:36:12,518 INFO MainThread:289 [wandb_run.py:_config_callback():1347] config_cb None None {'vocab_size': 250880, 'hidden_size': 1536, 'n_layer': 24, 'n_head': 16, 'layer_norm_epsilon': 1e-05, 'initializer_range': 0.02, 'use_cache': False, 'pretraining_tp': 1, 'apply_residual_connection_post_layernorm': False, 'hidden_dropout': 0.0, 'attention_dropout': 0.0, 'bos_token_id': 1, 'eos_token_id': 2, 'slow_but_exact': False, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'bfloat16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['BloomForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'pad_token_id': 3, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'bigscience/bloomz-1b1', 'transformers_version': '4.39.3', 'attention_softmax_in_fp32': True, 'bias_dropout_fusion': True, 'unk_token_id': 0, 'masked_softmax_fusion': True, 'model_type': 'bloom', 'n_inner': None, 'offset_alibi': 100, 'seq_length': 2048, 'skip_bias_add': True, 'skip_bias_add_qkv': False, 'quantization_config': {'quant_method': 'QuantizationMethod.BITS_AND_BYTES', '_load_in_8bit': False, '_load_in_4bit': True, 'llm_int8_threshold': 6.0, 'llm_int8_skip_modules': None, 'llm_int8_enable_fp32_cpu_offload': False, 'llm_int8_has_fp16_weight': False, 'bnb_4bit_quant_type': 'nf4', 'bnb_4bit_use_double_quant': True, 'bnb_4bit_compute_dtype': 'float16', 'bnb_4bit_quant_storage': 'uint8', 'load_in_4bit': True, 'load_in_8bit': False}, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': False, 'do_predict': False, 'evaluation_strategy': 'no', 'prediction_loss_only': False, 'per_device_train_batch_size': 2, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 3e-05, 'weight_decay': 0.0001, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 0.3, 'num_train_epochs': 5, 'max_steps': 20000, 'lr_scheduler_type': 'cosine', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.03, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Apr12_07-35-52_e5a48bec8248', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 20, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 20, 'save_total_limit': 1, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'paged_adamw_8bit', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': 'Femboyuwu2000/bloomz-1b1-vn-chat', 'hub_strategy': 'checkpoint', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': True, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None}
wandb/run-20240412_073555-bw7oy9ix/run-bw7oy9ix.wandb ADDED
Binary file (6.68 kB). View file