diff --git "a/experiment/rwkv-x-exp/v5-memory/v5-L96-D1024-E1e-1-ctx4k-part5.ipynb" "b/experiment/rwkv-x-exp/v5-memory/v5-L96-D1024-E1e-1-ctx4k-part5.ipynb" --- "a/experiment/rwkv-x-exp/v5-memory/v5-L96-D1024-E1e-1-ctx4k-part5.ipynb" +++ "b/experiment/rwkv-x-exp/v5-memory/v5-L96-D1024-E1e-1-ctx4k-part5.ipynb" @@ -3,13 +3,13 @@ { "attachments": {}, "cell_type": "markdown", - "id": "0aa470f4", + "id": "f1d7ccda", "metadata": { "papermill": { - "duration": 0.002959, - "end_time": "2023-09-01T14:53:06.162658", + "duration": 0.002567, + "end_time": "2023-09-02T06:16:28.958047", "exception": false, - "start_time": "2023-09-01T14:53:06.159699", + "start_time": "2023-09-02T06:16:28.955480", "status": "completed" }, "tags": [] @@ -28,13 +28,13 @@ { "attachments": {}, "cell_type": "markdown", - "id": "30b4429d", + "id": "54efc4fa", "metadata": { "papermill": { - "duration": 0.001695, - "end_time": "2023-09-01T14:53:06.166400", + "duration": 0.001679, + "end_time": "2023-09-02T06:16:28.961787", "exception": false, - "start_time": "2023-09-01T14:53:06.164705", + "start_time": "2023-09-02T06:16:28.960108", "status": "completed" }, "tags": [] @@ -46,19 +46,19 @@ { "cell_type": "code", "execution_count": 1, - "id": "b9d8c411", + "id": "8b4caf1b", "metadata": { "execution": { - "iopub.execute_input": "2023-09-01T14:53:06.171899Z", - "iopub.status.busy": "2023-09-01T14:53:06.171621Z", - "iopub.status.idle": "2023-09-01T14:53:06.892025Z", - "shell.execute_reply": "2023-09-01T14:53:06.891053Z" + "iopub.execute_input": "2023-09-02T06:16:28.967200Z", + "iopub.status.busy": "2023-09-02T06:16:28.966328Z", + "iopub.status.idle": "2023-09-02T06:16:29.684585Z", + "shell.execute_reply": "2023-09-02T06:16:29.683732Z" }, "papermill": { - "duration": 0.72586, - "end_time": "2023-09-01T14:53:06.894118", + "duration": 0.722814, + "end_time": "2023-09-02T06:16:29.686619", "exception": false, - "start_time": "2023-09-01T14:53:06.168258", + "start_time": "2023-09-02T06:16:28.963805", "status": "completed" }, "tags": [] @@ -74,19 +74,19 @@ { "cell_type": "code", "execution_count": 2, - "id": "d6a7b76f", + "id": "6cc3a721", "metadata": { "execution": { - "iopub.execute_input": "2023-09-01T14:53:06.899776Z", - "iopub.status.busy": "2023-09-01T14:53:06.899528Z", - "iopub.status.idle": "2023-09-01T14:53:09.788640Z", - "shell.execute_reply": "2023-09-01T14:53:09.787798Z" + "iopub.execute_input": "2023-09-02T06:16:29.691742Z", + "iopub.status.busy": "2023-09-02T06:16:29.691539Z", + "iopub.status.idle": "2023-09-02T06:16:32.527158Z", + "shell.execute_reply": "2023-09-02T06:16:32.526387Z" }, "papermill": { - "duration": 2.894122, - "end_time": "2023-09-01T14:53:09.790308", + "duration": 2.840419, + "end_time": "2023-09-02T06:16:32.529061", "exception": false, - "start_time": "2023-09-01T14:53:06.896186", + "start_time": "2023-09-02T06:16:29.688642", "status": "completed" }, "tags": [] @@ -118,19 +118,19 @@ { "cell_type": "code", "execution_count": 3, - "id": "7f700082", + "id": "6c3b0fbe", "metadata": { "execution": { - "iopub.execute_input": "2023-09-01T14:53:09.796934Z", - "iopub.status.busy": "2023-09-01T14:53:09.795855Z", - "iopub.status.idle": "2023-09-01T14:53:09.802854Z", - "shell.execute_reply": "2023-09-01T14:53:09.802131Z" + "iopub.execute_input": "2023-09-02T06:16:32.534904Z", + "iopub.status.busy": "2023-09-02T06:16:32.534704Z", + "iopub.status.idle": "2023-09-02T06:16:32.541251Z", + "shell.execute_reply": "2023-09-02T06:16:32.540683Z" }, "papermill": { - "duration": 0.011279, - "end_time": "2023-09-01T14:53:09.804026", + "duration": 0.010568, + "end_time": "2023-09-02T06:16:32.542402", "exception": false, - "start_time": "2023-09-01T14:53:09.792747", + "start_time": "2023-09-02T06:16:32.531834", "status": "completed" }, "tags": [] @@ -193,19 +193,19 @@ { "cell_type": "code", "execution_count": 4, - "id": "06ddc114", + "id": "2bb9275c", "metadata": { "execution": { - "iopub.execute_input": "2023-09-01T14:53:09.809533Z", - "iopub.status.busy": "2023-09-01T14:53:09.808951Z", - "iopub.status.idle": "2023-09-01T14:53:10.832978Z", - "shell.execute_reply": "2023-09-01T14:53:10.832125Z" + "iopub.execute_input": "2023-09-02T06:16:32.547321Z", + "iopub.status.busy": "2023-09-02T06:16:32.547162Z", + "iopub.status.idle": "2023-09-02T06:17:41.766730Z", + "shell.execute_reply": "2023-09-02T06:17:41.765728Z" }, "papermill": { - "duration": 1.028292, - "end_time": "2023-09-01T14:53:10.834448", + "duration": 69.223979, + "end_time": "2023-09-02T06:17:41.768492", "exception": false, - "start_time": "2023-09-01T14:53:09.806156", + "start_time": "2023-09-02T06:16:32.544513", "status": "completed" }, "tags": [] @@ -215,22 +215,35 @@ "name": "stdout", "output_type": "stream", "text": [ - "--2023-09-01 14:53:09-- https://huggingface.co/rwkv-x-dev/rwkv-x-playground/resolve/main/experiment/rwkv-x-exp/v5-memory/v5-L96-D1024-E0_1-mem-ctx-4k.pth\r\n", - "Resolving huggingface.co (huggingface.co)... 18.172.134.4, 18.172.134.88, 18.172.134.124, ...\r\n", - "Connecting to huggingface.co (huggingface.co)|18.172.134.4|:443... " + "--2023-09-02 06:16:32-- https://huggingface.co/rwkv-x-dev/rwkv-x-playground/resolve/main/experiment/rwkv-x-exp/v5-memory/v5-L96-D1024-E0_1-mem-ctx-4k.pth\r\n", + "Resolving huggingface.co (huggingface.co)... " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "connected.\r\n" + "18.165.122.101, 18.165.122.11, 18.165.122.120, ...\r\n", + "Connecting to huggingface.co (huggingface.co)|18.165.122.101|:443... connected.\r\n", + "HTTP request sent, awaiting response... " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "302 Found\r\n", + "Location: https://cdn-lfs.huggingface.co/repos/2e/f7/2ef78555202aa92abdbdf476ce3d0fd5a8b15f7245edf0b80d4d30572355f30d/20fb328bd798a1e23967f80661a0b6a277f1d45ed2d90760cec68d32dfcbd516?response-content-disposition=attachment%3B+filename*%3DUTF-8%27%27v5-L96-D1024-E0_1-mem-ctx-4k.pth%3B+filename%3D%22v5-L96-D1024-E0_1-mem-ctx-4k.pth%22%3B&Expires=1693894592&Policy=eyJTdGF0ZW1lbnQiOlt7IkNvbmRpdGlvbiI6eyJEYXRlTGVzc1RoYW4iOnsiQVdTOkVwb2NoVGltZSI6MTY5Mzg5NDU5Mn19LCJSZXNvdXJjZSI6Imh0dHBzOi8vY2RuLWxmcy5odWdnaW5nZmFjZS5jby9yZXBvcy8yZS9mNy8yZWY3ODU1NTIwMmFhOTJhYmRiZGY0NzZjZTNkMGZkNWE4YjE1ZjcyNDVlZGYwYjgwZDRkMzA1NzIzNTVmMzBkLzIwZmIzMjhiZDc5OGExZTIzOTY3ZjgwNjYxYTBiNmEyNzdmMWQ0NWVkMmQ5MDc2MGNlYzY4ZDMyZGZjYmQ1MTY%7EcmVzcG9uc2UtY29udGVudC1kaXNwb3NpdGlvbj0qIn1dfQ__&Signature=0lvghwVju7ugxwlMlvj%7EIeK4lhVAMO1bcdYQsP0FxGPt5AK88smuTfGJvprL1l1HiQVj4a7f6U8yvc97nQxIdWdTF0RC8s4SAyFcZoupxdMbzOQ7kEr7OgXQPIdUhXWFAQN8Jrd-gkTfbl1ZAR2HjikcsW0MZI2do5UoTezTx6aBHfDhSwsx0SOCKOXHdWmyvofJts71OggEykKwEqY0AH3CHBro5v88luav9avT6Ha3nrtoJQKwKBxYLkC5RVwbsw7L6Br%7EV14-DYVOr1EIrDRifwMfNjtbDrx5eoJiXJF4mpngng5zcaDIofe1LUprJ8oOglZbD8IwmUJyW6JiRw__&Key-Pair-Id=KVTP0A1DKRTAX [following]\r\n", + "--2023-09-02 06:16:32-- https://cdn-lfs.huggingface.co/repos/2e/f7/2ef78555202aa92abdbdf476ce3d0fd5a8b15f7245edf0b80d4d30572355f30d/20fb328bd798a1e23967f80661a0b6a277f1d45ed2d90760cec68d32dfcbd516?response-content-disposition=attachment%3B+filename*%3DUTF-8%27%27v5-L96-D1024-E0_1-mem-ctx-4k.pth%3B+filename%3D%22v5-L96-D1024-E0_1-mem-ctx-4k.pth%22%3B&Expires=1693894592&Policy=eyJTdGF0ZW1lbnQiOlt7IkNvbmRpdGlvbiI6eyJEYXRlTGVzc1RoYW4iOnsiQVdTOkVwb2NoVGltZSI6MTY5Mzg5NDU5Mn19LCJSZXNvdXJjZSI6Imh0dHBzOi8vY2RuLWxmcy5odWdnaW5nZmFjZS5jby9yZXBvcy8yZS9mNy8yZWY3ODU1NTIwMmFhOTJhYmRiZGY0NzZjZTNkMGZkNWE4YjE1ZjcyNDVlZGYwYjgwZDRkMzA1NzIzNTVmMzBkLzIwZmIzMjhiZDc5OGExZTIzOTY3ZjgwNjYxYTBiNmEyNzdmMWQ0NWVkMmQ5MDc2MGNlYzY4ZDMyZGZjYmQ1MTY%7EcmVzcG9uc2UtY29udGVudC1kaXNwb3NpdGlvbj0qIn1dfQ__&Signature=0lvghwVju7ugxwlMlvj%7EIeK4lhVAMO1bcdYQsP0FxGPt5AK88smuTfGJvprL1l1HiQVj4a7f6U8yvc97nQxIdWdTF0RC8s4SAyFcZoupxdMbzOQ7kEr7OgXQPIdUhXWFAQN8Jrd-gkTfbl1ZAR2HjikcsW0MZI2do5UoTezTx6aBHfDhSwsx0SOCKOXHdWmyvofJts71OggEykKwEqY0AH3CHBro5v88luav9avT6Ha3nrtoJQKwKBxYLkC5RVwbsw7L6Br%7EV14-DYVOr1EIrDRifwMfNjtbDrx5eoJiXJF4mpngng5zcaDIofe1LUprJ8oOglZbD8IwmUJyW6JiRw__&Key-Pair-Id=KVTP0A1DKRTAX\r\n", + "Resolving cdn-lfs.huggingface.co (cdn-lfs.huggingface.co)... " ] }, { "name": "stdout", "output_type": "stream", "text": [ + "108.156.22.119, 108.156.22.58, 108.156.22.7, ...\r\n", + "Connecting to cdn-lfs.huggingface.co (cdn-lfs.huggingface.co)|108.156.22.119|:443... connected.\r\n", "HTTP request sent, awaiting response... " ] }, @@ -238,7813 +251,34881 @@ "name": "stdout", "output_type": "stream", "text": [ - "404 Not Found\r\n", - "2023-09-01 14:53:10 ERROR 404: Not Found.\r\n", - "\r\n" + "200 OK\r\n", + "Length: 2825976699 (2.6G) [binary/octet-stream]\r\n", + "Saving to: ‘v5-L96-D1024-E0_1-mem-ctx-4k.pth’\r\n", + "\r\n", + "\r", + " v5-L96-D1 0%[ ] 0 --.-KB/s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "total 4.0K\r\n", - "drwxr-xr-x 2 root root 10 Sep 1 14:53 .\r\n", - "drwxr-xr-x 19 root root 4.0K Sep 1 14:53 ..\r\n" + "\r", + " v5-L96-D10 0%[ ] 128.27K 472KB/s " ] - } - ], - "source": [ - "# Download the model directly (stop gap till HF sync issues is resolved)\n", - "!cd \"{TRAINER_DIR}\" && cd \"../model/\" && \\\n", - " wget -nc \"https://huggingface.co/rwkv-x-dev/rwkv-x-playground/resolve/main/experiment/rwkv-x-exp/v5-memory/{FILENAME_PREFIX}-mem-ctx-4k.pth\"\n", - "\n", - "!cd \"{TRAINER_DIR}\" && cd \"../model/\" && \\\n", - " ls -alh ." - ] - }, - { - "cell_type": "markdown", - "id": "d36a2b7c", - "metadata": { - "papermill": { - "duration": 0.002214, - "end_time": "2023-09-01T14:53:10.839328", - "exception": false, - "start_time": "2023-09-01T14:53:10.837114", - "status": "completed" }, - "tags": [] - }, - "source": [ - "## Tune 6 : Ramping up the ctx size (8192), memory training\n", - "\n", - "- Tune 6: Large ctx size (8192), Scaling up!" - ] - }, - { - "cell_type": "code", - "execution_count": 5, - "id": "7a59bf6c", - "metadata": { - "execution": { - "iopub.execute_input": "2023-09-01T14:53:10.845705Z", - "iopub.status.busy": "2023-09-01T14:53:10.844902Z", - "iopub.status.idle": "2023-09-01T14:53:17.948023Z", - "shell.execute_reply": "2023-09-01T14:53:17.947224Z" + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5-L96-D102 0%[ ] 1.18M 2.49MB/s " + ] }, - "papermill": { - "duration": 7.16737, - "end_time": "2023-09-01T14:53:18.008993", - "exception": false, - "start_time": "2023-09-01T14:53:10.841623", - "status": "completed" + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5-L96-D1024 0%[ ] 3.65M 5.41MB/s " + ] }, - "tags": [] - }, - "outputs": [ { "name": "stdout", "output_type": "stream", "text": [ - "## Generating word reptition dataset ##\n" + "\r", + " v5-L96-D1024- 0%[ ] 10.10M 11.5MB/s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 88 samples (1 token repeat) - 30 max words - at ../dataset/shuffle-word-30-count.jsonl\n" + "\r", + " v5-L96-D1024-E 0%[ ] 20.48M 18.5MB/s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 49 samples (1 token repeat) - 50 max words - at ../dataset/shuffle-word-50-count.jsonl\n" + "\r", + " v5-L96-D1024-E0 1%[ ] 28.53M 21.6MB/s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated JSONL file with - 20 max words, 50 samples - at ../dataset/gen-word-20-count.jsonl\n" + "\r", + " v5-L96-D1024-E0_ 1%[ ] 38.20M 25.0MB/s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated JSONL file with - 2 max words, 50 samples - at ../dataset/word-2-count.jsonl\n" + "\r", + " v5-L96-D1024-E0_1 1%[ ] 45.14M 26.1MB/s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated JSONL file with - 5 max words, 50 samples - at ../dataset/gen-word-5-count.jsonl\n" + "\r", + " v5-L96-D1024-E0_1- 1%[ ] 53.31M 27.3MB/s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated JSONL file with - 15 max words, 50 samples - at ../dataset/gen-word-15-count.jsonl\n" + "\r", + "v5-L96-D1024-E0_1-m 2%[ ] 61.80M 28.7MB/s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated JSONL file with - 10 max words, 50 samples - at ../dataset/gen-word-10-count.jsonl\n" + "\r", + "5-L96-D1024-E0_1-me 2%[ ] 68.40M 29.1MB/s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated JSONL file with - 40 max words, 50 samples - at ../dataset/gen-word-40-count.jsonl\n" + "\r", + "-L96-D1024-E0_1-mem 2%[ ] 78.76M 30.8MB/s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated JSONL file with - 25 max words, 50 samples - at ../dataset/gen-word-25-count.jsonl\n" + "\r", + "L96-D1024-E0_1-mem- 3%[ ] 85.85M 31.1MB/s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated JSONL file with - 35 max words, 50 samples - at ../dataset/gen-word-35-count.jsonl\n" + "\r", + "96-D1024-E0_1-mem-c 3%[ ] 95.06M 31.9MB/s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated JSONL file with - 55 max words, 50 samples - at ../dataset/gen-word-55-count.jsonl\n" + "\r", + "6-D1024-E0_1-mem-ct 3%[ ] 103.00M 32.3MB/s eta 80s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 48 samples (1 token repeat) - 55 max words - at ../dataset/shuffle-word-55-count.jsonl\n" + "\r", + "-D1024-E0_1-mem-ctx 4%[ ] 110.60M 32.4MB/s eta 80s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 42 samples (1 token repeat) - 60 max words - at ../dataset/shuffle-word-60-count.jsonl\n" + "\r", + "D1024-E0_1-mem-ctx- 4%[ ] 121.50M 36.9MB/s eta 80s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 31 samples (1 token repeat) - 80 max words - at ../dataset/shuffle-word-80-count.jsonl\n" + "\r", + "1024-E0_1-mem-ctx-4 4%[ ] 130.07M 39.0MB/s eta 80s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated JSONL file with - 45 max words, 50 samples - at ../dataset/gen-word-45-count.jsonl\n" + "\r", + "024-E0_1-mem-ctx-4k 5%[> ] 137.57M 39.9MB/s eta 80s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 108 samples (1 token repeat) - 25 max words - at ../dataset/shuffle-word-25-count.jsonl\n" + "\r", + "24-E0_1-mem-ctx-4k. 5%[> ] 147.07M 40.5MB/s eta 74s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 37 samples (1 token repeat) - 70 max words - at ../dataset/shuffle-word-70-count.jsonl\n" + "\r", + "4-E0_1-mem-ctx-4k.p 5%[> ] 156.38M 40.1MB/s eta 74s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 58 samples (1 token repeat) - 45 max words - at ../dataset/shuffle-word-45-count.jsonl\n" + "\r", + "-E0_1-mem-ctx-4k.pt 6%[> ] 164.59M 40.0MB/s eta 74s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated JSONL file with - 85 max words, 50 samples - at ../dataset/gen-word-85-count.jsonl\n" + "\r", + "E0_1-mem-ctx-4k.pth 6%[> ] 171.63M 39.2MB/s eta 74s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated JSONL file with - 65 max words, 50 samples - at ../dataset/gen-word-65-count.jsonl\n" + "\r", + "0_1-mem-ctx-4k.pth 6%[> ] 180.27M 40.0MB/s eta 74s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 63 samples (1 token repeat) - 40 max words - at ../dataset/shuffle-word-40-count.jsonl\n" + "\r", + "_1-mem-ctx-4k.pth 6%[> ] 187.18M 39.2MB/s eta 71s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated JSONL file with - 130 max words, 50 samples - at ../dataset/gen-word-130-count.jsonl\n" + "\r", + "1-mem-ctx-4k.pth 7%[> ] 193.83M 39.0MB/s eta 71s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 177 samples (1 token repeat) - 15 max words - at ../dataset/shuffle-word-15-count.jsonl\n" + "\r", + "-mem-ctx-4k.pth 7%[> ] 202.32M 39.3MB/s eta 71s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 134 samples (1 token repeat) - 20 max words - at ../dataset/shuffle-word-20-count.jsonl\n" + "\r", + "mem-ctx-4k.pth 7%[> ] 210.41M 39.0MB/s eta 71s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 77 samples (1 token repeat) - 35 max words - at ../dataset/shuffle-word-35-count.jsonl\n" + "\r", + "em-ctx-4k.pth 8%[> ] 219.40M 39.4MB/s eta 71s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 44 samples (1 token repeat) - 65 max words - at ../dataset/shuffle-word-65-count.jsonl\n" + "\r", + "m-ctx-4k.pth 8%[> ] 226.92M 39.2MB/s eta 69s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated JSONL file with - 60 max words, 50 samples - at ../dataset/gen-word-60-count.jsonl\n" + "\r", + "-ctx-4k.pth 8%[> ] 235.87M 39.2MB/s eta 69s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated JSONL file with - 30 max words, 50 samples - at ../dataset/gen-word-30-count.jsonl\n" + "\r", + "ctx-4k.pth 9%[> ] 244.37M 39.2MB/s eta 69s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated JSONL file with - 50 max words, 50 samples - at ../dataset/gen-word-50-count.jsonl\n" + "\r", + "tx-4k.pth 9%[> ] 252.94M 39.4MB/s eta 69s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 28 samples (1 token repeat) - 95 max words - at ../dataset/shuffle-word-95-count.jsonl\n" + "\r", + "x-4k.pth 9%[> ] 260.46M 39.1MB/s eta 69s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated JSONL file with - 90 max words, 50 samples - at ../dataset/gen-word-90-count.jsonl\n" + "\r", + "-4k.pth 9%[> ] 268.38M 39.0MB/s eta 67s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated JSONL file with - 70 max words, 50 samples - at ../dataset/gen-word-70-count.jsonl\n" + "\r", + "4k.pth 10%[=> ] 278.50M 39.2MB/s eta 67s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 38 samples (1 token repeat) - 75 max words - at ../dataset/shuffle-word-75-count.jsonl\n" + "\r", + "k.pth 10%[=> ] 286.82M 39.2MB/s eta 67s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 17 samples (1 token repeat) - 125 max words - at ../dataset/shuffle-word-125-count.jsonl\n" + "\r", + ".pth 10%[=> ] 294.98M 39.1MB/s eta 67s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated JSONL file with - 75 max words, 50 samples - at ../dataset/gen-word-75-count.jsonl\n" + "\r", + "pth 11%[=> ] 303.99M 40.1MB/s eta 67s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated JSONL file with - 80 max words, 50 samples - at ../dataset/gen-word-80-count.jsonl\n" + "\r", + "th 11%[=> ] 312.18M 39.9MB/s eta 64s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated JSONL file with - 95 max words, 50 samples - at ../dataset/gen-word-95-count.jsonl\n" + "\r", + "h 11%[=> ] 320.37M 40.7MB/s eta 64s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 20 samples (1 token repeat) - 110 max words - at ../dataset/shuffle-word-110-count.jsonl\n" + "\r", + " 12%[=> ] 327.41M 40.2MB/s eta 64s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 560 samples (1 token repeat) - 5 max words - at ../dataset/shuffle-word-5-count.jsonl\n" + "\r", + " v 12%[=> ] 334.85M 40.1MB/s eta 64s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 19 samples (1 token repeat) - 115 max words - at ../dataset/shuffle-word-115-count.jsonl\n" + "\r", + " v5 12%[=> ] 343.29M 40.2MB/s eta 64s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 258 samples (1 token repeat) - 10 max words - at ../dataset/shuffle-word-10-count.jsonl\n" + "\r", + " v5- 12%[=> ] 350.02M 39.8MB/s eta 63s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 17 samples (1 token repeat) - 130 max words - at ../dataset/shuffle-word-130-count.jsonl\n" + "\r", + " v5-L 13%[=> ] 359.43M 40.0MB/s eta 63s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated JSONL file with - 105 max words, 50 samples - at ../dataset/gen-word-105-count.jsonl\n" + "\r", + " v5-L9 13%[=> ] 366.85M 39.7MB/s eta 63s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated JSONL file with - 120 max words, 50 samples - at ../dataset/gen-word-120-count.jsonl\n" + "\r", + " v5-L96 13%[=> ] 374.94M 39.6MB/s eta 63s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated JSONL file with - 110 max words, 50 samples - at ../dataset/gen-word-110-count.jsonl\n" + "\r", + " v5-L96- 14%[=> ] 383.94M 39.8MB/s eta 63s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated JSONL file with - 100 max words, 50 samples - at ../dataset/gen-word-100-count.jsonl\n" + "\r", + " v5-L96-D 14%[=> ] 389.22M 39.0MB/s eta 62s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 32 samples (1 token repeat) - 90 max words - at ../dataset/shuffle-word-90-count.jsonl\n" + "\r", + " v5-L96-D1 14%[=> ] 399.01M 39.5MB/s eta 62s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 31 samples (1 token repeat) - 85 max words - at ../dataset/shuffle-word-85-count.jsonl\n" + "\r", + " v5-L96-D10 14%[=> ] 400.87M 37.2MB/s eta 62s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated JSONL file with - 140 max words, 50 samples - at ../dataset/gen-word-140-count.jsonl\n" + "\r", + " v5-L96-D102 15%[==> ] 409.16M 37.3MB/s eta 62s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 16 samples (1 token repeat) - 150 max words - at ../dataset/shuffle-word-150-count.jsonl\n" + "\r", + " v5-L96-D1024 15%[==> ] 422.11M 38.5MB/s eta 62s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 14 samples (1 token repeat) - 165 max words - at ../dataset/shuffle-word-165-count.jsonl\n" + "\r", + " v5-L96-D1024- 15%[==> ] 430.96M 38.8MB/s eta 60s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated JSONL file with - 180 max words, 50 samples - at ../dataset/gen-word-180-count.jsonl\n" + "\r", + " v5-L96-D1024-E 16%[==> ] 438.90M 38.4MB/s eta 60s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 16 samples (1 token repeat) - 145 max words - at ../dataset/shuffle-word-145-count.jsonl\n" + "\r", + " v5-L96-D1024-E0 16%[==> ] 448.99M 39.2MB/s eta 60s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 19 samples (1 token repeat) - 120 max words - at ../dataset/shuffle-word-120-count.jsonl\n" + "\r", + " v5-L96-D1024-E0_ 16%[==> ] 454.42M 38.5MB/s eta 60s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 20 samples (1 token repeat) - 105 max words - at ../dataset/shuffle-word-105-count.jsonl\n" + "\r", + " v5-L96-D1024-E0_1 17%[==> ] 463.45M 39.3MB/s eta 60s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 27 samples (1 token repeat) - 100 max words - at ../dataset/shuffle-word-100-count.jsonl\n" + "\r", + " v5-L96-D1024-E0_1- 17%[==> ] 472.14M 39.5MB/s eta 59s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated JSONL file with - 135 max words, 50 samples - at ../dataset/gen-word-135-count.jsonl\n" + "\r", + "v5-L96-D1024-E0_1-m 17%[==> ] 480.86M 39.7MB/s eta 59s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated JSONL file with - 115 max words, 50 samples - at ../dataset/gen-word-115-count.jsonl\n" + "\r", + "5-L96-D1024-E0_1-me 18%[==> ] 488.78M 39.6MB/s eta 59s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 5 samples (1 token repeat) - 565 max words - at ../dataset/shuffle-word-565-count.jsonl\n" + "\r", + "-L96-D1024-E0_1-mem 18%[==> ] 498.34M 40.4MB/s eta 59s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 17 samples (1 token repeat) - 135 max words - at ../dataset/shuffle-word-135-count.jsonl\n" + "\r", + "L96-D1024-E0_1-mem- 18%[==> ] 507.16M 40.0MB/s eta 59s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated JSONL file with - 165 max words, 50 samples - at ../dataset/gen-word-165-count.jsonl\n" + "\r", + "96-D1024-E0_1-mem-c 19%[==> ] 515.72M 40.0MB/s eta 57s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 11 samples (1 token repeat) - 205 max words - at ../dataset/shuffle-word-205-count.jsonl\n" + "\r", + "6-D1024-E0_1-mem-ct 19%[==> ] 523.23M 40.3MB/s eta 57s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated JSONL file with - 145 max words, 50 samples - at ../dataset/gen-word-145-count.jsonl\n" + "\r", + "-D1024-E0_1-mem-ctx 19%[==> ] 530.88M 40.2MB/s eta 57s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated JSONL file with - 305 max words, 50 samples - at ../dataset/gen-word-305-count.jsonl\n" + "\r", + "D1024-E0_1-mem-ctx- 20%[===> ] 541.15M 42.6MB/s eta 57s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated JSONL file with - 445 max words, 50 samples - at ../dataset/gen-word-445-count.jsonl\n" + "\r", + "1024-E0_1-mem-ctx-4 20%[===> ] 549.61M 40.9MB/s eta 57s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 15 samples (1 token repeat) - 155 max words - at ../dataset/shuffle-word-155-count.jsonl\n" + "\r", + "024-E0_1-mem-ctx-4k 20%[===> ] 557.83M 40.4MB/s eta 56s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 5 samples (1 token repeat) - 585 max words - at ../dataset/shuffle-word-585-count.jsonl\n" + "\r", + "24-E0_1-mem-ctx-4k. 21%[===> ] 566.19M 40.5MB/s eta 56s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 5 samples (1 token repeat) - 570 max words - at ../dataset/shuffle-word-570-count.jsonl\n" + "\r", + "4-E0_1-mem-ctx-4k.p 21%[===> ] 576.30M 40.9MB/s eta 56s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 5 samples (1 token repeat) - 520 max words - at ../dataset/shuffle-word-520-count.jsonl\n" + "\r", + "-E0_1-mem-ctx-4k.pt 21%[===> ] 584.55M 40.6MB/s eta 56s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 9 samples (1 token repeat) - 255 max words - at ../dataset/shuffle-word-255-count.jsonl\n" + "\r", + "E0_1-mem-ctx-4k.pth 22%[===> ] 593.19M 40.8MB/s eta 56s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated JSONL file with - 150 max words, 50 samples - at ../dataset/gen-word-150-count.jsonl\n" + "\r", + "0_1-mem-ctx-4k.pth 22%[===> ] 601.02M 41.2MB/s eta 55s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated JSONL file with - 155 max words, 50 samples - at ../dataset/gen-word-155-count.jsonl\n" + "\r", + "_1-mem-ctx-4k.pth 22%[===> ] 607.94M 40.3MB/s eta 55s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 14 samples (1 token repeat) - 175 max words - at ../dataset/shuffle-word-175-count.jsonl\n" + "\r", + "1-mem-ctx-4k.pth 22%[===> ] 617.53M 40.9MB/s eta 55s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 4 samples (1 token repeat) - 625 max words - at ../dataset/shuffle-word-625-count.jsonl\n" + "\r", + "-mem-ctx-4k.pth 23%[===> ] 626.19M 41.1MB/s eta 55s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 7 samples (1 token repeat) - 380 max words - at ../dataset/shuffle-word-380-count.jsonl\n" + "\r", + "mem-ctx-4k.pth 23%[===> ] 631.94M 40.2MB/s eta 55s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 6 samples (1 token repeat) - 470 max words - at ../dataset/shuffle-word-470-count.jsonl\n" + "\r", + "em-ctx-4k.pth 23%[===> ] 642.25M 40.8MB/s eta 53s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 9 samples (1 token repeat) - 280 max words - at ../dataset/shuffle-word-280-count.jsonl\n" + "\r", + "m-ctx-4k.pth 24%[===> ] 650.46M 40.2MB/s eta 53s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 15 samples (1 token repeat) - 170 max words - at ../dataset/shuffle-word-170-count.jsonl\n" + "\r", + "-ctx-4k.pth 24%[===> ] 660.40M 41.1MB/s eta 53s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 15 samples (1 token repeat) - 160 max words - at ../dataset/shuffle-word-160-count.jsonl\n" + "\r", + "ctx-4k.pth 24%[===> ] 669.04M 41.3MB/s eta 53s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated JSONL file with - 535 max words, 50 samples - at ../dataset/gen-word-535-count.jsonl\n" + "\r", + "tx-4k.pth 25%[====> ] 677.22M 40.6MB/s eta 53s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated JSONL file with - 745 max words, 50 samples - at ../dataset/gen-word-745-count.jsonl\n" + "\r", + "x-4k.pth 25%[====> ] 685.40M 40.8MB/s eta 52s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated JSONL file with - 490 max words, 50 samples - at ../dataset/gen-word-490-count.jsonl\n" + "\r", + "-4k.pth 25%[====> ] 692.90M 40.2MB/s eta 52s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 5 samples (1 token repeat) - 535 max words - at ../dataset/shuffle-word-535-count.jsonl\n" + "\r", + "4k.pth 26%[====> ] 701.72M 40.5MB/s eta 52s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 6 samples (1 token repeat) - 405 max words - at ../dataset/shuffle-word-405-count.jsonl\n" + "\r", + "k.pth 26%[====> ] 710.47M 40.8MB/s eta 52s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 6 samples (1 token repeat) - 475 max words - at ../dataset/shuffle-word-475-count.jsonl\n" + "\r", + ".pth 26%[====> ] 719.50M 40.4MB/s eta 52s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 3 samples (1 token repeat) - 920 max words - at ../dataset/shuffle-word-920-count.jsonl\n" + "\r", + "pth 26%[====> ] 727.67M 40.6MB/s eta 51s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated JSONL file with - 260 max words, 50 samples - at ../dataset/gen-word-260-count.jsonl\n" + "\r", + "th 27%[====> ] 735.40M 40.5MB/s eta 51s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated JSONL file with - 405 max words, 50 samples - at ../dataset/gen-word-405-count.jsonl\n" + "\r", + "h 27%[====> ] 745.58M 41.2MB/s eta 51s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 4 samples (1 token repeat) - 690 max words - at ../dataset/shuffle-word-690-count.jsonl\n" + "\r", + " 27%[====> ] 754.27M 41.0MB/s eta 51s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 4 samples (1 token repeat) - 770 max words - at ../dataset/shuffle-word-770-count.jsonl\n" + "\r", + " v 28%[====> ] 762.58M 40.9MB/s eta 51s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 6 samples (1 token repeat) - 455 max words - at ../dataset/shuffle-word-455-count.jsonl\n" + "\r", + " v5 28%[====> ] 770.36M 40.8MB/s eta 50s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 3 samples (1 token repeat) - 950 max words - at ../dataset/shuffle-word-950-count.jsonl\n" + "\r", + " v5- 28%[====> ] 778.11M 40.5MB/s eta 50s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated JSONL file with - 240 max words, 50 samples - at ../dataset/gen-word-240-count.jsonl\n" + "\r", + " v5-L 29%[====> ] 787.60M 40.9MB/s eta 50s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated JSONL file with - 555 max words, 50 samples - at ../dataset/gen-word-555-count.jsonl\n" + "\r", + " v5-L9 29%[====> ] 795.51M 40.5MB/s eta 50s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 7 samples (1 token repeat) - 365 max words - at ../dataset/shuffle-word-365-count.jsonl\n" + "\r", + " v5-L96 29%[====> ] 804.15M 40.9MB/s eta 50s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 3 samples (1 token repeat) - 820 max words - at ../dataset/shuffle-word-820-count.jsonl\n" + "\r", + " v5-L96- 30%[=====> ] 811.94M 40.2MB/s eta 48s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 10 samples (1 token repeat) - 215 max words - at ../dataset/shuffle-word-215-count.jsonl\n" + "\r", + " v5-L96-D 30%[=====> ] 819.92M 40.3MB/s eta 48s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 5 samples (1 token repeat) - 555 max words - at ../dataset/shuffle-word-555-count.jsonl\n" + "\r", + " v5-L96-D1 30%[=====> ] 828.12M 40.4MB/s eta 48s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated JSONL file with - 215 max words, 50 samples - at ../dataset/gen-word-215-count.jsonl\n" + "\r", + " v5-L96-D10 31%[=====> ] 837.98M 40.4MB/s eta 48s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated JSONL file with - 175 max words, 50 samples - at ../dataset/gen-word-175-count.jsonl\n" + "\r", + " v5-L96-D102 31%[=====> ] 846.44M 40.3MB/s eta 48s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated JSONL file with - 480 max words, 50 samples - at ../dataset/gen-word-480-count.jsonl\n" + "\r", + " v5-L96-D1024 31%[=====> ] 855.32M 40.7MB/s eta 47s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 4 samples (1 token repeat) - 730 max words - at ../dataset/shuffle-word-730-count.jsonl\n" + "\r", + " v5-L96-D1024- 32%[=====> ] 863.40M 40.7MB/s eta 47s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 20 samples (20 token repeat) - 3700 max words - at ../dataset/shuffle-word-3700-count.jsonl\n" + "\r", + " v5-L96-D1024-E 32%[=====> ] 871.21M 40.2MB/s eta 47s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated JSONL file with - 340 max words, 50 samples - at ../dataset/gen-word-340-count.jsonl\n" + "\r", + " v5-L96-D1024-E0 32%[=====> ] 880.97M 41.1MB/s eta 47s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 4 samples (1 token repeat) - 710 max words - at ../dataset/shuffle-word-710-count.jsonl\n" + "\r", + " v5-L96-D1024-E0_ 32%[=====> ] 889.22M 40.5MB/s eta 47s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated JSONL file with - 890 max words, 50 samples - at ../dataset/gen-word-890-count.jsonl\n" + "\r", + " v5-L96-D1024-E0_1 33%[=====> ] 897.03M 40.2MB/s eta 46s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 10 samples (1 token repeat) - 250 max words - at ../dataset/shuffle-word-250-count.jsonl\n" + "\r", + " v5-L96-D1024-E0_1- 33%[=====> ] 898.93M 38.4MB/s eta 46s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 7 samples (1 token repeat) - 370 max words - at ../dataset/shuffle-word-370-count.jsonl\n" + "\r", + "v5-L96-D1024-E0_1-m 33%[=====> ] 915.07M 40.6MB/s eta 46s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 9 samples (1 token repeat) - 295 max words - at ../dataset/shuffle-word-295-count.jsonl\n" + "\r", + "5-L96-D1024-E0_1-me 34%[=====> ] 924.14M 40.6MB/s eta 46s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated JSONL file with - 230 max words, 50 samples - at ../dataset/gen-word-230-count.jsonl\n" + "\r", + "-L96-D1024-E0_1-mem 34%[=====> ] 932.36M 40.8MB/s eta 46s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 6 samples (1 token repeat) - 430 max words - at ../dataset/shuffle-word-430-count.jsonl\n" + "\r", + "L96-D1024-E0_1-mem- 34%[=====> ] 941.37M 40.7MB/s eta 45s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated JSONL file with - 695 max words, 50 samples - at ../dataset/gen-word-695-count.jsonl\n" + "\r", + "96-D1024-E0_1-mem-c 35%[======> ] 949.66M 40.9MB/s eta 45s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 9 samples (1 token repeat) - 270 max words - at ../dataset/shuffle-word-270-count.jsonl\n" + "\r", + "6-D1024-E0_1-mem-ct 35%[======> ] 958.50M 40.8MB/s eta 45s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 3 samples (1 token repeat) - 865 max words - at ../dataset/shuffle-word-865-count.jsonl\n" + "\r", + "-D1024-E0_1-mem-ctx 35%[======> ] 966.79M 40.7MB/s eta 45s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 5 samples (1 token repeat) - 560 max words - at ../dataset/shuffle-word-560-count.jsonl\n" + "\r", + "D1024-E0_1-mem-ctx- 36%[======> ] 976.33M 41.2MB/s eta 45s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 4 samples (1 token repeat) - 665 max words - at ../dataset/shuffle-word-665-count.jsonl\n" + "\r", + "1024-E0_1-mem-ctx-4 36%[======> ] 985.30M 41.5MB/s eta 44s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 5 samples (1 token repeat) - 550 max words - at ../dataset/shuffle-word-550-count.jsonl\n" + "\r", + "024-E0_1-mem-ctx-4k 36%[======> ] 993.33M 41.0MB/s eta 44s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 4 samples (1 token repeat) - 735 max words - at ../dataset/shuffle-word-735-count.jsonl\n" + "\r", + "24-E0_1-mem-ctx-4k. 37%[======> ] 1002M 41.1MB/s eta 44s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated JSONL file with - 645 max words, 50 samples - at ../dataset/gen-word-645-count.jsonl\n" + "\r", + "4-E0_1-mem-ctx-4k.p 37%[======> ] 1011M 41.0MB/s eta 44s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 4 samples (1 token repeat) - 755 max words - at ../dataset/shuffle-word-755-count.jsonl\n" + "\r", + "-E0_1-mem-ctx-4k.pt 37%[======> ] 1019M 40.9MB/s eta 44s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 6 samples (1 token repeat) - 485 max words - at ../dataset/shuffle-word-485-count.jsonl\n" + "\r", + "E0_1-mem-ctx-4k.pth 38%[======> ] 1.00G 40.9MB/s eta 42s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 6 samples (1 token repeat) - 440 max words - at ../dataset/shuffle-word-440-count.jsonl\n" + "\r", + "0_1-mem-ctx-4k.pth 38%[======> ] 1.01G 42.5MB/s eta 42s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated JSONL file with - 255 max words, 50 samples - at ../dataset/gen-word-255-count.jsonl\n" + "\r", + "_1-mem-ctx-4k.pth 38%[======> ] 1.02G 41.9MB/s eta 42s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated JSONL file with - 610 max words, 50 samples - at ../dataset/gen-word-610-count.jsonl\n" + "\r", + "1-mem-ctx-4k.pth 39%[======> ] 1.03G 40.6MB/s eta 42s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated JSONL file with - 860 max words, 50 samples - at ../dataset/gen-word-860-count.jsonl\n" + "\r", + "-mem-ctx-4k.pth 39%[======> ] 1.04G 40.4MB/s eta 42s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated JSONL file with - 320 max words, 50 samples - at ../dataset/gen-word-320-count.jsonl\n" + "\r", + "mem-ctx-4k.pth 39%[======> ] 1.04G 39.5MB/s eta 41s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated JSONL file with - 390 max words, 50 samples - at ../dataset/gen-word-390-count.jsonl\n" + "\r", + "em-ctx-4k.pth 40%[=======> ] 1.05G 40.4MB/s eta 41s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 6 samples (1 token repeat) - 415 max words - at ../dataset/shuffle-word-415-count.jsonl\n" + "\r", + "m-ctx-4k.pth 40%[=======> ] 1.06G 40.0MB/s eta 41s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 9 samples (1 token repeat) - 260 max words - at ../dataset/shuffle-word-260-count.jsonl\n" + "\r", + "-ctx-4k.pth 40%[=======> ] 1.07G 40.5MB/s eta 41s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated JSONL file with - 925 max words, 50 samples - at ../dataset/gen-word-925-count.jsonl\n" + "\r", + "ctx-4k.pth 40%[=======> ] 1.08G 40.1MB/s eta 41s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated a single JSONL file with 7 samples (1 token repeat) - 345 max words - at ../dataset/shuffle-word-345-count.jsonl\n" + "\r", + "tx-4k.pth 41%[=======> ] 1.09G 40.8MB/s eta 40s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated JSONL file with - 675 max words, 50 samples - at ../dataset/gen-word-675-count.jsonl\n" + "\r", + "x-4k.pth 41%[=======> ] 1.09G 40.6MB/s eta 40s " ] }, { "name": "stdout", "output_type": "stream", "text": [ - "Generated JSONL file with - 915 max words, 50 samples - at ../dataset/gen-word-915-count.jsonl\n" + "\r", + "-4k.pth 41%[=======> ] 1.10G 40.2MB/s eta 40s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "4k.pth 42%[=======> ] 1.11G 40.0MB/s eta 40s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "k.pth 42%[=======> ] 1.12G 39.8MB/s eta 40s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + ".pth 42%[=======> ] 1.12G 39.2MB/s eta 39s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "pth 43%[=======> ] 1.13G 39.1MB/s eta 39s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "th 43%[=======> ] 1.14G 38.9MB/s eta 39s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "h 43%[=======> ] 1.15G 39.3MB/s eta 39s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " 44%[=======> ] 1.16G 38.9MB/s eta 39s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v 44%[=======> ] 1.17G 39.4MB/s eta 38s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5 44%[=======> ] 1.17G 39.8MB/s eta 38s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5- 44%[=======> ] 1.18G 39.8MB/s eta 38s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5-L 45%[========> ] 1.19G 39.6MB/s eta 38s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5-L9 45%[========> ] 1.20G 39.7MB/s eta 38s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5-L96 45%[========> ] 1.21G 39.9MB/s eta 37s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5-L96- 46%[========> ] 1.22G 39.7MB/s eta 37s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5-L96-D 46%[========> ] 1.23G 39.2MB/s eta 37s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5-L96-D1 46%[========> ] 1.23G 39.5MB/s eta 37s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5-L96-D10 47%[========> ] 1.24G 39.9MB/s eta 37s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5-L96-D102 47%[========> ] 1.25G 39.6MB/s eta 36s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5-L96-D1024 47%[========> ] 1.26G 39.9MB/s eta 36s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5-L96-D1024- 48%[========> ] 1.27G 40.6MB/s eta 36s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5-L96-D1024-E 48%[========> ] 1.27G 40.1MB/s eta 36s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5-L96-D1024-E0 48%[========> ] 1.28G 40.5MB/s eta 36s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5-L96-D1024-E0_ 49%[========> ] 1.29G 40.5MB/s eta 35s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5-L96-D1024-E0_1 49%[========> ] 1.30G 40.6MB/s eta 35s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5-L96-D1024-E0_1- 49%[========> ] 1.31G 40.0MB/s eta 35s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "v5-L96-D1024-E0_1-m 49%[========> ] 1.31G 37.9MB/s eta 35s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "5-L96-D1024-E0_1-me 49%[========> ] 1.31G 37.3MB/s eta 35s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "-L96-D1024-E0_1-mem 50%[=========> ] 1.33G 39.3MB/s eta 34s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "L96-D1024-E0_1-mem- 50%[=========> ] 1.34G 38.9MB/s eta 34s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "96-D1024-E0_1-mem-c 51%[=========> ] 1.34G 38.8MB/s eta 34s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "6-D1024-E0_1-mem-ct 51%[=========> ] 1.35G 39.0MB/s eta 34s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "-D1024-E0_1-mem-ctx 51%[=========> ] 1.36G 39.1MB/s eta 34s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "D1024-E0_1-mem-ctx- 52%[=========> ] 1.37G 38.5MB/s eta 33s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "1024-E0_1-mem-ctx-4 52%[=========> ] 1.38G 39.2MB/s eta 33s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "024-E0_1-mem-ctx-4k 52%[=========> ] 1.39G 39.4MB/s eta 33s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "24-E0_1-mem-ctx-4k. 53%[=========> ] 1.40G 39.5MB/s eta 33s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "4-E0_1-mem-ctx-4k.p 53%[=========> ] 1.40G 39.4MB/s eta 33s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "-E0_1-mem-ctx-4k.pt 53%[=========> ] 1.41G 39.4MB/s eta 32s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "E0_1-mem-ctx-4k.pth 53%[=========> ] 1.42G 39.0MB/s eta 32s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "0_1-mem-ctx-4k.pth 54%[=========> ] 1.43G 40.3MB/s eta 32s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "_1-mem-ctx-4k.pth 54%[=========> ] 1.44G 40.6MB/s eta 32s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "1-mem-ctx-4k.pth 54%[=========> ] 1.44G 42.8MB/s eta 32s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "-mem-ctx-4k.pth 55%[==========> ] 1.45G 42.5MB/s eta 31s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "mem-ctx-4k.pth 55%[==========> ] 1.46G 40.5MB/s eta 31s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "em-ctx-4k.pth 55%[==========> ] 1.47G 40.7MB/s eta 31s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "m-ctx-4k.pth 56%[==========> ] 1.48G 40.5MB/s eta 31s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "-ctx-4k.pth 56%[==========> ] 1.48G 39.5MB/s eta 31s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "ctx-4k.pth 56%[==========> ] 1.49G 39.4MB/s eta 30s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "tx-4k.pth 57%[==========> ] 1.50G 40.0MB/s eta 30s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "x-4k.pth 57%[==========> ] 1.51G 39.8MB/s eta 30s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "-4k.pth 57%[==========> ] 1.52G 39.9MB/s eta 30s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "4k.pth 57%[==========> ] 1.53G 39.6MB/s eta 30s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "k.pth 58%[==========> ] 1.54G 40.3MB/s eta 28s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + ".pth 58%[==========> ] 1.54G 39.8MB/s eta 28s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "pth 58%[==========> ] 1.55G 40.0MB/s eta 28s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "th 59%[==========> ] 1.56G 39.6MB/s eta 28s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "h 59%[==========> ] 1.57G 39.9MB/s eta 28s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " 59%[==========> ] 1.58G 39.5MB/s eta 27s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v 60%[===========> ] 1.58G 39.8MB/s eta 27s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5 60%[===========> ] 1.59G 39.8MB/s eta 27s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5- 60%[===========> ] 1.60G 39.9MB/s eta 27s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5-L 61%[===========> ] 1.61G 40.0MB/s eta 27s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5-L9 61%[===========> ] 1.62G 40.5MB/s eta 26s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5-L96 61%[===========> ] 1.63G 40.7MB/s eta 26s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5-L96- 62%[===========> ] 1.63G 40.6MB/s eta 26s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5-L96-D 62%[===========> ] 1.64G 41.2MB/s eta 26s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5-L96-D1 62%[===========> ] 1.65G 41.0MB/s eta 26s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5-L96-D10 63%[===========> ] 1.66G 40.7MB/s eta 25s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5-L96-D102 63%[===========> ] 1.67G 40.7MB/s eta 25s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5-L96-D1024 63%[===========> ] 1.67G 40.2MB/s eta 25s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5-L96-D1024- 63%[===========> ] 1.68G 40.6MB/s eta 25s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5-L96-D1024-E 64%[===========> ] 1.69G 40.3MB/s eta 25s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5-L96-D1024-E0 64%[===========> ] 1.70G 40.5MB/s eta 24s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5-L96-D1024-E0_ 64%[===========> ] 1.71G 40.5MB/s eta 24s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5-L96-D1024-E0_1 65%[============> ] 1.72G 40.9MB/s eta 24s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5-L96-D1024-E0_1- 65%[============> ] 1.73G 40.6MB/s eta 24s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "v5-L96-D1024-E0_1-m 65%[============> ] 1.73G 40.4MB/s eta 24s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "5-L96-D1024-E0_1-me 66%[============> ] 1.74G 39.9MB/s eta 23s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "-L96-D1024-E0_1-mem 66%[============> ] 1.75G 39.8MB/s eta 23s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "L96-D1024-E0_1-mem- 66%[============> ] 1.75G 37.0MB/s eta 23s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "96-D1024-E0_1-mem-c 67%[============> ] 1.76G 39.3MB/s eta 23s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "6-D1024-E0_1-mem-ct 67%[============> ] 1.77G 39.7MB/s eta 23s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "-D1024-E0_1-mem-ctx 67%[============> ] 1.78G 38.8MB/s eta 22s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "D1024-E0_1-mem-ctx- 67%[============> ] 1.79G 38.9MB/s eta 22s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "1024-E0_1-mem-ctx-4 68%[============> ] 1.79G 38.7MB/s eta 22s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "024-E0_1-mem-ctx-4k 68%[============> ] 1.81G 39.0MB/s eta 22s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "24-E0_1-mem-ctx-4k. 68%[============> ] 1.81G 39.4MB/s eta 22s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "4-E0_1-mem-ctx-4k.p 69%[============> ] 1.82G 39.3MB/s eta 21s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "-E0_1-mem-ctx-4k.pt 69%[============> ] 1.83G 38.9MB/s eta 21s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "E0_1-mem-ctx-4k.pth 69%[============> ] 1.84G 39.5MB/s eta 21s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "0_1-mem-ctx-4k.pth 70%[=============> ] 1.85G 38.9MB/s eta 21s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "_1-mem-ctx-4k.pth 70%[=============> ] 1.85G 39.1MB/s eta 21s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "1-mem-ctx-4k.pth 70%[=============> ] 1.86G 38.3MB/s eta 20s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "-mem-ctx-4k.pth 70%[=============> ] 1.87G 38.7MB/s eta 20s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "mem-ctx-4k.pth 71%[=============> ] 1.87G 38.1MB/s eta 20s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "em-ctx-4k.pth 71%[=============> ] 1.88G 38.3MB/s eta 20s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "m-ctx-4k.pth 71%[=============> ] 1.89G 39.9MB/s eta 20s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "-ctx-4k.pth 72%[=============> ] 1.90G 38.3MB/s eta 19s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "ctx-4k.pth 72%[=============> ] 1.91G 38.7MB/s eta 19s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "tx-4k.pth 72%[=============> ] 1.91G 38.6MB/s eta 19s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "x-4k.pth 73%[=============> ] 1.92G 39.5MB/s eta 19s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "-4k.pth 73%[=============> ] 1.93G 39.2MB/s eta 19s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "4k.pth 73%[=============> ] 1.94G 38.9MB/s eta 18s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "k.pth 74%[=============> ] 1.95G 39.1MB/s eta 18s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + ".pth 74%[=============> ] 1.96G 39.1MB/s eta 18s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "pth 74%[=============> ] 1.97G 39.4MB/s eta 18s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "th 75%[==============> ] 1.97G 38.7MB/s eta 18s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "h 75%[==============> ] 1.98G 39.3MB/s eta 17s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " 75%[==============> ] 1.99G 40.3MB/s eta 17s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v 76%[==============> ] 2.00G 40.0MB/s eta 17s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5 76%[==============> ] 2.01G 40.8MB/s eta 17s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5- 76%[==============> ] 2.02G 40.8MB/s eta 17s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5-L 77%[==============> ] 2.03G 40.8MB/s eta 16s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5-L9 77%[==============> ] 2.03G 41.2MB/s eta 16s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5-L96 77%[==============> ] 2.04G 41.2MB/s eta 16s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5-L96- 77%[==============> ] 2.05G 41.8MB/s eta 16s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5-L96-D 78%[==============> ] 2.06G 40.4MB/s eta 16s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5-L96-D1 78%[==============> ] 2.06G 39.9MB/s eta 15s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5-L96-D10 78%[==============> ] 2.08G 41.5MB/s eta 15s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5-L96-D102 79%[==============> ] 2.08G 40.8MB/s eta 15s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5-L96-D1024 79%[==============> ] 2.09G 40.5MB/s eta 15s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5-L96-D1024- 79%[==============> ] 2.10G 39.8MB/s eta 15s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5-L96-D1024-E 80%[===============> ] 2.11G 40.0MB/s eta 14s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5-L96-D1024-E0 80%[===============> ] 2.12G 39.9MB/s eta 14s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5-L96-D1024-E0_ 80%[===============> ] 2.12G 40.2MB/s eta 14s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5-L96-D1024-E0_1 81%[===============> ] 2.13G 39.1MB/s eta 14s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5-L96-D1024-E0_1- 81%[===============> ] 2.14G 39.9MB/s eta 14s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "v5-L96-D1024-E0_1-m 81%[===============> ] 2.15G 39.5MB/s eta 13s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "5-L96-D1024-E0_1-me 82%[===============> ] 2.16G 39.8MB/s eta 13s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "-L96-D1024-E0_1-mem 82%[===============> ] 2.17G 39.3MB/s eta 13s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "L96-D1024-E0_1-mem- 82%[===============> ] 2.17G 39.2MB/s eta 13s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "96-D1024-E0_1-mem-c 82%[===============> ] 2.18G 39.5MB/s eta 13s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "6-D1024-E0_1-mem-ct 83%[===============> ] 2.19G 39.7MB/s eta 11s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "-D1024-E0_1-mem-ctx 83%[===============> ] 2.20G 40.6MB/s eta 11s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "D1024-E0_1-mem-ctx- 83%[===============> ] 2.21G 39.8MB/s eta 11s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "1024-E0_1-mem-ctx-4 84%[===============> ] 2.21G 39.9MB/s eta 11s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "024-E0_1-mem-ctx-4k 84%[===============> ] 2.22G 40.4MB/s eta 11s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "24-E0_1-mem-ctx-4k. 84%[===============> ] 2.23G 40.4MB/s eta 10s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "4-E0_1-mem-ctx-4k.p 85%[================> ] 2.24G 40.8MB/s eta 10s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "-E0_1-mem-ctx-4k.pt 85%[================> ] 2.25G 40.5MB/s eta 10s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "E0_1-mem-ctx-4k.pth 85%[================> ] 2.26G 40.4MB/s eta 10s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "0_1-mem-ctx-4k.pth 86%[================> ] 2.27G 40.8MB/s eta 10s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "_1-mem-ctx-4k.pth 86%[================> ] 2.27G 41.0MB/s eta 9s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "1-mem-ctx-4k.pth 86%[================> ] 2.28G 40.1MB/s eta 9s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "-mem-ctx-4k.pth 87%[================> ] 2.29G 40.1MB/s eta 9s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "mem-ctx-4k.pth 87%[================> ] 2.30G 40.0MB/s eta 9s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "em-ctx-4k.pth 87%[================> ] 2.31G 40.4MB/s eta 9s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "m-ctx-4k.pth 87%[================> ] 2.31G 40.1MB/s eta 8s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "-ctx-4k.pth 88%[================> ] 2.32G 39.9MB/s eta 8s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "ctx-4k.pth 88%[================> ] 2.33G 40.0MB/s eta 8s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "tx-4k.pth 88%[================> ] 2.34G 39.2MB/s eta 8s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "x-4k.pth 89%[================> ] 2.35G 39.7MB/s eta 8s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "-4k.pth 89%[================> ] 2.36G 39.4MB/s eta 7s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "4k.pth 89%[================> ] 2.36G 39.4MB/s eta 7s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "k.pth 90%[=================> ] 2.37G 39.6MB/s eta 7s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + ".pth 90%[=================> ] 2.38G 39.4MB/s eta 7s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "pth 90%[=================> ] 2.39G 39.2MB/s eta 7s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "th 91%[=================> ] 2.40G 39.2MB/s eta 6s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "h 91%[=================> ] 2.41G 39.4MB/s eta 6s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " 91%[=================> ] 2.41G 40.0MB/s eta 6s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v 92%[=================> ] 2.42G 39.8MB/s eta 6s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5 92%[=================> ] 2.43G 39.7MB/s eta 6s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5- 92%[=================> ] 2.44G 39.5MB/s eta 5s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5-L 92%[=================> ] 2.45G 40.6MB/s eta 5s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5-L9 93%[=================> ] 2.45G 39.8MB/s eta 5s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5-L96 93%[=================> ] 2.46G 39.0MB/s eta 5s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5-L96- 93%[=================> ] 2.46G 37.8MB/s eta 5s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5-L96-D 94%[=================> ] 2.48G 40.1MB/s eta 4s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5-L96-D1 94%[=================> ] 2.49G 40.3MB/s eta 4s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5-L96-D10 94%[=================> ] 2.50G 40.4MB/s eta 4s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5-L96-D102 95%[==================> ] 2.51G 40.3MB/s eta 4s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5-L96-D1024 95%[==================> ] 2.51G 40.4MB/s eta 4s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5-L96-D1024- 95%[==================> ] 2.52G 41.1MB/s eta 3s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5-L96-D1024-E 96%[==================> ] 2.53G 40.8MB/s eta 3s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5-L96-D1024-E0 96%[==================> ] 2.54G 40.5MB/s eta 3s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5-L96-D1024-E0_ 96%[==================> ] 2.55G 40.5MB/s eta 3s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5-L96-D1024-E0_1 97%[==================> ] 2.55G 40.3MB/s eta 3s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + " v5-L96-D1024-E0_1- 97%[==================> ] 2.56G 40.3MB/s eta 2s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "v5-L96-D1024-E0_1-m 97%[==================> ] 2.57G 40.5MB/s eta 2s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "5-L96-D1024-E0_1-me 98%[==================> ] 2.58G 40.9MB/s eta 2s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "-L96-D1024-E0_1-mem 98%[==================> ] 2.59G 40.8MB/s eta 2s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "L96-D1024-E0_1-mem- 98%[==================> ] 2.60G 42.5MB/s eta 2s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "96-D1024-E0_1-mem-c 99%[==================> ] 2.61G 42.8MB/s eta 1s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "6-D1024-E0_1-mem-ct 99%[==================> ] 2.61G 40.8MB/s eta 1s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "-D1024-E0_1-mem-ctx 99%[==================> ] 2.62G 40.9MB/s eta 1s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "D1024-E0_1-mem-ctx- 99%[==================> ] 2.63G 40.8MB/s eta 1s " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "v5-L96-D1024-E0_1-m 100%[===================>] 2.63G 41.1MB/s in 68s \r\n", + "\r\n", + "2023-09-02 06:17:41 (39.7 MB/s) - ‘v5-L96-D1024-E0_1-mem-ctx-4k.pth’ saved [2825976699/2825976699]\r\n", + "\r\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "total 2.7G\r\n", + "drwxr-xr-x 2 root root 54 Sep 2 06:16 .\r\n", + "drwxr-xr-x 19 root root 4.0K Sep 2 06:16 ..\r\n", + "-rw-r--r-- 1 root root 2.7G Sep 2 05:37 v5-L96-D1024-E0_1-mem-ctx-4k.pth\r\n" + ] + } + ], + "source": [ + "# Download the model directly (stop gap till HF sync issues is resolved)\n", + "!cd \"{TRAINER_DIR}\" && cd \"../model/\" && \\\n", + " wget -nc \"https://huggingface.co/rwkv-x-dev/rwkv-x-playground/resolve/main/experiment/rwkv-x-exp/v5-memory/{FILENAME_PREFIX}-mem-ctx-4k.pth\"\n", + "\n", + "!cd \"{TRAINER_DIR}\" && cd \"../model/\" && \\\n", + " ls -alh ." + ] + }, + { + "cell_type": "markdown", + "id": "adf68d89", + "metadata": { + "papermill": { + "duration": 0.01742, + "end_time": "2023-09-02T06:17:41.804733", + "exception": false, + "start_time": "2023-09-02T06:17:41.787313", + "status": "completed" + }, + "tags": [] + }, + "source": [ + "## Tune 6 : Ramping up the ctx size (8192), memory training\n", + "\n", + "- Tune 6: Large ctx size (8192), Scaling up!" + ] + }, + { + "cell_type": "code", + "execution_count": 5, + "id": "51c58e54", + "metadata": { + "execution": { + "iopub.execute_input": "2023-09-02T06:17:41.841382Z", + "iopub.status.busy": "2023-09-02T06:17:41.841191Z", + "iopub.status.idle": "2023-09-02T06:17:48.969203Z", + "shell.execute_reply": "2023-09-02T06:17:48.967550Z" + }, + "papermill": { + "duration": 7.217868, + "end_time": "2023-09-02T06:17:49.040024", + "exception": false, + "start_time": "2023-09-02T06:17:41.822156", + "status": "completed" + }, + "tags": [] + }, + "outputs": [ + { + "name": "stdout", + "output_type": "stream", + "text": [ + "## Generating word reptition dataset ##\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 2 max words, 50 samples - at ../dataset/word-2-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 10 max words, 50 samples - at ../dataset/gen-word-10-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 35 max words, 50 samples - at ../dataset/gen-word-35-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 566 samples (1 token repeat) - 5 max words - at ../dataset/shuffle-word-5-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 5 max words, 50 samples - at ../dataset/gen-word-5-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 60 max words, 50 samples - at ../dataset/gen-word-60-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 30 max words, 50 samples - at ../dataset/gen-word-30-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 84 samples (1 token repeat) - 30 max words - at ../dataset/shuffle-word-30-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 77 samples (1 token repeat) - 35 max words - at ../dataset/shuffle-word-35-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 25 max words, 50 samples - at ../dataset/gen-word-25-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 45 max words, 50 samples - at ../dataset/gen-word-45-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 63 samples (1 token repeat) - 40 max words - at ../dataset/shuffle-word-40-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 50 max words, 50 samples - at ../dataset/gen-word-50-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 55 max words, 50 samples - at ../dataset/gen-word-55-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 15 max words, 50 samples - at ../dataset/gen-word-15-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 120 max words, 50 samples - at ../dataset/gen-word-120-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 129 samples (1 token repeat) - 20 max words - at ../dataset/shuffle-word-20-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 56 samples (1 token repeat) - 50 max words - at ../dataset/shuffle-word-50-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 264 samples (1 token repeat) - 10 max words - at ../dataset/shuffle-word-10-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 40 max words, 50 samples - at ../dataset/gen-word-40-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 189 samples (1 token repeat) - 15 max words - at ../dataset/shuffle-word-15-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 19 samples (1 token repeat) - 115 max words - at ../dataset/shuffle-word-115-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 58 samples (1 token repeat) - 45 max words - at ../dataset/shuffle-word-45-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 20 max words, 50 samples - at ../dataset/gen-word-20-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 20 samples (1 token repeat) - 110 max words - at ../dataset/shuffle-word-110-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 10 samples (1 token repeat) - 220 max words - at ../dataset/shuffle-word-220-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 110 max words, 50 samples - at ../dataset/gen-word-110-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 50 samples (1 token repeat) - 55 max words - at ../dataset/shuffle-word-55-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 105 max words, 50 samples - at ../dataset/gen-word-105-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 85 max words, 50 samples - at ../dataset/gen-word-85-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 100 max words, 50 samples - at ../dataset/gen-word-100-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 235 max words, 50 samples - at ../dataset/gen-word-235-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 19 samples (1 token repeat) - 120 max words - at ../dataset/shuffle-word-120-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 140 max words, 50 samples - at ../dataset/gen-word-140-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 125 max words, 50 samples - at ../dataset/gen-word-125-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 80 max words, 50 samples - at ../dataset/gen-word-80-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 95 max words, 50 samples - at ../dataset/gen-word-95-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 105 samples (1 token repeat) - 25 max words - at ../dataset/shuffle-word-25-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 10 samples (1 token repeat) - 215 max words - at ../dataset/shuffle-word-215-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 10 samples (1 token repeat) - 265 max words - at ../dataset/shuffle-word-265-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 21 samples (1 token repeat) - 105 max words - at ../dataset/shuffle-word-105-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 310 max words, 50 samples - at ../dataset/gen-word-310-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 70 max words, 50 samples - at ../dataset/gen-word-70-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 115 max words, 50 samples - at ../dataset/gen-word-115-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 5 samples (1 token repeat) - 590 max words - at ../dataset/shuffle-word-590-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 200 max words, 50 samples - at ../dataset/gen-word-200-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 5 samples (1 token repeat) - 595 max words - at ../dataset/shuffle-word-595-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 9 samples (1 token repeat) - 255 max words - at ../dataset/shuffle-word-255-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 175 max words, 50 samples - at ../dataset/gen-word-175-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 26 samples (1 token repeat) - 95 max words - at ../dataset/shuffle-word-95-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 6 samples (1 token repeat) - 450 max words - at ../dataset/shuffle-word-450-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 270 max words, 50 samples - at ../dataset/gen-word-270-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 44 samples (1 token repeat) - 60 max words - at ../dataset/shuffle-word-60-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 415 max words, 50 samples - at ../dataset/gen-word-415-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 585 max words, 50 samples - at ../dataset/gen-word-585-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 7 samples (1 token repeat) - 350 max words - at ../dataset/shuffle-word-350-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 430 max words, 50 samples - at ../dataset/gen-word-430-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 17 samples (1 token repeat) - 135 max words - at ../dataset/shuffle-word-135-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 5 samples (1 token repeat) - 525 max words - at ../dataset/shuffle-word-525-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 5 samples (1 token repeat) - 530 max words - at ../dataset/shuffle-word-530-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 8 samples (1 token repeat) - 305 max words - at ../dataset/shuffle-word-305-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 11 samples (1 token repeat) - 210 max words - at ../dataset/shuffle-word-210-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 285 max words, 50 samples - at ../dataset/gen-word-285-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 330 max words, 50 samples - at ../dataset/gen-word-330-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 6 samples (1 token repeat) - 495 max words - at ../dataset/shuffle-word-495-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 400 max words, 50 samples - at ../dataset/gen-word-400-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 9 samples (1 token repeat) - 240 max words - at ../dataset/shuffle-word-240-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 5 samples (1 token repeat) - 580 max words - at ../dataset/shuffle-word-580-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 520 max words, 50 samples - at ../dataset/gen-word-520-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 370 max words, 50 samples - at ../dataset/gen-word-370-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 425 max words, 50 samples - at ../dataset/gen-word-425-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 7 samples (1 token repeat) - 365 max words - at ../dataset/shuffle-word-365-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 10 samples (1 token repeat) - 225 max words - at ../dataset/shuffle-word-225-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 10 samples (1 token repeat) - 230 max words - at ../dataset/shuffle-word-230-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 290 max words, 50 samples - at ../dataset/gen-word-290-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 7 samples (1 token repeat) - 400 max words - at ../dataset/shuffle-word-400-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 6 samples (1 token repeat) - 445 max words - at ../dataset/shuffle-word-445-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 180 max words, 50 samples - at ../dataset/gen-word-180-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 65 max words, 50 samples - at ../dataset/gen-word-65-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 5 samples (1 token repeat) - 545 max words - at ../dataset/shuffle-word-545-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 18 samples (1 token repeat) - 130 max words - at ../dataset/shuffle-word-130-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 560 max words, 50 samples - at ../dataset/gen-word-560-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 480 max words, 50 samples - at ../dataset/gen-word-480-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 32 samples (1 token repeat) - 85 max words - at ../dataset/shuffle-word-85-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 36 samples (1 token repeat) - 75 max words - at ../dataset/shuffle-word-75-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 15 samples (1 token repeat) - 155 max words - at ../dataset/shuffle-word-155-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 4 samples (1 token repeat) - 800 max words - at ../dataset/shuffle-word-800-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 365 max words, 50 samples - at ../dataset/gen-word-365-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 5 samples (1 token repeat) - 600 max words - at ../dataset/shuffle-word-600-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 18 samples (1 token repeat) - 125 max words - at ../dataset/shuffle-word-125-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 16 samples (1 token repeat) - 140 max words - at ../dataset/shuffle-word-140-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 155 max words, 50 samples - at ../dataset/gen-word-155-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 130 max words, 50 samples - at ../dataset/gen-word-130-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 445 max words, 50 samples - at ../dataset/gen-word-445-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 6 samples (1 token repeat) - 470 max words - at ../dataset/shuffle-word-470-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 580 max words, 50 samples - at ../dataset/gen-word-580-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 4 samples (1 token repeat) - 610 max words - at ../dataset/shuffle-word-610-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 405 max words, 50 samples - at ../dataset/gen-word-405-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 6 samples (1 token repeat) - 490 max words - at ../dataset/shuffle-word-490-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 4 samples (1 token repeat) - 630 max words - at ../dataset/shuffle-word-630-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 755 max words, 50 samples - at ../dataset/gen-word-755-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 255 max words, 50 samples - at ../dataset/gen-word-255-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 4 samples (1 token repeat) - 745 max words - at ../dataset/shuffle-word-745-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 9 samples (1 token repeat) - 280 max words - at ../dataset/shuffle-word-280-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 4 samples (1 token repeat) - 690 max words - at ../dataset/shuffle-word-690-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 15 samples (1 token repeat) - 160 max words - at ../dataset/shuffle-word-160-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 740 max words, 50 samples - at ../dataset/gen-word-740-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 145 max words, 50 samples - at ../dataset/gen-word-145-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 7 samples (1 token repeat) - 360 max words - at ../dataset/shuffle-word-360-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 3 samples (1 token repeat) - 835 max words - at ../dataset/shuffle-word-835-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 210 max words, 50 samples - at ../dataset/gen-word-210-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 5 samples (1 token repeat) - 585 max words - at ../dataset/shuffle-word-585-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 39 samples (1 token repeat) - 65 max words - at ../dataset/shuffle-word-65-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 9 samples (1 token repeat) - 250 max words - at ../dataset/shuffle-word-250-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 450 max words, 50 samples - at ../dataset/gen-word-450-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 5 samples (1 token repeat) - 550 max words - at ../dataset/shuffle-word-550-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 455 max words, 50 samples - at ../dataset/gen-word-455-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 435 max words, 50 samples - at ../dataset/gen-word-435-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 4 samples (1 token repeat) - 760 max words - at ../dataset/shuffle-word-760-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 165 max words, 50 samples - at ../dataset/gen-word-165-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 14 samples (1 token repeat) - 170 max words - at ../dataset/shuffle-word-170-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 15 samples (1 token repeat) - 180 max words - at ../dataset/shuffle-word-180-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 75 max words, 50 samples - at ../dataset/gen-word-75-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 28 samples (1 token repeat) - 100 max words - at ../dataset/shuffle-word-100-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 39 samples (1 token repeat) - 70 max words - at ../dataset/shuffle-word-70-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 550 max words, 50 samples - at ../dataset/gen-word-550-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 3 samples (1 token repeat) - 955 max words - at ../dataset/shuffle-word-955-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 355 max words, 50 samples - at ../dataset/gen-word-355-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 9 samples (1 token repeat) - 295 max words - at ../dataset/shuffle-word-295-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 4 samples (1 token repeat) - 830 max words - at ../dataset/shuffle-word-830-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 4 samples (1 token repeat) - 770 max words - at ../dataset/shuffle-word-770-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 9 samples (1 token repeat) - 285 max words - at ../dataset/shuffle-word-285-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 470 max words, 50 samples - at ../dataset/gen-word-470-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 345 max words, 50 samples - at ../dataset/gen-word-345-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 7 samples (1 token repeat) - 355 max words - at ../dataset/shuffle-word-355-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 410 max words, 50 samples - at ../dataset/gen-word-410-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 5 samples (1 token repeat) - 535 max words - at ../dataset/shuffle-word-535-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 15 samples (1 token repeat) - 165 max words - at ../dataset/shuffle-word-165-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 995 max words, 50 samples - at ../dataset/gen-word-995-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 60 samples (20 token repeat) - 1100 max words - at ../dataset/shuffle-word-1100-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 225 max words, 50 samples - at ../dataset/gen-word-225-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 245 max words, 50 samples - at ../dataset/gen-word-245-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 13 samples (1 token repeat) - 190 max words - at ../dataset/shuffle-word-190-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 13 samples (1 token repeat) - 200 max words - at ../dataset/shuffle-word-200-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 7 samples (1 token repeat) - 310 max words - at ../dataset/shuffle-word-310-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 230 max words, 50 samples - at ../dataset/gen-word-230-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 59 samples (20 token repeat) - 1200 max words - at ../dataset/shuffle-word-1200-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 390 max words, 50 samples - at ../dataset/gen-word-390-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 5 samples (1 token repeat) - 500 max words - at ../dataset/shuffle-word-500-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 685 max words, 50 samples - at ../dataset/gen-word-685-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 635 max words, 50 samples - at ../dataset/gen-word-635-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 545 max words, 50 samples - at ../dataset/gen-word-545-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 20 samples (20 token repeat) - 4200 max words - at ../dataset/shuffle-word-4200-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 4 samples (1 token repeat) - 715 max words - at ../dataset/shuffle-word-715-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 625 max words, 50 samples - at ../dataset/gen-word-625-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 4 samples (1 token repeat) - 795 max words - at ../dataset/shuffle-word-795-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 4 samples (1 token repeat) - 705 max words - at ../dataset/shuffle-word-705-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 33 samples (1 token repeat) - 80 max words - at ../dataset/shuffle-word-80-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 905 max words, 50 samples - at ../dataset/gen-word-905-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 4 samples (1 token repeat) - 780 max words - at ../dataset/shuffle-word-780-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 5 samples (1 token repeat) - 520 max words - at ../dataset/shuffle-word-520-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 375 max words, 50 samples - at ../dataset/gen-word-375-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 3 samples (1 token repeat) - 940 max words - at ../dataset/shuffle-word-940-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 150 max words, 50 samples - at ../dataset/gen-word-150-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 7 samples (1 token repeat) - 345 max words - at ../dataset/shuffle-word-345-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 275 max words, 50 samples - at ../dataset/gen-word-275-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 6 samples (1 token repeat) - 425 max words - at ../dataset/shuffle-word-425-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 3 samples (1 token repeat) - 945 max words - at ../dataset/shuffle-word-945-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 3 samples (1 token repeat) - 815 max words - at ../dataset/shuffle-word-815-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 3 samples (1 token repeat) - 960 max words - at ../dataset/shuffle-word-960-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 14 samples (1 token repeat) - 195 max words - at ../dataset/shuffle-word-195-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 3 samples (1 token repeat) - 825 max words - at ../dataset/shuffle-word-825-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 3 samples (1 token repeat) - 980 max words - at ../dataset/shuffle-word-980-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 7 samples (1 token repeat) - 335 max words - at ../dataset/shuffle-word-335-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 830 max words, 50 samples - at ../dataset/gen-word-830-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 3 samples (1 token repeat) - 810 max words - at ../dataset/shuffle-word-810-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 3 samples (1 token repeat) - 820 max words - at ../dataset/shuffle-word-820-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 3 samples (1 token repeat) - 990 max words - at ../dataset/shuffle-word-990-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 725 max words, 50 samples - at ../dataset/gen-word-725-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 3 samples (1 token repeat) - 975 max words - at ../dataset/shuffle-word-975-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 7 samples (1 token repeat) - 315 max words - at ../dataset/shuffle-word-315-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 4 samples (1 token repeat) - 680 max words - at ../dataset/shuffle-word-680-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 20 samples (20 token repeat) - 5400 max words - at ../dataset/shuffle-word-5400-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 525 max words, 50 samples - at ../dataset/gen-word-525-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 20 samples (20 token repeat) - 7300 max words - at ../dataset/shuffle-word-7300-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 170 max words, 50 samples - at ../dataset/gen-word-170-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 250 max words, 50 samples - at ../dataset/gen-word-250-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 380 max words, 50 samples - at ../dataset/gen-word-380-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 4 samples (1 token repeat) - 740 max words - at ../dataset/shuffle-word-740-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 4 samples (1 token repeat) - 725 max words - at ../dataset/shuffle-word-725-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 715 max words, 50 samples - at ../dataset/gen-word-715-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 760 max words, 50 samples - at ../dataset/gen-word-760-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 5 samples (1 token repeat) - 510 max words - at ../dataset/shuffle-word-510-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 7 samples (1 token repeat) - 375 max words - at ../dataset/shuffle-word-375-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 3 samples (1 token repeat) - 860 max words - at ../dataset/shuffle-word-860-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 7 samples (1 token repeat) - 330 max words - at ../dataset/shuffle-word-330-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 4 samples (1 token repeat) - 755 max words - at ../dataset/shuffle-word-755-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 3 samples (1 token repeat) - 905 max words - at ../dataset/shuffle-word-905-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 915 max words, 50 samples - at ../dataset/gen-word-915-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 4 samples (1 token repeat) - 660 max words - at ../dataset/shuffle-word-660-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 20 samples (20 token repeat) - 4000 max words - at ../dataset/shuffle-word-4000-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 315 max words, 50 samples - at ../dataset/gen-word-315-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 880 max words, 50 samples - at ../dataset/gen-word-880-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 195 max words, 50 samples - at ../dataset/gen-word-195-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 4 samples (1 token repeat) - 625 max words - at ../dataset/shuffle-word-625-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 40 samples (20 token repeat) - 2100 max words - at ../dataset/shuffle-word-2100-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 325 max words, 50 samples - at ../dataset/gen-word-325-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 4 samples (1 token repeat) - 735 max words - at ../dataset/shuffle-word-735-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 795 max words, 50 samples - at ../dataset/gen-word-795-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 890 max words, 50 samples - at ../dataset/gen-word-890-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 4 samples (1 token repeat) - 685 max words - at ../dataset/shuffle-word-685-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 4 samples (1 token repeat) - 620 max words - at ../dataset/shuffle-word-620-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 4 samples (1 token repeat) - 785 max words - at ../dataset/shuffle-word-785-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 9 samples (1 token repeat) - 275 max words - at ../dataset/shuffle-word-275-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 4 samples (1 token repeat) - 655 max words - at ../dataset/shuffle-word-655-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 340 max words, 50 samples - at ../dataset/gen-word-340-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 3 samples (1 token repeat) - 920 max words - at ../dataset/shuffle-word-920-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 20 samples (20 token repeat) - 7100 max words - at ../dataset/shuffle-word-7100-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 655 max words, 50 samples - at ../dataset/gen-word-655-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 4 samples (1 token repeat) - 720 max words - at ../dataset/shuffle-word-720-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 3 samples (1 token repeat) - 970 max words - at ../dataset/shuffle-word-970-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 845 max words, 50 samples - at ../dataset/gen-word-845-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 20 samples (20 token repeat) - 3000 max words - at ../dataset/shuffle-word-3000-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 6 samples (1 token repeat) - 430 max words - at ../dataset/shuffle-word-430-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 14 samples (1 token repeat) - 175 max words - at ../dataset/shuffle-word-175-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 4 samples (1 token repeat) - 790 max words - at ../dataset/shuffle-word-790-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 90 max words, 50 samples - at ../dataset/gen-word-90-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 3 samples (1 token repeat) - 895 max words - at ../dataset/shuffle-word-895-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 37 samples (20 token repeat) - 2500 max words - at ../dataset/shuffle-word-2500-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 3 samples (1 token repeat) - 865 max words - at ../dataset/shuffle-word-865-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 51 samples (20 token repeat) - 1300 max words - at ../dataset/shuffle-word-1300-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 660 max words, 50 samples - at ../dataset/gen-word-660-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 220 max words, 50 samples - at ../dataset/gen-word-220-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 6 samples (1 token repeat) - 485 max words - at ../dataset/shuffle-word-485-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 695 max words, 50 samples - at ../dataset/gen-word-695-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 840 max words, 50 samples - at ../dataset/gen-word-840-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 745 max words, 50 samples - at ../dataset/gen-word-745-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 600 max words, 50 samples - at ../dataset/gen-word-600-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 790 max words, 50 samples - at ../dataset/gen-word-790-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 540 max words, 50 samples - at ../dataset/gen-word-540-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 420 max words, 50 samples - at ../dataset/gen-word-420-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 825 max words, 50 samples - at ../dataset/gen-word-825-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 805 max words, 50 samples - at ../dataset/gen-word-805-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 765 max words, 50 samples - at ../dataset/gen-word-765-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 4 samples (1 token repeat) - 855 max words - at ../dataset/shuffle-word-855-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 965 max words, 50 samples - at ../dataset/gen-word-965-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 20 samples (20 token repeat) - 5900 max words - at ../dataset/shuffle-word-5900-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 40 samples (20 token repeat) - 1700 max words - at ../dataset/shuffle-word-1700-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 705 max words, 50 samples - at ../dataset/gen-word-705-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 20 samples (20 token repeat) - 7900 max words - at ../dataset/shuffle-word-7900-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 40 samples (20 token repeat) - 2300 max words - at ../dataset/shuffle-word-2300-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 385 max words, 50 samples - at ../dataset/gen-word-385-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 990 max words, 50 samples - at ../dataset/gen-word-990-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 515 max words, 50 samples - at ../dataset/gen-word-515-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 640 max words, 50 samples - at ../dataset/gen-word-640-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 730 max words, 50 samples - at ../dataset/gen-word-730-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 930 max words, 50 samples - at ../dataset/gen-word-930-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 20 samples (20 token repeat) - 4600 max words - at ../dataset/shuffle-word-4600-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 590 max words, 50 samples - at ../dataset/gen-word-590-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 595 max words, 50 samples - at ../dataset/gen-word-595-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 20 samples (20 token repeat) - 3300 max words - at ../dataset/shuffle-word-3300-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 820 max words, 50 samples - at ../dataset/gen-word-820-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 320 max words, 50 samples - at ../dataset/gen-word-320-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 20 samples (20 token repeat) - 5600 max words - at ../dataset/shuffle-word-5600-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 4 samples (1 token repeat) - 750 max words - at ../dataset/shuffle-word-750-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 20 samples (20 token repeat) - 5300 max words - at ../dataset/shuffle-word-5300-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 9 samples (1 token repeat) - 300 max words - at ../dataset/shuffle-word-300-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 7 samples (1 token repeat) - 325 max words - at ../dataset/shuffle-word-325-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 670 max words, 50 samples - at ../dataset/gen-word-670-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 4 samples (1 token repeat) - 730 max words - at ../dataset/shuffle-word-730-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 20 samples (20 token repeat) - 4100 max words - at ../dataset/shuffle-word-4100-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 11 samples (1 token repeat) - 205 max words - at ../dataset/shuffle-word-205-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 980 max words, 50 samples - at ../dataset/gen-word-980-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 935 max words, 50 samples - at ../dataset/gen-word-935-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 1000 max words, 50 samples - at ../dataset/gen-word-1000-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 40 samples (20 token repeat) - 1900 max words - at ../dataset/shuffle-word-1900-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 895 max words, 50 samples - at ../dataset/gen-word-895-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 3 samples (1 token repeat) - 870 max words - at ../dataset/shuffle-word-870-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 750 max words, 50 samples - at ../dataset/gen-word-750-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 9 samples (1 token repeat) - 245 max words - at ../dataset/shuffle-word-245-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 9 samples (1 token repeat) - 270 max words - at ../dataset/shuffle-word-270-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 7 samples (1 token repeat) - 380 max words - at ../dataset/shuffle-word-380-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 6 samples (1 token repeat) - 405 max words - at ../dataset/shuffle-word-405-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 6 samples (1 token repeat) - 420 max words - at ../dataset/shuffle-word-420-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 300 max words, 50 samples - at ../dataset/gen-word-300-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 910 max words, 50 samples - at ../dataset/gen-word-910-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 735 max words, 50 samples - at ../dataset/gen-word-735-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 40 samples (20 token repeat) - 2000 max words - at ../dataset/shuffle-word-2000-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 39 samples (20 token repeat) - 2400 max words - at ../dataset/shuffle-word-2400-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 20 samples (20 token repeat) - 8000 max words - at ../dataset/shuffle-word-8000-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 4 samples (1 token repeat) - 645 max words - at ../dataset/shuffle-word-645-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 630 max words, 50 samples - at ../dataset/gen-word-630-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 20 samples (20 token repeat) - 7400 max words - at ../dataset/shuffle-word-7400-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 305 max words, 50 samples - at ../dataset/gen-word-305-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 6 samples (1 token repeat) - 455 max words - at ../dataset/shuffle-word-455-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 860 max words, 50 samples - at ../dataset/gen-word-860-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 20 samples (20 token repeat) - 7700 max words - at ../dataset/shuffle-word-7700-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 185 max words, 50 samples - at ../dataset/gen-word-185-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 260 max words, 50 samples - at ../dataset/gen-word-260-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 920 max words, 50 samples - at ../dataset/gen-word-920-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 665 max words, 50 samples - at ../dataset/gen-word-665-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 565 max words, 50 samples - at ../dataset/gen-word-565-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 215 max words, 50 samples - at ../dataset/gen-word-215-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 5 samples (1 token repeat) - 540 max words - at ../dataset/shuffle-word-540-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 800 max words, 50 samples - at ../dataset/gen-word-800-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 190 max words, 50 samples - at ../dataset/gen-word-190-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 5 samples (1 token repeat) - 560 max words - at ../dataset/shuffle-word-560-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 20 samples (20 token repeat) - 4700 max words - at ../dataset/shuffle-word-4700-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 495 max words, 50 samples - at ../dataset/gen-word-495-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 360 max words, 50 samples - at ../dataset/gen-word-360-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 950 max words, 50 samples - at ../dataset/gen-word-950-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 4 samples (1 token repeat) - 615 max words - at ../dataset/shuffle-word-615-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 20 samples (20 token repeat) - 6300 max words - at ../dataset/shuffle-word-6300-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 3 samples (1 token repeat) - 845 max words - at ../dataset/shuffle-word-845-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 505 max words, 50 samples - at ../dataset/gen-word-505-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 20 samples (20 token repeat) - 3900 max words - at ../dataset/shuffle-word-3900-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 4 samples (1 token repeat) - 670 max words - at ../dataset/shuffle-word-670-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 20 samples (20 token repeat) - 6400 max words - at ../dataset/shuffle-word-6400-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 815 max words, 50 samples - at ../dataset/gen-word-815-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 6 samples (1 token repeat) - 440 max words - at ../dataset/shuffle-word-440-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 40 samples (20 token repeat) - 1500 max words - at ../dataset/shuffle-word-1500-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 3 samples (1 token repeat) - 925 max words - at ../dataset/shuffle-word-925-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 6 samples (1 token repeat) - 465 max words - at ../dataset/shuffle-word-465-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 690 max words, 50 samples - at ../dataset/gen-word-690-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 395 max words, 50 samples - at ../dataset/gen-word-395-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 265 max words, 50 samples - at ../dataset/gen-word-265-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 955 max words, 50 samples - at ../dataset/gen-word-955-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 885 max words, 50 samples - at ../dataset/gen-word-885-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 14 samples (1 token repeat) - 185 max words - at ../dataset/shuffle-word-185-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 240 max words, 50 samples - at ../dataset/gen-word-240-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 20 samples (20 token repeat) - 6600 max words - at ../dataset/shuffle-word-6600-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 440 max words, 50 samples - at ../dataset/gen-word-440-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 135 max words, 50 samples - at ../dataset/gen-word-135-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 555 max words, 50 samples - at ../dataset/gen-word-555-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 605 max words, 50 samples - at ../dataset/gen-word-605-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 465 max words, 50 samples - at ../dataset/gen-word-465-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 295 max words, 50 samples - at ../dataset/gen-word-295-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 7 samples (1 token repeat) - 390 max words - at ../dataset/shuffle-word-390-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 900 max words, 50 samples - at ../dataset/gen-word-900-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 975 max words, 50 samples - at ../dataset/gen-word-975-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 945 max words, 50 samples - at ../dataset/gen-word-945-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 6 samples (1 token repeat) - 480 max words - at ../dataset/shuffle-word-480-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 20 samples (20 token repeat) - 4900 max words - at ../dataset/shuffle-word-4900-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 20 samples (20 token repeat) - 5800 max words - at ../dataset/shuffle-word-5800-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 970 max words, 50 samples - at ../dataset/gen-word-970-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 4 samples (1 token repeat) - 605 max words - at ../dataset/shuffle-word-605-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 650 max words, 50 samples - at ../dataset/gen-word-650-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 3 samples (1 token repeat) - 985 max words - at ../dataset/shuffle-word-985-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 20 samples (20 token repeat) - 3800 max words - at ../dataset/shuffle-word-3800-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 9 samples (1 token repeat) - 235 max words - at ../dataset/shuffle-word-235-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 20 samples (20 token repeat) - 7000 max words - at ../dataset/shuffle-word-7000-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 5 samples (1 token repeat) - 570 max words - at ../dataset/shuffle-word-570-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 9 samples (1 token repeat) - 260 max words - at ../dataset/shuffle-word-260-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 280 max words, 50 samples - at ../dataset/gen-word-280-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 530 max words, 50 samples - at ../dataset/gen-word-530-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 20 samples (20 token repeat) - 2900 max words - at ../dataset/shuffle-word-2900-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 4 samples (1 token repeat) - 775 max words - at ../dataset/shuffle-word-775-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 610 max words, 50 samples - at ../dataset/gen-word-610-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 675 max words, 50 samples - at ../dataset/gen-word-675-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 350 max words, 50 samples - at ../dataset/gen-word-350-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 710 max words, 50 samples - at ../dataset/gen-word-710-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 15 samples (1 token repeat) - 150 max words - at ../dataset/shuffle-word-150-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 475 max words, 50 samples - at ../dataset/gen-word-475-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 575 max words, 50 samples - at ../dataset/gen-word-575-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 30 samples (1 token repeat) - 90 max words - at ../dataset/shuffle-word-90-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 5 samples (1 token repeat) - 575 max words - at ../dataset/shuffle-word-575-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 3 samples (1 token repeat) - 805 max words - at ../dataset/shuffle-word-805-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 7 samples (1 token repeat) - 320 max words - at ../dataset/shuffle-word-320-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 5 samples (1 token repeat) - 665 max words - at ../dataset/shuffle-word-665-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 960 max words, 50 samples - at ../dataset/gen-word-960-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 490 max words, 50 samples - at ../dataset/gen-word-490-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 810 max words, 50 samples - at ../dataset/gen-word-810-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 4 samples (1 token repeat) - 695 max words - at ../dataset/shuffle-word-695-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 3 samples (1 token repeat) - 995 max words - at ../dataset/shuffle-word-995-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 770 max words, 50 samples - at ../dataset/gen-word-770-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 3 samples (1 token repeat) - 875 max words - at ../dataset/shuffle-word-875-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 7 samples (1 token repeat) - 395 max words - at ../dataset/shuffle-word-395-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 41 samples (20 token repeat) - 1400 max words - at ../dataset/shuffle-word-1400-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 460 max words, 50 samples - at ../dataset/gen-word-460-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 20 samples (20 token repeat) - 7600 max words - at ../dataset/shuffle-word-7600-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 20 samples (20 token repeat) - 7200 max words - at ../dataset/shuffle-word-7200-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 20 samples (20 token repeat) - 4800 max words - at ../dataset/shuffle-word-4800-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 20 samples (20 token repeat) - 3600 max words - at ../dataset/shuffle-word-3600-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 535 max words, 50 samples - at ../dataset/gen-word-535-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 570 max words, 50 samples - at ../dataset/gen-word-570-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 4 samples (1 token repeat) - 700 max words - at ../dataset/shuffle-word-700-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 985 max words, 50 samples - at ../dataset/gen-word-985-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 510 max words, 50 samples - at ../dataset/gen-word-510-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 16 samples (1 token repeat) - 145 max words - at ../dataset/shuffle-word-145-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 7 samples (1 token repeat) - 340 max words - at ../dataset/shuffle-word-340-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 5 samples (1 token repeat) - 515 max words - at ../dataset/shuffle-word-515-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 6 samples (1 token repeat) - 435 max words - at ../dataset/shuffle-word-435-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 4 samples (1 token repeat) - 640 max words - at ../dataset/shuffle-word-640-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 775 max words, 50 samples - at ../dataset/gen-word-775-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 5 samples (1 token repeat) - 565 max words - at ../dataset/shuffle-word-565-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 6 samples (1 token repeat) - 415 max words - at ../dataset/shuffle-word-415-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 835 max words, 50 samples - at ../dataset/gen-word-835-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 20 samples (20 token repeat) - 6700 max words - at ../dataset/shuffle-word-6700-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 335 max words, 50 samples - at ../dataset/gen-word-335-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 620 max words, 50 samples - at ../dataset/gen-word-620-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 870 max words, 50 samples - at ../dataset/gen-word-870-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 3 samples (1 token repeat) - 890 max words - at ../dataset/shuffle-word-890-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 5 samples (1 token repeat) - 460 max words - at ../dataset/shuffle-word-460-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 4 samples (1 token repeat) - 765 max words - at ../dataset/shuffle-word-765-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 4 samples (1 token repeat) - 635 max words - at ../dataset/shuffle-word-635-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 500 max words, 50 samples - at ../dataset/gen-word-500-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 9 samples (1 token repeat) - 290 max words - at ../dataset/shuffle-word-290-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 720 max words, 50 samples - at ../dataset/gen-word-720-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 4 samples (1 token repeat) - 850 max words - at ../dataset/shuffle-word-850-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 3 samples (1 token repeat) - 1000 max words - at ../dataset/shuffle-word-1000-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 20 samples (20 token repeat) - 3700 max words - at ../dataset/shuffle-word-3700-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 160 max words, 50 samples - at ../dataset/gen-word-160-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 5 samples (1 token repeat) - 505 max words - at ../dataset/shuffle-word-505-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 30 samples (20 token repeat) - 2600 max words - at ../dataset/shuffle-word-2600-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 3 samples (1 token repeat) - 910 max words - at ../dataset/shuffle-word-910-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 22 samples (20 token repeat) - 2700 max words - at ../dataset/shuffle-word-2700-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 700 max words, 50 samples - at ../dataset/gen-word-700-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 20 samples (20 token repeat) - 5200 max words - at ../dataset/shuffle-word-5200-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 4 samples (1 token repeat) - 710 max words - at ../dataset/shuffle-word-710-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 3 samples (1 token repeat) - 935 max words - at ../dataset/shuffle-word-935-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 3 samples (1 token repeat) - 885 max words - at ../dataset/shuffle-word-885-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 20 samples (20 token repeat) - 3200 max words - at ../dataset/shuffle-word-3200-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 40 samples (20 token repeat) - 1800 max words - at ../dataset/shuffle-word-1800-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 785 max words, 50 samples - at ../dataset/gen-word-785-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 6 samples (1 token repeat) - 475 max words - at ../dataset/shuffle-word-475-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 40 samples (20 token repeat) - 2200 max words - at ../dataset/shuffle-word-2200-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 20 samples (20 token repeat) - 6000 max words - at ../dataset/shuffle-word-6000-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 645 max words, 50 samples - at ../dataset/gen-word-645-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 205 max words, 50 samples - at ../dataset/gen-word-205-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 4 samples (1 token repeat) - 650 max words - at ../dataset/shuffle-word-650-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 5 samples (1 token repeat) - 555 max words - at ../dataset/shuffle-word-555-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 875 max words, 50 samples - at ../dataset/gen-word-875-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 3 samples (1 token repeat) - 900 max words - at ../dataset/shuffle-word-900-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 20 samples (20 token repeat) - 4400 max words - at ../dataset/shuffle-word-4400-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 3 samples (1 token repeat) - 965 max words - at ../dataset/shuffle-word-965-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 3 samples (1 token repeat) - 950 max words - at ../dataset/shuffle-word-950-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 20 samples (20 token repeat) - 3500 max words - at ../dataset/shuffle-word-3500-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 615 max words, 50 samples - at ../dataset/gen-word-615-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 20 samples (20 token repeat) - 5500 max words - at ../dataset/shuffle-word-5500-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 4 samples (1 token repeat) - 675 max words - at ../dataset/shuffle-word-675-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 6 samples (1 token repeat) - 410 max words - at ../dataset/shuffle-word-410-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 7 samples (1 token repeat) - 370 max words - at ../dataset/shuffle-word-370-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 7 samples (1 token repeat) - 385 max words - at ../dataset/shuffle-word-385-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 855 max words, 50 samples - at ../dataset/gen-word-855-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 20 samples (20 token repeat) - 7500 max words - at ../dataset/shuffle-word-7500-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 20 samples (20 token repeat) - 3400 max words - at ../dataset/shuffle-word-3400-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 40 samples (20 token repeat) - 1600 max words - at ../dataset/shuffle-word-1600-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 3 samples (1 token repeat) - 880 max words - at ../dataset/shuffle-word-880-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 20 samples (20 token repeat) - 6800 max words - at ../dataset/shuffle-word-6800-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 3 samples (1 token repeat) - 930 max words - at ../dataset/shuffle-word-930-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 20 samples (20 token repeat) - 6200 max words - at ../dataset/shuffle-word-6200-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 20 samples (20 token repeat) - 3100 max words - at ../dataset/shuffle-word-3100-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 780 max words, 50 samples - at ../dataset/gen-word-780-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 20 samples (20 token repeat) - 5000 max words - at ../dataset/shuffle-word-5000-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 21 samples (20 token repeat) - 2800 max words - at ../dataset/shuffle-word-2800-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 20 samples (20 token repeat) - 6500 max words - at ../dataset/shuffle-word-6500-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 20 samples (20 token repeat) - 6900 max words - at ../dataset/shuffle-word-6900-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 925 max words, 50 samples - at ../dataset/gen-word-925-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 940 max words, 50 samples - at ../dataset/gen-word-940-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 680 max words, 50 samples - at ../dataset/gen-word-680-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 20 samples (20 token repeat) - 6100 max words - at ../dataset/shuffle-word-6100-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 485 max words, 50 samples - at ../dataset/gen-word-485-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 3 samples (1 token repeat) - 840 max words - at ../dataset/shuffle-word-840-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 850 max words, 50 samples - at ../dataset/gen-word-850-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 20 samples (20 token repeat) - 4300 max words - at ../dataset/shuffle-word-4300-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 20 samples (20 token repeat) - 7800 max words - at ../dataset/shuffle-word-7800-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 865 max words, 50 samples - at ../dataset/gen-word-865-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 3 samples (1 token repeat) - 915 max words - at ../dataset/shuffle-word-915-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 20 samples (20 token repeat) - 5700 max words - at ../dataset/shuffle-word-5700-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 20 samples (20 token repeat) - 5100 max words - at ../dataset/shuffle-word-5100-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated a single JSONL file with 20 samples (20 token repeat) - 4500 max words - at ../dataset/shuffle-word-4500-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 1100 max words, 2000 samples - at ../dataset/gen-word-1100-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 1400 max words, 2000 samples - at ../dataset/gen-word-1400-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 1300 max words, 2000 samples - at ../dataset/gen-word-1300-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 1200 max words, 2000 samples - at ../dataset/gen-word-1200-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 1500 max words, 2000 samples - at ../dataset/gen-word-1500-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 1700 max words, 2000 samples - at ../dataset/gen-word-1700-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 1800 max words, 2000 samples - at ../dataset/gen-word-1800-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 1600 max words, 2000 samples - at ../dataset/gen-word-1600-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 1900 max words, 2000 samples - at ../dataset/gen-word-1900-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 2100 max words, 2000 samples - at ../dataset/gen-word-2100-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 2000 max words, 2000 samples - at ../dataset/gen-word-2000-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 2200 max words, 2000 samples - at ../dataset/gen-word-2200-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 2300 max words, 2000 samples - at ../dataset/gen-word-2300-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 2500 max words, 2000 samples - at ../dataset/gen-word-2500-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 2400 max words, 2000 samples - at ../dataset/gen-word-2400-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 2600 max words, 2000 samples - at ../dataset/gen-word-2600-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 2800 max words, 2000 samples - at ../dataset/gen-word-2800-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 2700 max words, 2000 samples - at ../dataset/gen-word-2700-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 3300 max words, 2000 samples - at ../dataset/gen-word-3300-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 3000 max words, 2000 samples - at ../dataset/gen-word-3000-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 3100 max words, 2000 samples - at ../dataset/gen-word-3100-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 2900 max words, 2000 samples - at ../dataset/gen-word-2900-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 3400 max words, 2000 samples - at ../dataset/gen-word-3400-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 4100 max words, 2000 samples - at ../dataset/gen-word-4100-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 3200 max words, 2000 samples - at ../dataset/gen-word-3200-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 3600 max words, 2000 samples - at ../dataset/gen-word-3600-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 3500 max words, 2000 samples - at ../dataset/gen-word-3500-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 3800 max words, 2000 samples - at ../dataset/gen-word-3800-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 4000 max words, 2000 samples - at ../dataset/gen-word-4000-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 3700 max words, 2000 samples - at ../dataset/gen-word-3700-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 3900 max words, 2000 samples - at ../dataset/gen-word-3900-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 4500 max words, 2000 samples - at ../dataset/gen-word-4500-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 4300 max words, 2000 samples - at ../dataset/gen-word-4300-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 4400 max words, 2000 samples - at ../dataset/gen-word-4400-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 5300 max words, 2000 samples - at ../dataset/gen-word-5300-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 4600 max words, 2000 samples - at ../dataset/gen-word-4600-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 4800 max words, 2000 samples - at ../dataset/gen-word-4800-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 4200 max words, 2000 samples - at ../dataset/gen-word-4200-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 5000 max words, 2000 samples - at ../dataset/gen-word-5000-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 4700 max words, 2000 samples - at ../dataset/gen-word-4700-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 5200 max words, 2000 samples - at ../dataset/gen-word-5200-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 5700 max words, 2000 samples - at ../dataset/gen-word-5700-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 5600 max words, 2000 samples - at ../dataset/gen-word-5600-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 5100 max words, 2000 samples - at ../dataset/gen-word-5100-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 6800 max words, 2000 samples - at ../dataset/gen-word-6800-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 4900 max words, 2000 samples - at ../dataset/gen-word-4900-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 5400 max words, 2000 samples - at ../dataset/gen-word-5400-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 5800 max words, 2000 samples - at ../dataset/gen-word-5800-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 6900 max words, 2000 samples - at ../dataset/gen-word-6900-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 6300 max words, 2000 samples - at ../dataset/gen-word-6300-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 6400 max words, 2000 samples - at ../dataset/gen-word-6400-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 5900 max words, 2000 samples - at ../dataset/gen-word-5900-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 6200 max words, 2000 samples - at ../dataset/gen-word-6200-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 5500 max words, 2000 samples - at ../dataset/gen-word-5500-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 7800 max words, 2000 samples - at ../dataset/gen-word-7800-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 6500 max words, 2000 samples - at ../dataset/gen-word-6500-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 6700 max words, 2000 samples - at ../dataset/gen-word-6700-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 6100 max words, 2000 samples - at ../dataset/gen-word-6100-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 7000 max words, 2000 samples - at ../dataset/gen-word-7000-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 7200 max words, 2000 samples - at ../dataset/gen-word-7200-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 6000 max words, 2000 samples - at ../dataset/gen-word-6000-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 6600 max words, 2000 samples - at ../dataset/gen-word-6600-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 7100 max words, 2000 samples - at ../dataset/gen-word-7100-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 7400 max words, 2000 samples - at ../dataset/gen-word-7400-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 7300 max words, 2000 samples - at ../dataset/gen-word-7300-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 7500 max words, 2000 samples - at ../dataset/gen-word-7500-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 7900 max words, 2000 samples - at ../dataset/gen-word-7900-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 8000 max words, 2000 samples - at ../dataset/gen-word-8000-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 7700 max words, 2000 samples - at ../dataset/gen-word-7700-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Generated JSONL file with - 7600 max words, 2000 samples - at ../dataset/gen-word-7600-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "## Done ##\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "total 6.1G\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 11K Sep 2 06:17 gen-word-10-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 56K Sep 2 06:17 gen-word-100-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 487K Sep 2 06:17 gen-word-1000-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 58K Sep 2 06:17 gen-word-105-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 58K Sep 2 06:17 gen-word-110-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 22M Sep 2 06:17 gen-word-1100-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 62K Sep 2 06:17 gen-word-115-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 63K Sep 2 06:17 gen-word-120-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 23M Sep 2 06:17 gen-word-1200-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 65K Sep 2 06:17 gen-word-125-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 68K Sep 2 06:17 gen-word-130-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 25M Sep 2 06:17 gen-word-1300-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 72K Sep 2 06:17 gen-word-135-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 73K Sep 2 06:17 gen-word-140-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27M Sep 2 06:17 gen-word-1400-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 76K Sep 2 06:17 gen-word-145-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 13K Sep 2 06:17 gen-word-15-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 81K Sep 2 06:17 gen-word-150-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 29M Sep 2 06:17 gen-word-1500-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 82K Sep 2 06:17 gen-word-155-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 83K Sep 2 06:17 gen-word-160-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 31M Sep 2 06:17 gen-word-1600-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 87K Sep 2 06:17 gen-word-165-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 86K Sep 2 06:17 gen-word-170-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 33M Sep 2 06:17 gen-word-1700-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 88K Sep 2 06:17 gen-word-175-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 93K Sep 2 06:17 gen-word-180-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 35M Sep 2 06:17 gen-word-1800-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 92K Sep 2 06:17 gen-word-185-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 99K Sep 2 06:17 gen-word-190-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 37M Sep 2 06:17 gen-word-1900-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 102K Sep 2 06:17 gen-word-195-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 15K Sep 2 06:17 gen-word-20-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 103K Sep 2 06:17 gen-word-200-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 39M Sep 2 06:17 gen-word-2000-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 103K Sep 2 06:17 gen-word-205-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 108K Sep 2 06:17 gen-word-210-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 41M Sep 2 06:17 gen-word-2100-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 113K Sep 2 06:17 gen-word-215-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 114K Sep 2 06:17 gen-word-220-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 42M Sep 2 06:17 gen-word-2200-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 116K Sep 2 06:17 gen-word-225-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 114K Sep 2 06:17 gen-word-230-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 44M Sep 2 06:17 gen-word-2300-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 117K Sep 2 06:17 gen-word-235-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 124K Sep 2 06:17 gen-word-240-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 46M Sep 2 06:17 gen-word-2400-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 122K Sep 2 06:17 gen-word-245-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 17K Sep 2 06:17 gen-word-25-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 125K Sep 2 06:17 gen-word-250-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 48M Sep 2 06:17 gen-word-2500-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 128K Sep 2 06:17 gen-word-255-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 134K Sep 2 06:17 gen-word-260-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 50M Sep 2 06:17 gen-word-2600-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 135K Sep 2 06:17 gen-word-265-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 134K Sep 2 06:17 gen-word-270-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 52M Sep 2 06:17 gen-word-2700-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 140K Sep 2 06:17 gen-word-275-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 143K Sep 2 06:17 gen-word-280-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 54M Sep 2 06:17 gen-word-2800-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 141K Sep 2 06:17 gen-word-285-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 144K Sep 2 06:17 gen-word-290-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 56M Sep 2 06:17 gen-word-2900-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 150K Sep 2 06:17 gen-word-295-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 20K Sep 2 06:17 gen-word-30-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 155K Sep 2 06:17 gen-word-300-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 58M Sep 2 06:17 gen-word-3000-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 155K Sep 2 06:17 gen-word-305-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 156K Sep 2 06:17 gen-word-310-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 60M Sep 2 06:17 gen-word-3100-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 157K Sep 2 06:17 gen-word-315-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 159K Sep 2 06:17 gen-word-320-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 61M Sep 2 06:17 gen-word-3200-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 166K Sep 2 06:17 gen-word-325-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 166K Sep 2 06:17 gen-word-330-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 63M Sep 2 06:17 gen-word-3300-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 165K Sep 2 06:17 gen-word-335-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 171K Sep 2 06:17 gen-word-340-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 65M Sep 2 06:17 gen-word-3400-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 171K Sep 2 06:17 gen-word-345-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 21K Sep 2 06:17 gen-word-35-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 175K Sep 2 06:17 gen-word-350-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 67M Sep 2 06:17 gen-word-3500-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 175K Sep 2 06:17 gen-word-355-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 180K Sep 2 06:17 gen-word-360-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 69M Sep 2 06:17 gen-word-3600-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 181K Sep 2 06:17 gen-word-365-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 185K Sep 2 06:17 gen-word-370-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 71M Sep 2 06:17 gen-word-3700-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 190K Sep 2 06:17 gen-word-375-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 189K Sep 2 06:17 gen-word-380-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 73M Sep 2 06:17 gen-word-3800-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 190K Sep 2 06:17 gen-word-385-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 195K Sep 2 06:17 gen-word-390-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 75M Sep 2 06:17 gen-word-3900-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 199K Sep 2 06:17 gen-word-395-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 25K Sep 2 06:17 gen-word-40-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 200K Sep 2 06:17 gen-word-400-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 77M Sep 2 06:17 gen-word-4000-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 205K Sep 2 06:17 gen-word-405-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 208K Sep 2 06:17 gen-word-410-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 79M Sep 2 06:17 gen-word-4100-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 206K Sep 2 06:17 gen-word-415-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 212K Sep 2 06:17 gen-word-420-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 80M Sep 2 06:17 gen-word-4200-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 214K Sep 2 06:17 gen-word-425-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 214K Sep 2 06:17 gen-word-430-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 82M Sep 2 06:17 gen-word-4300-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 214K Sep 2 06:17 gen-word-435-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 219K Sep 2 06:17 gen-word-440-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 84M Sep 2 06:17 gen-word-4400-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 222K Sep 2 06:17 gen-word-445-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 28K Sep 2 06:17 gen-word-45-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 223K Sep 2 06:17 gen-word-450-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 86M Sep 2 06:17 gen-word-4500-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 226K Sep 2 06:17 gen-word-455-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 231K Sep 2 06:17 gen-word-460-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 88M Sep 2 06:17 gen-word-4600-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 233K Sep 2 06:17 gen-word-465-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 230K Sep 2 06:17 gen-word-470-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 90M Sep 2 06:17 gen-word-4700-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 238K Sep 2 06:17 gen-word-475-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 236K Sep 2 06:17 gen-word-480-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 92M Sep 2 06:17 gen-word-4800-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 241K Sep 2 06:17 gen-word-485-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 239K Sep 2 06:17 gen-word-490-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 94M Sep 2 06:17 gen-word-4900-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 245K Sep 2 06:17 gen-word-495-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 7.3K Sep 2 06:17 gen-word-5-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 31K Sep 2 06:17 gen-word-50-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 248K Sep 2 06:17 gen-word-500-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 96M Sep 2 06:17 gen-word-5000-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 252K Sep 2 06:17 gen-word-505-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 256K Sep 2 06:17 gen-word-510-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 98M Sep 2 06:17 gen-word-5100-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 261K Sep 2 06:17 gen-word-515-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 262K Sep 2 06:17 gen-word-520-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 99M Sep 2 06:17 gen-word-5200-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 261K Sep 2 06:17 gen-word-525-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 262K Sep 2 06:17 gen-word-530-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 101M Sep 2 06:17 gen-word-5300-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 266K Sep 2 06:17 gen-word-535-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 264K Sep 2 06:17 gen-word-540-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 103M Sep 2 06:17 gen-word-5400-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 273K Sep 2 06:17 gen-word-545-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 31K Sep 2 06:17 gen-word-55-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 277K Sep 2 06:17 gen-word-550-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 105M Sep 2 06:17 gen-word-5500-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 278K Sep 2 06:17 gen-word-555-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 272K Sep 2 06:17 gen-word-560-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 107M Sep 2 06:17 gen-word-5600-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 282K Sep 2 06:17 gen-word-565-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 283K Sep 2 06:17 gen-word-570-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 109M Sep 2 06:17 gen-word-5700-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 285K Sep 2 06:17 gen-word-575-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 290K Sep 2 06:17 gen-word-580-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 111M Sep 2 06:17 gen-word-5800-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 286K Sep 2 06:17 gen-word-585-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 290K Sep 2 06:17 gen-word-590-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 113M Sep 2 06:17 gen-word-5900-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 291K Sep 2 06:17 gen-word-595-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 35K Sep 2 06:17 gen-word-60-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 302K Sep 2 06:17 gen-word-600-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 115M Sep 2 06:17 gen-word-6000-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 298K Sep 2 06:17 gen-word-605-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 304K Sep 2 06:17 gen-word-610-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 117M Sep 2 06:17 gen-word-6100-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 306K Sep 2 06:17 gen-word-615-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 304K Sep 2 06:17 gen-word-620-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 118M Sep 2 06:17 gen-word-6200-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 306K Sep 2 06:17 gen-word-625-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 310K Sep 2 06:17 gen-word-630-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 120M Sep 2 06:17 gen-word-6300-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 312K Sep 2 06:17 gen-word-635-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 317K Sep 2 06:17 gen-word-640-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 122M Sep 2 06:17 gen-word-6400-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 322K Sep 2 06:17 gen-word-645-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 36K Sep 2 06:17 gen-word-65-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 322K Sep 2 06:17 gen-word-650-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 124M Sep 2 06:17 gen-word-6500-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 328K Sep 2 06:17 gen-word-655-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 325K Sep 2 06:17 gen-word-660-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 126M Sep 2 06:17 gen-word-6600-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 329K Sep 2 06:17 gen-word-665-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 334K Sep 2 06:17 gen-word-670-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 128M Sep 2 06:17 gen-word-6700-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 332K Sep 2 06:17 gen-word-675-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 333K Sep 2 06:17 gen-word-680-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 130M Sep 2 06:17 gen-word-6800-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 341K Sep 2 06:17 gen-word-685-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 338K Sep 2 06:17 gen-word-690-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 132M Sep 2 06:17 gen-word-6900-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 340K Sep 2 06:17 gen-word-695-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 39K Sep 2 06:17 gen-word-70-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 342K Sep 2 06:17 gen-word-700-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 134M Sep 2 06:17 gen-word-7000-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 351K Sep 2 06:17 gen-word-705-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 347K Sep 2 06:17 gen-word-710-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 136M Sep 2 06:17 gen-word-7100-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 352K Sep 2 06:17 gen-word-715-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 355K Sep 2 06:17 gen-word-720-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 137M Sep 2 06:17 gen-word-7200-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 354K Sep 2 06:17 gen-word-725-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 358K Sep 2 06:17 gen-word-730-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 139M Sep 2 06:17 gen-word-7300-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 372K Sep 2 06:17 gen-word-735-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 363K Sep 2 06:17 gen-word-740-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 141M Sep 2 06:17 gen-word-7400-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 364K Sep 2 06:17 gen-word-745-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 42K Sep 2 06:17 gen-word-75-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 372K Sep 2 06:17 gen-word-750-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 143M Sep 2 06:17 gen-word-7500-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 366K Sep 2 06:17 gen-word-755-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 376K Sep 2 06:17 gen-word-760-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 145M Sep 2 06:17 gen-word-7600-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 372K Sep 2 06:17 gen-word-765-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 381K Sep 2 06:17 gen-word-770-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 147M Sep 2 06:17 gen-word-7700-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 387K Sep 2 06:17 gen-word-775-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 381K Sep 2 06:17 gen-word-780-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 149M Sep 2 06:17 gen-word-7800-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 381K Sep 2 06:17 gen-word-785-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 390K Sep 2 06:17 gen-word-790-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 151M Sep 2 06:17 gen-word-7900-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 391K Sep 2 06:17 gen-word-795-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 44K Sep 2 06:17 gen-word-80-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 391K Sep 2 06:17 gen-word-800-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 153M Sep 2 06:17 gen-word-8000-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 396K Sep 2 06:17 gen-word-805-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 400K Sep 2 06:17 gen-word-810-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 398K Sep 2 06:17 gen-word-815-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 409K Sep 2 06:17 gen-word-820-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 403K Sep 2 06:17 gen-word-825-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 409K Sep 2 06:17 gen-word-830-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 411K Sep 2 06:17 gen-word-835-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 414K Sep 2 06:17 gen-word-840-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 413K Sep 2 06:17 gen-word-845-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 46K Sep 2 06:17 gen-word-85-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 412K Sep 2 06:17 gen-word-850-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 418K Sep 2 06:17 gen-word-855-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 417K Sep 2 06:17 gen-word-860-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 432K Sep 2 06:17 gen-word-865-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 429K Sep 2 06:17 gen-word-870-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 429K Sep 2 06:17 gen-word-875-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 432K Sep 2 06:17 gen-word-880-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 437K Sep 2 06:17 gen-word-885-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 432K Sep 2 06:17 gen-word-890-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 446K Sep 2 06:17 gen-word-895-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 49K Sep 2 06:17 gen-word-90-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 448K Sep 2 06:17 gen-word-900-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 448K Sep 2 06:17 gen-word-905-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 441K Sep 2 06:17 gen-word-910-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 449K Sep 2 06:17 gen-word-915-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 450K Sep 2 06:17 gen-word-920-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 460K Sep 2 06:17 gen-word-925-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 464K Sep 2 06:17 gen-word-930-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 458K Sep 2 06:17 gen-word-935-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 465K Sep 2 06:17 gen-word-940-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 467K Sep 2 06:17 gen-word-945-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 53K Sep 2 06:17 gen-word-95-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 466K Sep 2 06:17 gen-word-950-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 464K Sep 2 06:17 gen-word-955-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 473K Sep 2 06:17 gen-word-960-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 474K Sep 2 06:17 gen-word-965-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 475K Sep 2 06:17 gen-word-970-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 477K Sep 2 06:17 gen-word-975-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 483K Sep 2 06:17 gen-word-980-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 485K Sep 2 06:17 gen-word-985-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 486K Sep 2 06:17 gen-word-990-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 486K Sep 2 06:17 gen-word-995-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 52K Sep 2 06:17 shuffle-word-10-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 28K Sep 2 06:17 shuffle-word-100-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 26K Sep 2 06:17 shuffle-word-1000-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 28K Sep 2 06:17 shuffle-word-105-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 28K Sep 2 06:17 shuffle-word-110-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 523K Sep 2 06:17 shuffle-word-1100-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 28K Sep 2 06:17 shuffle-word-115-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 28K Sep 2 06:17 shuffle-word-120-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 523K Sep 2 06:17 shuffle-word-1200-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-125-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 29K Sep 2 06:17 shuffle-word-130-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 519K Sep 2 06:17 shuffle-word-1300-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 28K Sep 2 06:17 shuffle-word-135-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 28K Sep 2 06:17 shuffle-word-140-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 518K Sep 2 06:17 shuffle-word-1400-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-145-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 45K Sep 2 06:17 shuffle-word-15-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 29K Sep 2 06:17 shuffle-word-150-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 520K Sep 2 06:17 shuffle-word-1500-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-155-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 29K Sep 2 06:17 shuffle-word-160-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 520K Sep 2 06:17 shuffle-word-1600-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 28K Sep 2 06:17 shuffle-word-165-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 28K Sep 2 06:17 shuffle-word-170-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 522K Sep 2 06:17 shuffle-word-1700-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-175-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-180-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 522K Sep 2 06:17 shuffle-word-1800-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 29K Sep 2 06:17 shuffle-word-185-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-190-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 520K Sep 2 06:17 shuffle-word-1900-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 28K Sep 2 06:17 shuffle-word-195-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 40K Sep 2 06:17 shuffle-word-20-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-200-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 521K Sep 2 06:17 shuffle-word-2000-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-205-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 28K Sep 2 06:17 shuffle-word-210-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 522K Sep 2 06:17 shuffle-word-2100-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 28K Sep 2 06:17 shuffle-word-215-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-220-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 520K Sep 2 06:17 shuffle-word-2200-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 26K Sep 2 06:17 shuffle-word-225-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 28K Sep 2 06:17 shuffle-word-230-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 520K Sep 2 06:17 shuffle-word-2300-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-235-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-240-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 522K Sep 2 06:17 shuffle-word-2400-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 28K Sep 2 06:17 shuffle-word-245-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 35K Sep 2 06:17 shuffle-word-25-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 26K Sep 2 06:17 shuffle-word-250-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 518K Sep 2 06:17 shuffle-word-2500-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-255-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 28K Sep 2 06:17 shuffle-word-260-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 516K Sep 2 06:17 shuffle-word-2600-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-265-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-270-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 510K Sep 2 06:17 shuffle-word-2700-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-275-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-280-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 509K Sep 2 06:17 shuffle-word-2800-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-285-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 26K Sep 2 06:17 shuffle-word-290-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 509K Sep 2 06:17 shuffle-word-2900-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-295-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 34K Sep 2 06:17 shuffle-word-30-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 28K Sep 2 06:17 shuffle-word-300-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 509K Sep 2 06:17 shuffle-word-3000-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 28K Sep 2 06:17 shuffle-word-305-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-310-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 509K Sep 2 06:17 shuffle-word-3100-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-315-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-320-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 509K Sep 2 06:17 shuffle-word-3200-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 26K Sep 2 06:17 shuffle-word-325-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-330-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 508K Sep 2 06:17 shuffle-word-3300-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 28K Sep 2 06:17 shuffle-word-335-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-340-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 509K Sep 2 06:17 shuffle-word-3400-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-345-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 33K Sep 2 06:17 shuffle-word-35-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-350-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 509K Sep 2 06:17 shuffle-word-3500-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-355-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-360-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 509K Sep 2 06:17 shuffle-word-3600-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-365-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 26K Sep 2 06:17 shuffle-word-370-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 509K Sep 2 06:17 shuffle-word-3700-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-375-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-380-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 509K Sep 2 06:17 shuffle-word-3800-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-385-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-390-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 509K Sep 2 06:17 shuffle-word-3900-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-395-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 32K Sep 2 06:17 shuffle-word-40-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 26K Sep 2 06:17 shuffle-word-400-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 509K Sep 2 06:17 shuffle-word-4000-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-405-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-410-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 509K Sep 2 06:17 shuffle-word-4100-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 26K Sep 2 06:17 shuffle-word-415-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 26K Sep 2 06:17 shuffle-word-420-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 509K Sep 2 06:17 shuffle-word-4200-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-425-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-430-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 509K Sep 2 06:17 shuffle-word-4300-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 26K Sep 2 06:17 shuffle-word-435-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-440-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 509K Sep 2 06:17 shuffle-word-4400-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 26K Sep 2 06:17 shuffle-word-445-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 33K Sep 2 06:17 shuffle-word-45-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-450-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 509K Sep 2 06:17 shuffle-word-4500-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 28K Sep 2 06:17 shuffle-word-455-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-460-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 508K Sep 2 06:17 shuffle-word-4600-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-465-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 28K Sep 2 06:17 shuffle-word-470-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 509K Sep 2 06:17 shuffle-word-4700-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-475-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-480-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 509K Sep 2 06:17 shuffle-word-4800-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 28K Sep 2 06:17 shuffle-word-485-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-490-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 509K Sep 2 06:17 shuffle-word-4900-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-495-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 85K Sep 2 06:17 shuffle-word-5-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 32K Sep 2 06:17 shuffle-word-50-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-500-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 508K Sep 2 06:17 shuffle-word-5000-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-505-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-510-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 509K Sep 2 06:17 shuffle-word-5100-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-515-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 26K Sep 2 06:17 shuffle-word-520-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 509K Sep 2 06:17 shuffle-word-5200-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 26K Sep 2 06:17 shuffle-word-525-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-530-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 509K Sep 2 06:17 shuffle-word-5300-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-535-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-540-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 508K Sep 2 06:17 shuffle-word-5400-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-545-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 31K Sep 2 06:17 shuffle-word-55-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-550-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 508K Sep 2 06:17 shuffle-word-5500-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 28K Sep 2 06:17 shuffle-word-555-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 26K Sep 2 06:17 shuffle-word-560-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 509K Sep 2 06:17 shuffle-word-5600-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-565-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-570-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 509K Sep 2 06:17 shuffle-word-5700-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-575-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-580-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 509K Sep 2 06:17 shuffle-word-5800-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 26K Sep 2 06:17 shuffle-word-585-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-590-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 509K Sep 2 06:17 shuffle-word-5900-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 26K Sep 2 06:17 shuffle-word-595-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 30K Sep 2 06:17 shuffle-word-60-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-600-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 509K Sep 2 06:17 shuffle-word-6000-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-605-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-610-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 509K Sep 2 06:17 shuffle-word-6100-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-615-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-620-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 509K Sep 2 06:17 shuffle-word-6200-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 26K Sep 2 06:17 shuffle-word-625-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-630-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 509K Sep 2 06:17 shuffle-word-6300-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-635-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 26K Sep 2 06:17 shuffle-word-640-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 509K Sep 2 06:17 shuffle-word-6400-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-645-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 31K Sep 2 06:17 shuffle-word-65-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-650-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 509K Sep 2 06:17 shuffle-word-6500-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 26K Sep 2 06:17 shuffle-word-655-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 26K Sep 2 06:17 shuffle-word-660-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 508K Sep 2 06:17 shuffle-word-6600-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-665-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-670-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 508K Sep 2 06:17 shuffle-word-6700-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-675-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 26K Sep 2 06:17 shuffle-word-680-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 509K Sep 2 06:17 shuffle-word-6800-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-685-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-690-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 509K Sep 2 06:17 shuffle-word-6900-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-695-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 30K Sep 2 06:17 shuffle-word-70-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-700-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 508K Sep 2 06:17 shuffle-word-7000-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-705-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 26K Sep 2 06:17 shuffle-word-710-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 509K Sep 2 06:17 shuffle-word-7100-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 26K Sep 2 06:17 shuffle-word-715-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-720-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 509K Sep 2 06:17 shuffle-word-7200-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 26K Sep 2 06:17 shuffle-word-725-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-730-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 509K Sep 2 06:17 shuffle-word-7300-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 26K Sep 2 06:17 shuffle-word-735-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-740-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 508K Sep 2 06:17 shuffle-word-7400-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 26K Sep 2 06:17 shuffle-word-745-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 29K Sep 2 06:17 shuffle-word-75-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 26K Sep 2 06:17 shuffle-word-750-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 509K Sep 2 06:17 shuffle-word-7500-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-755-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 26K Sep 2 06:17 shuffle-word-760-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 509K Sep 2 06:17 shuffle-word-7600-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 26K Sep 2 06:17 shuffle-word-765-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-770-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 509K Sep 2 06:17 shuffle-word-7700-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 26K Sep 2 06:17 shuffle-word-775-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 26K Sep 2 06:17 shuffle-word-780-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 509K Sep 2 06:17 shuffle-word-7800-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-785-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-790-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 509K Sep 2 06:17 shuffle-word-7900-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-795-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 29K Sep 2 06:17 shuffle-word-80-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-800-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 509K Sep 2 06:17 shuffle-word-8000-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 26K Sep 2 06:17 shuffle-word-805-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-810-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 26K Sep 2 06:17 shuffle-word-815-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-820-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-825-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-830-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-835-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 26K Sep 2 06:17 shuffle-word-840-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-845-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 29K Sep 2 06:17 shuffle-word-85-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 28K Sep 2 06:17 shuffle-word-850-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-855-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 26K Sep 2 06:17 shuffle-word-860-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 26K Sep 2 06:17 shuffle-word-865-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 26K Sep 2 06:17 shuffle-word-870-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-875-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 26K Sep 2 06:17 shuffle-word-880-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-885-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-890-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-895-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 29K Sep 2 06:17 shuffle-word-90-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-900-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-905-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-910-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-915-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 26K Sep 2 06:17 shuffle-word-920-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-925-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-930-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 26K Sep 2 06:17 shuffle-word-935-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 26K Sep 2 06:17 shuffle-word-940-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 26K Sep 2 06:17 shuffle-word-945-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-95-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 26K Sep 2 06:17 shuffle-word-950-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-955-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-960-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-965-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 26K Sep 2 06:17 shuffle-word-970-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-975-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-980-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 26K Sep 2 06:17 shuffle-word-985-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 26K Sep 2 06:17 shuffle-word-990-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 27K Sep 2 06:17 shuffle-word-995-count.jsonl\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "-rw-r--r-- 1 root root 6.0K Sep 2 06:17 word-2-count.jsonl\n" + ] + } + ], + "source": [ + "%%script bash\n", + "\n", + "########################################\n", + "# Generate the required jsonl dataset\n", + "########################################\n", + "\n", + "# Reset the dataset dir\n", + "mkdir -p ../dataset\n", + "rm -rf ../dataset/*.jsonl\n", + "\n", + "# Generate the various datasets\n", + "echo \"## Generating word reptition dataset ##\"\n", + "\n", + "#\n", + "# We reduce the training set for < 50 words - and shift the focus upwards\n", + "# (aka 50-100 token * 2 : ~100 - 250 token ctx len)\n", + "#\n", + "python ../memory_script/gen_limited_prompt_completion_jsonl.py ../dataset/word-2-count.jsonl 2 50 &\n", + "for i in {5..1000..5} \n", + "do\n", + " python ../memory_script/gen_limited_prompt_completion_jsonl.py ../dataset/gen-word-$i-count.jsonl $i 50 & \n", + " python ../memory_script/shuffle_limited_prompt_completion_jsonl.py ../dataset/shuffle-word-$i-count.jsonl $i 1 & \n", + "done\n", + "\n", + "#\n", + "# Ramping up the 50+ - 4200 words dataset\n", + "# \n", + "for i in {1100..8000..100} \n", + "do\n", + " python ../memory_script/gen_limited_prompt_completion_jsonl.py ../dataset/gen-word-$i-count.jsonl $i 2000 & \n", + " python ../memory_script/shuffle_limited_prompt_completion_jsonl.py ../dataset/shuffle-word-$i-count.jsonl $i 20 & \n", + "done\n", + "\n", + "wait\n", + "echo \"## Done ##\"\n", + "\n", + "ls -lh ../dataset/" + ] + }, + { + "cell_type": "code", + "execution_count": 6, + "id": "2eb840b5", + "metadata": { + "execution": { + "iopub.execute_input": "2023-09-02T06:17:49.255288Z", + "iopub.status.busy": "2023-09-02T06:17:49.254615Z" + }, + "papermill": { + "duration": null, + "end_time": null, + "exception": false, + "start_time": "2023-09-02T06:17:49.184772", + "status": "running" + }, + "tags": [] + }, + "outputs": [ + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Setting ds_accelerator to cuda (auto detect)\r\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "[RWKV.model] Running RWKV model using 'torch-jit' with torch '2.0.1+cu118'\r\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "/usr/local/lib/python3.11/dist-packages/lightning/pytorch/cli.py:484: UserWarning: LightningCLI's args parameter is intended to run from within Python like if it were from the command line. To prevent mistakes it is not recommended to provide both args and command line arguments, got: sys.argv[1:]=['fit', '-c', '/actions-runner/_work/RWKV-infctx-trainer/RWKV-infctx-trainer/notebook/experiment/rwkv-x-exp/v5-memory/config-mem-template.yaml', '--trainer.logger.init_args.name=v5-L96-D1024-E0.1 - Mem-Tune ctx-8k (train-ctx=4k, deepspeed_stage_1)', '--trainer.strategy=deepspeed_stage_1', '--trainer.devices=auto', '--trainer.callbacks.init_args.dirpath=../checkpoint/v5-L96-D1024-E0_1-mem-ctx-8k/', '--model.lr_init=3e-4', '--model.lr_final=1e-4', '--data.max_token_size=8192', '--model.ctx_len=4096', '--model.bptt_learning_range=2', '--model.load_model=../model/v5-L96-D1024-E0_1-mem-ctx-4k.pth'], args=['fit', '-c', '/actions-runner/_work/RWKV-infctx-trainer/RWKV-infctx-trainer/notebook/experiment/rwkv-x-exp/v5-memory/config-mem-template.yaml', '--trainer.logger.init_args.name=v5-L96-D1024-E0.1 - Mem-Tune ctx-8k (train-ctx=4k, deepspeed_stage_1)', '--trainer.strategy=deepspeed_stage_1', '--trainer.devices=auto', '--trainer.callbacks.init_args.dirpath=../checkpoint/v5-L96-D1024-E0_1-mem-ctx-8k/', '--model.lr_init=3e-4', '--model.lr_final=1e-4', '--data.max_token_size=8192', '--model.ctx_len=4096', '--model.bptt_learning_range=2', '--model.load_model=../model/v5-L96-D1024-E0_1-mem-ctx-4k.pth'].\r\n", + " rank_zero_warn(\r\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "/usr/local/lib/python3.11/dist-packages/lightning/fabric/utilities/seed.py:39: UserWarning: No seed found, seed set to 4016710040\r\n", + " rank_zero_warn(f\"No seed found, seed set to {seed}\")\r\n", + "Global seed set to 4016710040\r\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\u001b[34m\u001b[1mwandb\u001b[0m: Currently logged in as: \u001b[33mpicocreator\u001b[0m (\u001b[33mrwkv-x-dev\u001b[0m). Use \u001b[1m`wandb login --relogin`\u001b[0m to force relogin\r\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\u001b[34m\u001b[1mwandb\u001b[0m: Tracking run with wandb version 0.15.9\r\n", + "\u001b[34m\u001b[1mwandb\u001b[0m: Run data is saved locally in \u001b[35m\u001b[1m./wandb/run-20230902_061755-f2kzi2o7\u001b[0m\r\n", + "\u001b[34m\u001b[1mwandb\u001b[0m: Run \u001b[1m`wandb offline`\u001b[0m to turn off syncing.\r\n", + "\u001b[34m\u001b[1mwandb\u001b[0m: Syncing run \u001b[33mv5-L96-D1024-E0.1 - Mem-Tune ctx-8k (train-ctx=4k, deepspeed_stage_1)\u001b[0m\r\n", + "\u001b[34m\u001b[1mwandb\u001b[0m: ⭐️ View project at \u001b[34m\u001b[4mhttps://wandb.ai/rwkv-x-dev/RWKV-X-Experiments\u001b[0m\r\n", + "\u001b[34m\u001b[1mwandb\u001b[0m: 🚀 View run at \u001b[34m\u001b[4mhttps://wandb.ai/rwkv-x-dev/RWKV-X-Experiments/runs/f2kzi2o7\u001b[0m\r\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "/usr/local/lib/python3.11/dist-packages/lightning/fabric/connector.py:562: UserWarning: bf16 is supported for historical reasons but its usage is discouraged. Please set your precision to bf16-mixed instead!\r\n", + " rank_zero_warn(\r\n", + "GPU available: True (cuda), used: True\r\n", + "TPU available: False, using: 0 TPU cores\r\n", + "IPU available: False, using: 0 IPUs\r\n", + "HPU available: False, using: 0 HPUs\r\n", + "\r\n", + "\r\n", + "[RWKV.Trainer] Applying 'target_batch_size' with the following:\r\n", + " - target_batch_size: 256\r\n", + " - num_nodes: 1\r\n", + " - num_devices: 8\r\n", + " - accumulate_grad_batches: 32\r\n", + " - effective_batch_size: 256\r\n", + "\r\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "Resolving data files: 0%| | 0/541 [00:00=12.1), as this is known to have freeze issues\r\n", + "# - The terms used in wandb / the progress bar can be confusing, see the github README.md for beter clarifications\r\n", + "# - When resuming from checkpoint, the estimated time is inaccurate\r\n", + "#\r\n", + "\r\n", + "[RWKV.model] Configuring optimizer with\r\n", + " - lr_init: 3.000e-04 (0.0003)\r\n", + " - lr_final: 1.000e-04 (0.0001)\r\n", + "\r\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Using /root/.cache/torch_extensions/py311_cu118 as PyTorch extensions root...\r\n", + "Using /root/.cache/torch_extensions/py311_cu118 as PyTorch extensions root...\r\n", + "Using /root/.cache/torch_extensions/py311_cu118 as PyTorch extensions root...\r\n", + "Using /root/.cache/torch_extensions/py311_cu118 as PyTorch extensions root...\r\n", + "Using /root/.cache/torch_extensions/py311_cu118 as PyTorch extensions root...\r\n", + "Using /root/.cache/torch_extensions/py311_cu118 as PyTorch extensions root...\r\n", + "Using /root/.cache/torch_extensions/py311_cu118 as PyTorch extensions root...\r\n", + "Using /root/.cache/torch_extensions/py311_cu118 as PyTorch extensions root...\r\n", + "Detected CUDA files, patching ldflags\r\n", + "Emitting ninja build file /root/.cache/torch_extensions/py311_cu118/fused_adam/build.ninja...\r\n", + "Building extension module fused_adam...\r\n", + "Allowing ninja to set a default number of workers... (overridable by setting the environment variable MAX_JOBS=N)\r\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "ninja: no work to do.\r\n", + "Loading extension module fused_adam...\r\n", + "Time to load fused_adam op: 0.06581354141235352 seconds\r\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Loading extension module fused_adam...\r\n", + "Time to load fused_adam op: 0.10160326957702637 seconds\r\n", + "Loading extension module fused_adam...\r\n", + "Loading extension module fused_adam...\r\n", + "Loading extension module fused_adam...\r\n", + "Loading extension module fused_adam...\r\n", + "Time to load fused_adam op: 0.10122323036193848 seconds\r\n", + "Time to load fused_adam op: 0.10131621360778809 seconds\r\n", + "Time to load fused_adam op: 0.10161590576171875 seconds\r\n", + "Time to load fused_adam op: 0.10172796249389648 seconds\r\n", + "Loading extension module fused_adam...\r\n", + "Time to load fused_adam op: 0.10164880752563477 seconds\r\n", + "Loading `train_dataloader` to estimate number of stepping batches.\r\n", + "Loading extension module fused_adam...\r\n", + "Time to load fused_adam op: 0.10137581825256348 seconds\r\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Using /root/.cache/torch_extensions/py311_cu118 as PyTorch extensions root...\r\n", + "Using /root/.cache/torch_extensions/py311_cu118 as PyTorch extensions root...\r\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Using /root/.cache/torch_extensions/py311_cu118 as PyTorch extensions root...\r\n", + "Using /root/.cache/torch_extensions/py311_cu118 as PyTorch extensions root...\r\n", + "Using /root/.cache/torch_extensions/py311_cu118 as PyTorch extensions root...\r\n", + "Using /root/.cache/torch_extensions/py311_cu118 as PyTorch extensions root...\r\n", + "Using /root/.cache/torch_extensions/py311_cu118 as PyTorch extensions root...\r\n", + "Using /root/.cache/torch_extensions/py311_cu118 as PyTorch extensions root...\r\n", + "Emitting ninja build file /root/.cache/torch_extensions/py311_cu118/utils/build.ninja...\r\n", + "Building extension module utils...\r\n", + "Allowing ninja to set a default number of workers... (overridable by setting the environment variable MAX_JOBS=N)\r\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "ninja: no work to do.\r\n", + "Loading extension module utils...\r\n", + "Time to load utils op: 0.0746297836303711 seconds\r\n", + "Loading extension module utils...\r\n", + "Time to load utils op: 0.10225343704223633 seconds\r\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Loading extension module utils...\r\n", + "Loading extension module utils...\r\n", + "Time to load utils op: 0.1022336483001709 seconds\r\n", + "Time to load utils op: 0.10247349739074707 seconds\r\n", + "Loading extension module utils...\r\n", + "Loading extension module utils...\r\n", + "Time to load utils op: 0.1032721996307373 seconds\r\n", + "Time to load utils op: 0.10237407684326172 seconds\r\n", + "Loading extension module utils...\r\n", + "Time to load utils op: 0.10254526138305664 seconds\r\n", + "Loading extension module utils...\r\n", + "Time to load utils op: 0.10282373428344727 seconds\r\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Rank: 5 partition count [8, 8] and sizes[(176584448, False), (384, False)] \r\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Rank: 1 partition count [8, 8] and sizes[(176584448, False), (384, False)] " + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Rank: 2 partition count [8, 8] and sizes[(176584448, False), (384, False)] \r\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Rank: 4 partition count [8, 8] and sizes[(176584448, False), (384, False)] \r\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Rank: 7 partition count [8, 8] and sizes[(176584448, False), (384, False)] \r\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Rank: 0 partition count [8, 8] and sizes[(176584448, False), (384, False)] \r\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Rank: 3 partition count [8, 8] and sizes[(176584448, False), (384, False)] \r\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Rank: 6 partition count [8, 8] and sizes[(176584448, False), (384, False)] \r\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Using /root/.cache/torch_extensions/py311_cu118 as PyTorch extensions root...\r\n", + "No modifications detected for re-loaded extension module utils, skipping build step...\r\n", + "Loading extension module utils...\r\n", + "Time to load utils op: 0.0006437301635742188 seconds\r\n", + "Using /root/.cache/torch_extensions/py311_cu118 as PyTorch extensions root...\r\n", + "No modifications detected for re-loaded extension module utils, skipping build step...\r\n", + "Loading extension module utils...\r\n", + "Time to load utils op: 0.0009703636169433594 seconds\r\n", + "Using /root/.cache/torch_extensions/py311_cu118 as PyTorch extensions root...\r\n", + "Using /root/.cache/torch_extensions/py311_cu118 as PyTorch extensions root...\r\n", + "No modifications detected for re-loaded extension module utils, skipping build step...\r\n", + "Loading extension module utils...\r\n", + "Time to load utils op: 0.0006546974182128906 seconds\r\n", + "No modifications detected for re-loaded extension module utils, skipping build step...\r\n", + "Loading extension module utils...\r\n", + "Time to load utils op: 0.0006368160247802734 seconds\r\n", + "Using /root/.cache/torch_extensions/py311_cu118 as PyTorch extensions root...\r\n", + "No modifications detected for re-loaded extension module utils, skipping build step...\r\n", + "Loading extension module utils...\r\n", + "Time to load utils op: 0.0006356239318847656 seconds\r\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Using /root/.cache/torch_extensions/py311_cu118 as PyTorch extensions root...\r\n", + "No modifications detected for re-loaded extension module utils, skipping build step...\r\n", + "Loading extension module utils...\r\n", + "Time to load utils op: 0.0006177425384521484 seconds\r\n", + "Using /root/.cache/torch_extensions/py311_cu118 as PyTorch extensions root...\r\n", + "No modifications detected for re-loaded extension module utils, skipping build step...\r\n", + "Loading extension module utils...\r\n", + "Time to load utils op: 0.0008919239044189453 seconds\r\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Using /root/.cache/torch_extensions/py311_cu118 as PyTorch extensions root...\r\n", + "No modifications detected for re-loaded extension module utils, skipping build step...\r\n", + "Loading extension module utils...\r\n", + "Time to load utils op: 0.0009381771087646484 seconds\r\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r\n", + " | Name | Type | Params\r\n", + "--------------------------------------\r\n", + "0 | emb | Embedding | 51.5 M\r\n", + "1 | blocks | ModuleList | 1.3 B \r\n", + "2 | ln_out | LayerNorm | 2.0 K \r\n", + "3 | head | Linear | 51.5 M\r\n", + "--------------------------------------\r\n", + "1.4 B Trainable params\r\n", + "0 Non-trainable params\r\n", + "1.4 B Total params\r\n", + "5,650.715 Total estimated model params size (MB)\r\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "\r", + "Training: 0it [00:00, ?it/s]\r", + "Training: 0%| | 0/12323 [00:00\r\n", - " cli_main()\r\n", - " File \"/actions-runner/_work/RWKV-infctx-trainer/RWKV-infctx-trainer/RWKV-v5/lightning_trainer.py\", line 233, in cli_main\r\n", - " LightningCLI(\r\n", - " File \"/usr/local/lib/python3.11/dist-packages/lightning/pytorch/cli.py\", line 350, in __init__\r\n", - " self.instantiate_classes()\r\n", - " File \"/usr/local/lib/python3.11/dist-packages/lightning/pytorch/cli.py\", line 499, in instantiate_classes\r\n", - " self.config_init = self.parser.instantiate_classes(self.config)\r\n", - " ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n", - " File \"/usr/local/lib/python3.11/dist-packages/jsonargparse/_deprecated.py\", line 139, in patched_instantiate_classes\r\n", - " cfg = self._unpatched_instantiate_classes(cfg, **kwargs)\r\n", - " ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n", - " File \"/usr/local/lib/python3.11/dist-packages/jsonargparse/_core.py\", line 1128, in instantiate_classes\r\n", - " cfg[subcommand] = subparser.instantiate_classes(cfg[subcommand], instantiate_groups=instantiate_groups)\r\n", - " ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n", - " File \"/usr/local/lib/python3.11/dist-packages/jsonargparse/_deprecated.py\", line 139, in patched_instantiate_classes\r\n", - " cfg = self._unpatched_instantiate_classes(cfg, **kwargs)\r\n", - " ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n", - " File \"/usr/local/lib/python3.11/dist-packages/jsonargparse/_core.py\", line 1122, in instantiate_classes\r\n", - " component.instantiate_class(component, cfg)\r\n", - " File \"/usr/local/lib/python3.11/dist-packages/jsonargparse/_signatures.py\", line 551, in group_instantiate_class\r\n", - " parent[key] = group.group_class(**value)\r\n", - " ^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n", - " File \"/actions-runner/_work/RWKV-infctx-trainer/RWKV-infctx-trainer/RWKV-v5/src/model.py\", line 553, in __init__\r\n", - " raise ValueError(f\"load_model file '{load_model}' does not exist\")\r\n", - "ValueError: load_model file '../model/v5-L96-D1024-E0_1-mem-ctx-4k.pth' does not exist\r\n", - "\u001b[34m\u001b[1mwandb\u001b[0m: Waiting for W&B process to finish... \u001b[31m(failed 1).\u001b[0m Press Control-C to abort syncing.\r\n" - ] - }, - { - "name": "stdout", - "output_type": "stream", - "text": [ - "\u001b[34m\u001b[1mwandb\u001b[0m: 🚀 View run \u001b[33mv5-L96-D1024-E0.1 - Mem-Tune ctx-8k (train-ctx=4k, deepspeed_stage_1)\u001b[0m at: \u001b[34m\u001b[4mhttps://wandb.ai/rwkv-x-dev/RWKV-X-Experiments/runs/1c6n4316\u001b[0m\r\n", - "\u001b[34m\u001b[1mwandb\u001b[0m: ️⚡ View job at \u001b[34m\u001b[4mhttps://wandb.ai/rwkv-x-dev/RWKV-X-Experiments/jobs/QXJ0aWZhY3RDb2xsZWN0aW9uOjkzMjg5ODA3/version_details/v17\u001b[0m\r\n", - "\u001b[34m\u001b[1mwandb\u001b[0m: Synced 5 W&B file(s), 0 media file(s), 0 artifact file(s) and 0 other file(s)\r\n", - "\u001b[34m\u001b[1mwandb\u001b[0m: Find logs at: \u001b[35m\u001b[1m./wandb/run-20230901_145324-1c6n4316/logs\u001b[0m\r\n" + "\r", + "Epoch 0: 8%| | 1028/12323 [1:44:40<19:10:09, 6.11s/it, v_num=i2o7, train/loss\r", + "Epoch 0: 8%| | 1028/12323 [1:44:40<19:10:09, 6.11s/it, v_num=i2o7, train/loss" ] } ], @@ -8069,55 +35150,19 @@ }, { "cell_type": "code", - "execution_count": 7, - "id": "7ac5ad40", + "execution_count": null, + "id": "065aea13", "metadata": { - "execution": { - "iopub.execute_input": "2023-09-01T14:53:38.115546Z", - "iopub.status.busy": "2023-09-01T14:53:38.115356Z", - "iopub.status.idle": "2023-09-01T14:53:40.987634Z", - "shell.execute_reply": "2023-09-01T14:53:40.986824Z" - }, "papermill": { - "duration": 2.928097, - "end_time": "2023-09-01T14:53:40.989318", - "exception": false, - "start_time": "2023-09-01T14:53:38.061221", - "status": "completed" + "duration": null, + "end_time": null, + "exception": null, + "start_time": null, + "status": "pending" }, "tags": [] }, - "outputs": [ - { - "name": "stdout", - "output_type": "stream", - "text": [ - "Setting ds_accelerator to cuda (auto detect)\r\n" - ] - }, - { - "name": "stdout", - "output_type": "stream", - "text": [ - "Traceback (most recent call last):\r\n", - " File \"/actions-runner/_work/RWKV-infctx-trainer/RWKV-infctx-trainer/RWKV-v5/export_checkpoint.py\", line 651, in \r\n", - " convert_zero_checkpoint_to_fp32_state_dict(args.checkpoint_dir, output_file, save_dtype=args.dtype)\r\n", - " File \"/actions-runner/_work/RWKV-infctx-trainer/RWKV-infctx-trainer/RWKV-v5/export_checkpoint.py\", line 542, in convert_zero_checkpoint_to_fp32_state_dict\r\n", - " state_dict = get_fp32_state_dict_from_zero_checkpoint(checkpoint_dir, tag)\r\n", - " ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n", - " File \"/actions-runner/_work/RWKV-infctx-trainer/RWKV-infctx-trainer/RWKV-v5/export_checkpoint.py\", line 516, in get_fp32_state_dict_from_zero_checkpoint\r\n", - " raise ValueError(f\"Unable to find 'latest' file at {latest_path}\")\r\n", - "ValueError: Unable to find 'latest' file at ../checkpoint/v5-L96-D1024-E0_1-mem-ctx-8k/last.ckpt/latest\r\n" - ] - }, - { - "name": "stdout", - "output_type": "stream", - "text": [ - "ls: cannot access '../model/v5-L96-D1024-E0_1-mem-ctx-8k.pth': No such file or directory\r\n" - ] - } - ], + "outputs": [], "source": [ "# Lets export the model from the checkpoint\n", "!cd \"{TRAINER_DIR}\" && \\\n", @@ -8129,61 +35174,19 @@ }, { "cell_type": "code", - "execution_count": 8, - "id": "912275ec", + "execution_count": null, + "id": "e497d3dc", "metadata": { - "execution": { - "iopub.execute_input": "2023-09-01T14:53:41.101106Z", - "iopub.status.busy": "2023-09-01T14:53:41.100914Z", - "iopub.status.idle": "2023-09-01T14:53:46.113705Z", - "shell.execute_reply": "2023-09-01T14:53:46.112987Z" - }, "papermill": { - "duration": 5.068771, - "end_time": "2023-09-01T14:53:46.115376", - "exception": false, - "start_time": "2023-09-01T14:53:41.046605", - "status": "completed" + "duration": null, + "end_time": null, + "exception": null, + "start_time": null, + "status": "pending" }, "tags": [] }, - "outputs": [ - { - "name": "stdout", - "output_type": "stream", - "text": [ - "Setting ds_accelerator to cuda (auto detect)\r\n" - ] - }, - { - "name": "stdout", - "output_type": "stream", - "text": [ - "[RWKV.model] Running RWKV model using 'torch-jit' with torch '2.0.1+cu118'\r\n", - "Traceback (most recent call last):\r\n", - " File \"/actions-runner/_work/RWKV-infctx-trainer/RWKV-infctx-trainer/notebook/experiment/rwkv-x-exp/v5-memory/../memory_script/eval_v5_memory_guided.py\", line 366, in \r\n", - " asyncio.run(main_function())\r\n", - " File \"/usr/lib/python3.11/asyncio/runners.py\", line 190, in run\r\n", - " return runner.run(main)\r\n", - " ^^^^^^^^^^^^^^^^\r\n", - " File \"/usr/lib/python3.11/asyncio/runners.py\", line 118, in run\r\n", - " return self._loop.run_until_complete(task)\r\n", - " ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n", - " File \"/usr/lib/python3.11/asyncio/base_events.py\", line 653, in run_until_complete\r\n", - " return future.result()\r\n", - " ^^^^^^^^^^^^^^^\r\n", - " File \"/actions-runner/_work/RWKV-infctx-trainer/RWKV-infctx-trainer/notebook/experiment/rwkv-x-exp/v5-memory/../memory_script/eval_v5_memory_guided.py\", line 58, in main_function\r\n", - " model = SimpleRWKV(model_path, device=\"cuda\")\r\n", - " ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n", - " File \"/actions-runner/_work/RWKV-infctx-trainer/RWKV-infctx-trainer/RWKV-v5/src/model.py\", line 1378, in __init__\r\n", - " self.model = RWKV(**model_config)\r\n", - " ^^^^^^^^^^^^^^^^^^^^\r\n", - " File \"/actions-runner/_work/RWKV-infctx-trainer/RWKV-infctx-trainer/RWKV-v5/src/model.py\", line 553, in __init__\r\n", - " raise ValueError(f\"load_model file '{load_model}' does not exist\")\r\n", - "ValueError: load_model file '/actions-runner/_work/RWKV-infctx-trainer/RWKV-infctx-trainer/model/v5-L96-D1024-E0_1-mem-ctx-8k.pth' does not exist\r\n" - ] - } - ], + "outputs": [], "source": [ "# Lets do a quick memory test\n", "!export RWKV_WAVENET_LAYERS=\"{RWKV_WAVENET_LAYERS}\" && \\\n", @@ -8192,61 +35195,19 @@ }, { "cell_type": "code", - "execution_count": 9, - "id": "71ea6284", + "execution_count": null, + "id": "e5ae95c7", "metadata": { - "execution": { - "iopub.execute_input": "2023-09-01T14:53:46.225301Z", - "iopub.status.busy": "2023-09-01T14:53:46.225111Z", - "iopub.status.idle": "2023-09-01T14:53:51.386811Z", - "shell.execute_reply": "2023-09-01T14:53:51.386023Z" - }, "papermill": { - "duration": 5.217798, - "end_time": "2023-09-01T14:53:51.388460", - "exception": false, - "start_time": "2023-09-01T14:53:46.170662", - "status": "completed" + "duration": null, + "end_time": null, + "exception": null, + "start_time": null, + "status": "pending" }, "tags": [] }, - "outputs": [ - { - "name": "stdout", - "output_type": "stream", - "text": [ - "Setting ds_accelerator to cuda (auto detect)\r\n" - ] - }, - { - "name": "stdout", - "output_type": "stream", - "text": [ - "[RWKV.model] Running RWKV model using 'torch-jit' with torch '2.0.1+cu118'\r\n", - "Traceback (most recent call last):\r\n", - " File \"/actions-runner/_work/RWKV-infctx-trainer/RWKV-infctx-trainer/notebook/experiment/rwkv-x-exp/v5-memory/../memory_script/eval_v5_memory_guided.py\", line 366, in \r\n", - " asyncio.run(main_function())\r\n", - " File \"/usr/lib/python3.11/asyncio/runners.py\", line 190, in run\r\n", - " return runner.run(main)\r\n", - " ^^^^^^^^^^^^^^^^\r\n", - " File \"/usr/lib/python3.11/asyncio/runners.py\", line 118, in run\r\n", - " return self._loop.run_until_complete(task)\r\n", - " ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n", - " File \"/usr/lib/python3.11/asyncio/base_events.py\", line 653, in run_until_complete\r\n", - " return future.result()\r\n", - " ^^^^^^^^^^^^^^^\r\n", - " File \"/actions-runner/_work/RWKV-infctx-trainer/RWKV-infctx-trainer/notebook/experiment/rwkv-x-exp/v5-memory/../memory_script/eval_v5_memory_guided.py\", line 58, in main_function\r\n", - " model = SimpleRWKV(model_path, device=\"cuda\")\r\n", - " ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n", - " File \"/actions-runner/_work/RWKV-infctx-trainer/RWKV-infctx-trainer/RWKV-v5/src/model.py\", line 1378, in __init__\r\n", - " self.model = RWKV(**model_config)\r\n", - " ^^^^^^^^^^^^^^^^^^^^\r\n", - " File \"/actions-runner/_work/RWKV-infctx-trainer/RWKV-infctx-trainer/RWKV-v5/src/model.py\", line 553, in __init__\r\n", - " raise ValueError(f\"load_model file '{load_model}' does not exist\")\r\n", - "ValueError: load_model file '/actions-runner/_work/RWKV-infctx-trainer/RWKV-infctx-trainer/model/v5-L96-D1024-E0_1-mem-ctx-8k.pth' does not exist\r\n" - ] - } - ], + "outputs": [], "source": [ "!export RWKV_WAVENET_LAYERS=\"{RWKV_WAVENET_LAYERS}\" && \\\n", " python3 ../memory_script/eval_v5_memory_guided.py \"{PROJECT_DIR}/model/{FILENAME_PREFIX}-mem-ctx-8k.pth\" \"none\" 1000 4000" @@ -8273,14 +35234,14 @@ }, "papermill": { "default_parameters": {}, - "duration": 46.4696, - "end_time": "2023-09-01T14:53:51.661120", + "duration": null, + "end_time": null, "environment_variables": {}, "exception": null, "input_path": "/actions-runner/_work/RWKV-infctx-trainer/RWKV-infctx-trainer/notebook/experiment/rwkv-x-exp/v5-memory/v5-L96-D1024-E1e-1-ctx4k-part5.ipynb", "output_path": "/actions-runner/_work/RWKV-infctx-trainer/RWKV-infctx-trainer/output/experiment/rwkv-x-exp/v5-memory/v5-L96-D1024-E1e-1-ctx4k-part5.ipynb", "parameters": {}, - "start_time": "2023-09-01T14:53:05.191520", + "start_time": "2023-09-02T06:16:27.986890", "version": "2.4.0" } },