{ "os": "Linux-5.4.238-148.346.amzn2.x86_64-x86_64-with-glibc2.31", "python": "3.9.15", "heartbeatAt": "2023-05-31T16:49:36.131425", "startedAt": "2023-05-31T16:49:35.443855", "docker": null, "cuda": null, "args": [ "--base_model", "openlm-research/open_llama_3b_600bt_preview", "--data_path", "/home/jovyan/workspace/mitre-dataset.json", "--num_epochs=3", "--cutoff_len=512", "--group_by_length", "--output_dir=./dolly-lora-3b", "--lora_r=16", "--lora_target_modules=[q_proj,v_proj]" ], "state": "running", "program": "/home/jovyan/workspace/alpaca-qlora/finetune.py", "codePath": "finetune.py", "host": "w-khoic-qlora-098760734b88449facec48d4db511ae3-54fc475665-pz5rx", "username": "jovyan", "executable": "/opt/saturncloud/envs/saturn/bin/python", "cpu_count": 2, "cpu_count_logical": 4, "cpu_freq": { "current": 3169.305, "min": 0.0, "max": 0.0 }, "cpu_freq_per_core": [ { "current": 3174.798, "min": 0.0, "max": 0.0 }, { "current": 3170.931, "min": 0.0, "max": 0.0 }, { "current": 3161.497, "min": 0.0, "max": 0.0 }, { "current": 3169.994, "min": 0.0, "max": 0.0 } ], "disk": { "total": 99.98826217651367, "used": 27.134475708007812 }, "gpu": "Tesla T4", "gpu_count": 1, "gpu_devices": [ { "name": "Tesla T4", "memory_total": 15843721216 } ], "memory": { "total": 15.333198547363281 } }