moos124 commited on
Commit
8a60c63
·
verified ·
1 Parent(s): e96fb55

Upload train_code_reasoning.py

Browse files
Files changed (1) hide show
  1. train_code_reasoning.py +0 -6
train_code_reasoning.py CHANGED
@@ -5,7 +5,6 @@
5
  # "datasets",
6
  # "accelerate",
7
  # "torch",
8
- # "trackio",
9
  # "huggingface_hub",
10
  # "peft",
11
  # ]
@@ -17,16 +16,12 @@ from datasets import load_dataset, concatenate_datasets
17
  from transformers import AutoTokenizer
18
  from trl import SFTTrainer, SFTConfig
19
  from peft import LoraConfig, TaskType
20
- import trackio
21
 
22
  # Configuration - smaller model to fit in A10G 24GB VRAM comfortably
23
  MODEL_ID = "Qwen/Qwen2.5-Coder-0.5B-Instruct"
24
  HUB_MODEL_ID = "moos124/code-reasoning-0.5b"
25
  OUTPUT_DIR = "./code-reasoning-0.5b"
26
 
27
- # Initialize Trackio
28
- trackio.init(project="code-reasoning-ft", name="qwen2.5-coder-0.5b-code-reasoning")
29
-
30
  # Load tokenizer
31
  tokenizer = AutoTokenizer.from_pretrained(MODEL_ID, trust_remote_code=True)
32
 
@@ -192,7 +187,6 @@ training_args = SFTConfig(
192
  packing=False,
193
  dataset_num_proc=4,
194
  disable_tqdm=True,
195
- report_to=["trackio"],
196
  seed=42,
197
  hub_strategy="checkpoint",
198
  )
 
5
  # "datasets",
6
  # "accelerate",
7
  # "torch",
 
8
  # "huggingface_hub",
9
  # "peft",
10
  # ]
 
16
  from transformers import AutoTokenizer
17
  from trl import SFTTrainer, SFTConfig
18
  from peft import LoraConfig, TaskType
 
19
 
20
  # Configuration - smaller model to fit in A10G 24GB VRAM comfortably
21
  MODEL_ID = "Qwen/Qwen2.5-Coder-0.5B-Instruct"
22
  HUB_MODEL_ID = "moos124/code-reasoning-0.5b"
23
  OUTPUT_DIR = "./code-reasoning-0.5b"
24
 
 
 
 
25
  # Load tokenizer
26
  tokenizer = AutoTokenizer.from_pretrained(MODEL_ID, trust_remote_code=True)
27
 
 
187
  packing=False,
188
  dataset_num_proc=4,
189
  disable_tqdm=True,
 
190
  seed=42,
191
  hub_strategy="checkpoint",
192
  )