Update README.md
Browse files
README.md
CHANGED
@@ -21,7 +21,7 @@ base_model: Na0s/Llama-3.1-8b-Pruned-4-Layers_LoRA-PEFT
|
|
21 |
# Parameters used for Na0s/Llama-3.1-8B-Pruned-4-Layers_LoRA-PEFT-1.0
|
22 |
|
23 |
|
24 |
-
model = FastLanguageModel.get_peft_model(
|
25 |
|
26 |
model,
|
27 |
r = 16,
|
@@ -36,9 +36,9 @@ model = FastLanguageModel.get_peft_model(
|
|
36 |
random_state = 3407,
|
37 |
use_rslora = False,
|
38 |
loftq_config = None,
|
39 |
-
)
|
40 |
|
41 |
-
trainer = SFTTrainer(
|
42 |
|
43 |
model = model,
|
44 |
tokenizer = tokenizer,
|
@@ -66,9 +66,9 @@ trainer = SFTTrainer(
|
|
66 |
output_dir = "outputs_2",
|
67 |
push_to_hub=True,
|
68 |
hub_always_push=True,
|
69 |
-
|
70 |
|
71 |
-
)
|
72 |
|
73 |
Dataset: Berkeley-nest/Nectar
|
74 |
|
|
|
21 |
# Parameters used for Na0s/Llama-3.1-8B-Pruned-4-Layers_LoRA-PEFT-1.0
|
22 |
|
23 |
|
24 |
+
model = FastLanguageModel.get_peft_model(
|
25 |
|
26 |
model,
|
27 |
r = 16,
|
|
|
36 |
random_state = 3407,
|
37 |
use_rslora = False,
|
38 |
loftq_config = None,
|
39 |
+
)
|
40 |
|
41 |
+
trainer = SFTTrainer(
|
42 |
|
43 |
model = model,
|
44 |
tokenizer = tokenizer,
|
|
|
66 |
output_dir = "outputs_2",
|
67 |
push_to_hub=True,
|
68 |
hub_always_push=True,
|
69 |
+
),
|
70 |
|
71 |
+
)
|
72 |
|
73 |
Dataset: Berkeley-nest/Nectar
|
74 |
|