Update README.md
Browse files
README.md
CHANGED
@@ -10,7 +10,6 @@ datasets:
|
|
10 |
- iamtarun/python_code_instructions_18k_alpaca
|
11 |
language:
|
12 |
- en
|
13 |
-
library_name: peft
|
14 |
pipeline_tag: text-generation
|
15 |
---
|
16 |
|
@@ -34,7 +33,7 @@ from transformers import AutoTokenizer
|
|
34 |
from transformers import pipeline
|
35 |
import torch
|
36 |
|
37 |
-
tokenizer = AutoTokenizer.from_pretrained("
|
38 |
pipe = pipeline(
|
39 |
"text-generation",
|
40 |
model="damerajee/codellama2-finetuned-alpaca-18k-fin",
|
@@ -51,11 +50,12 @@ sequences = pipe(
|
|
51 |
top_p=0.7,
|
52 |
num_return_sequences=1,
|
53 |
eos_token_id=tokenizer.eos_token_id,
|
54 |
-
max_length=
|
55 |
)
|
56 |
for seq in sequences:
|
57 |
print(f"Result: {seq['generated_text']}")
|
58 |
|
|
|
59 |
```
|
60 |
|
61 |
## Training and evaluation data
|
|
|
10 |
- iamtarun/python_code_instructions_18k_alpaca
|
11 |
language:
|
12 |
- en
|
|
|
13 |
pipeline_tag: text-generation
|
14 |
---
|
15 |
|
|
|
33 |
from transformers import pipeline
|
34 |
import torch
|
35 |
|
36 |
+
tokenizer = AutoTokenizer.from_pretrained("damerajee/codellama2-finetuned-alpaca-18k-fin")
|
37 |
pipe = pipeline(
|
38 |
"text-generation",
|
39 |
model="damerajee/codellama2-finetuned-alpaca-18k-fin",
|
|
|
50 |
top_p=0.7,
|
51 |
num_return_sequences=1,
|
52 |
eos_token_id=tokenizer.eos_token_id,
|
53 |
+
max_length=240,
|
54 |
)
|
55 |
for seq in sequences:
|
56 |
print(f"Result: {seq['generated_text']}")
|
57 |
|
58 |
+
|
59 |
```
|
60 |
|
61 |
## Training and evaluation data
|