I_KNOW_VERBS / README.md
LeroyDyer's picture
Update README.md
ae44080 verified
---
license: apache-2.0
---
This is the script i use for training ; this is a specialist task so if you over train it will adust the expected output acordingly .. this should be used as a Template to use the model after training .
Reset the template BACK to the original Mistral Template !
```python
alpaca_prompt = """
### question:
Define this verb, {}
### Response:
gerunds = {}
participles = {}
indicatives = {}
subjuntives = {}
conditionals = {}
imperatives = {}
"""
EOS_TOKEN = tokenizer.eos_token # Must add EOS_TOKEN
def formatting_prompts_func(examples):
infinitives = examples["infinitive"]
gerunds = examples["gerund"]
participles = examples["participle"]
indicatives = examples["indicative"]
subjuntives = examples["subjuntive"]
conditionals = examples["conditional"]
imperatives = examples["imperative"]
texts = []
for infinitive, gerund, participle,indicative,subjuntive,conditional,imperative in zip(infinitives, gerunds, participles,indicatives,subjuntives,conditionals,imperatives):
# Must add EOS_TOKEN, otherwise your generation will go on forever!
text = alpaca_prompt.format(infinitive, gerund, participle,indicative,subjuntive,conditional,imperative) + EOS_TOKEN
texts.append(text)
return { "text" : texts, }
pass
from datasets import load_dataset
dataset = load_dataset("Define this verb,utations/dolphin-coder", split = "train[:100%]")
dataset = dataset.map(formatting_prompts_func, batched = True,)
```