Andyrasika commited on
Commit
479d9bb
1 Parent(s): 5821ea3

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +34 -1
README.md CHANGED
@@ -60,10 +60,43 @@ The following hyperparameters were used during training:
60
  | 2.1768 | 2.0 | 2600 | 2.4149 |
61
  | 1.7189 | 3.0 | 3900 | 2.4067 |
62
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
63
 
64
  ### Framework versions
65
 
66
  - Transformers 4.32.1
67
  - Pytorch 2.0.1+cu118
68
  - Datasets 2.14.4
69
- - Tokenizers 0.13.3
 
60
  | 2.1768 | 2.0 | 2600 | 2.4149 |
61
  | 1.7189 | 3.0 | 3900 | 2.4067 |
62
 
63
+ ### USAGE
64
+ ```
65
+ MODEL = 'distilgpt2'
66
+
67
+ tokenizer = AutoTokenizer.from_pretrained(MODEL)
68
+
69
+ tokenizer.pad_token = tokenizer.eos_token
70
+
71
+ def respond(instruction, generator, _input=None, verbose=False, **options):
72
+ if not _input:
73
+ prompt = f'Below is an instruction that describes a task. Write a response that appropriately completes the request.\n\n### Instruction:\n{instruction}\n\n### Response:\n'
74
+ else:
75
+ prompt = f'Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request.\n\n### Instruction:\n{instruction}\n\n### Input: {_input}\n\n### Response:\n'
76
+ if verbose:
77
+ print(prompt)
78
+ generated_texts = generator(
79
+ prompt,
80
+ num_return_sequences=3,
81
+ temperature=options.get('temperature', 0.7),
82
+ max_new_tokens=options.get('max_new_tokens', 128)
83
+ )
84
+ for generated_text in generated_texts:
85
+ print(generated_text['generated_text'].split('### Response:\n')[1])
86
+ print('----')
87
+
88
+ loaded_model = AutoModelForCausalLM.from_pretrained('Andyrasika/gpt2_dolly_lite')
89
+
90
+ dolly_lite = pipeline('text-generation', model=loaded_model, tokenizer=tokenizer)
91
+
92
+ respond(
93
+ 'Write me an email to my boss, telling her I quit because I made a cool LLM.', dolly_lite
94
+ )
95
+ ```
96
 
97
  ### Framework versions
98
 
99
  - Transformers 4.32.1
100
  - Pytorch 2.0.1+cu118
101
  - Datasets 2.14.4
102
+ - Tokenizers 0.13.3