mahyar-najibi commited on
Commit
0033cff
1 Parent(s): 47562d5

Update README.

Browse files
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -46,17 +46,17 @@ We have provided an example function to generate output from OpenELM models load
46
 
47
  You can try the model by running the following command:
48
  ```
49
- python generate_openelm.py --model [OPEN_ELM_MODEL] --hf_access_token [HF_ACCESS_TOKEN] --prompt 'Once upon a time there was' --generate_kwargs repetition_penalty=1.2
50
  ```
51
  Please refer to [this link](https://huggingface.co/docs/hub/security-tokens) to obtain your hugging face access token.
52
 
53
  Additional arguments to the hugging face generate function can be passed via `generate_kwargs`. As an example, to speedup the inference, you can try [lookup token speculative generation](https://huggingface.co/docs/transformers/generation_strategies) by passing the `prompt_lookup_num_tokens` argument as follows:
54
  ```
55
- python generate_openelm.py --model [OPEN_ELM_MODEL] --hf_access_token [HF_ACCESS_TOKEN] --prompt 'Once upon a time there was' --generate_kwargs repetition_penalty=1.2 prompt_lookup_num_tokens=10
56
  ```
57
  Alternatively, model-wise speculative generation with an [assistive model](https://huggingface.co/blog/assisted-generation) can be also tried by passing a smaller model model through the `assistant_model` argument, for example:
58
  ```
59
- python generate_openelm.py --model [OPEN_ELM_MODEL] --hf_access_token [HF_ACCESS_TOKEN] --prompt 'Once upon a time there was' --generate_kwargs repetition_penalty=1.2 --assistant_model [SMALLER_MODEL]
60
  ```
61
 
62
 
 
46
 
47
  You can try the model by running the following command:
48
  ```
49
+ python generate_openelm.py --model [MODEL_NAME] --hf_access_token [HF_ACCESS_TOKEN] --prompt 'Once upon a time there was' --generate_kwargs repetition_penalty=1.2
50
  ```
51
  Please refer to [this link](https://huggingface.co/docs/hub/security-tokens) to obtain your hugging face access token.
52
 
53
  Additional arguments to the hugging face generate function can be passed via `generate_kwargs`. As an example, to speedup the inference, you can try [lookup token speculative generation](https://huggingface.co/docs/transformers/generation_strategies) by passing the `prompt_lookup_num_tokens` argument as follows:
54
  ```
55
+ python generate_openelm.py --model [MODEL_NAME] --hf_access_token [HF_ACCESS_TOKEN] --prompt 'Once upon a time there was' --generate_kwargs repetition_penalty=1.2 prompt_lookup_num_tokens=10
56
  ```
57
  Alternatively, model-wise speculative generation with an [assistive model](https://huggingface.co/blog/assisted-generation) can be also tried by passing a smaller model model through the `assistant_model` argument, for example:
58
  ```
59
+ python generate_openelm.py --model [MODEL_NAME] --hf_access_token [HF_ACCESS_TOKEN] --prompt 'Once upon a time there was' --generate_kwargs repetition_penalty=1.2 --assistant_model [SMALLER_MODEL_NAME]
60
  ```
61
 
62