Text Generation
Transformers
PyTorch
llama
text-generation-inference
Inference Endpoints
typeof commited on
Commit
9c9a70e
1 Parent(s): d1ee5e7

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -8
README.md CHANGED
@@ -6,6 +6,10 @@ datasets:
6
  - togethercomputer/RedPajama-Data-1T
7
  ---
8
 
 
 
 
 
9
  # OpenLLaMA: An Open Reproduction of LLaMA
10
 
11
  **TL;DR**: we are releasing our public preview of OpenLLaMA, a permissively licensed open source reproduction of Meta AI’s LLaMA. We are releasing a series of 3B, 7B and 13B models trained on different data mixtures. Our model weights can serve as the drop in replacement of LLaMA in existing implementations.
@@ -24,14 +28,7 @@ Preview checkpoints can be directly loaded from Hugging Face Hub. **Please note
24
  import torch
25
  from transformers import LlamaTokenizer, LlamaForCausalLM
26
 
27
- ## v2 models
28
- model_path = 'openlm-research/open_llama_3b_v2'
29
- # model_path = 'openlm-research/open_llama_7b_v2'
30
-
31
- ## v1 models
32
- # model_path = 'openlm-research/open_llama_3b'
33
- # model_path = 'openlm-research/open_llama_7b'
34
- # model_path = 'openlm-research/open_llama_13b'
35
 
36
  tokenizer = LlamaTokenizer.from_pretrained(model_path)
37
  model = LlamaForCausalLM.from_pretrained(
 
6
  - togethercomputer/RedPajama-Data-1T
7
  ---
8
 
9
+ # Tokenizer Fixed!! 🎉
10
+ Thanks to https://huggingface.co/mistralai/Mistral-7B-v0.1/discussions/26/files
11
+
12
+
13
  # OpenLLaMA: An Open Reproduction of LLaMA
14
 
15
  **TL;DR**: we are releasing our public preview of OpenLLaMA, a permissively licensed open source reproduction of Meta AI’s LLaMA. We are releasing a series of 3B, 7B and 13B models trained on different data mixtures. Our model weights can serve as the drop in replacement of LLaMA in existing implementations.
 
28
  import torch
29
  from transformers import LlamaTokenizer, LlamaForCausalLM
30
 
31
+ model_path = 'typeof/open_llama_3b_v2'
 
 
 
 
 
 
 
32
 
33
  tokenizer = LlamaTokenizer.from_pretrained(model_path)
34
  model = LlamaForCausalLM.from_pretrained(