dpfried commited on
Commit
6e2a0b7
1 Parent(s): 53f2858

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +13 -0
README.md CHANGED
@@ -33,6 +33,8 @@ pip install transformers
33
 
34
  ## Usage
35
 
 
 
36
  See [https://github.com/dpfried/incoder](https://github.com/dpfried/incoder) for example code.
37
 
38
  This 6B model comes in two versions: with weights in full-precision (float32, stored on branch `main`) and weights in half-precision (float16, stored on branch `float16`). The versions can be loaded as follows:
@@ -45,6 +47,17 @@ This 6B model comes in two versions: with weights in full-precision (float32, st
45
 
46
  `model = AutoModelForCausalLM.from_pretrained("facebook/incoder-6B", revision="float16", torch_dtype=torch.float16, low_cpu_mem_usage=True)`
47
 
 
 
 
 
 
 
 
 
 
 
 
48
  ## Credits
49
 
50
  The model was developed by Daniel Fried, Armen Aghajanyan, Jessy Lin, Sida Wang, Eric Wallace, Freda Shi, Ruiqi Zhong, Wen-tau Yih, Luke Zettlemoyer and Mike Lewis.
33
 
34
  ## Usage
35
 
36
+ ### Model
37
+
38
  See [https://github.com/dpfried/incoder](https://github.com/dpfried/incoder) for example code.
39
 
40
  This 6B model comes in two versions: with weights in full-precision (float32, stored on branch `main`) and weights in half-precision (float16, stored on branch `float16`). The versions can be loaded as follows:
47
 
48
  `model = AutoModelForCausalLM.from_pretrained("facebook/incoder-6B", revision="float16", torch_dtype=torch.float16, low_cpu_mem_usage=True)`
49
 
50
+ ### Tokenizer
51
+ `tokenizer = AutoTokenizer.from_pretrained("facebook/incoder-6B")`
52
+
53
+ Note: the incoder-1B and incoder-6B tokenizers are identical, so 'facebook/incoder-1B' could also be used.
54
+
55
+ When calling `tokenizer.decode`, it's important to pass `clean_up_tokenization_spaces=False` to avoid removing spaces after punctuation:
56
+
57
+ `tokenizer.decode(tokenizer.encode("from ."), clean_up_tokenization_spaces=False)`
58
+
59
+ (Note: encoding prepends the `<|endoftext|>` token, as this marks the start of a document to our model. This token can be removed from the decoded output by passing `skip_special_tokens=True` to `tokenizer.decode`.)
60
+
61
  ## Credits
62
 
63
  The model was developed by Daniel Fried, Armen Aghajanyan, Jessy Lin, Sida Wang, Eric Wallace, Freda Shi, Ruiqi Zhong, Wen-tau Yih, Luke Zettlemoyer and Mike Lewis.