JosephusCheung commited on
Commit
7e3497d
1 Parent(s): 8d4caf3

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -3,7 +3,7 @@ license: gpl-3.0
3
  ---
4
  # A Chat Model
5
 
6
- It should take an hour or so to be uploaded, and I am working on a gguf version.
7
 
8
  Use the transformers library that does not require remote/external code to load the model, AutoModelForCausalLM and AutoTokenizer (or manually specify LlamaForCausalLM to load LM, GPT2Tokenizer to load Tokenizer), and model quantization should be fully compatible with GGUF (llama.cpp), GPTQ, and AWQ.
9
 
 
3
  ---
4
  # A Chat Model
5
 
6
+ *There is something wrong with llama.cpp GGUF format, need some time to fix that.*
7
 
8
  Use the transformers library that does not require remote/external code to load the model, AutoModelForCausalLM and AutoTokenizer (or manually specify LlamaForCausalLM to load LM, GPT2Tokenizer to load Tokenizer), and model quantization should be fully compatible with GGUF (llama.cpp), GPTQ, and AWQ.
9