Transformers
GGUF
English
tinyllama
TheBloke commited on
Commit
0442cb0
1 Parent(s): 9d19493

Upload README.md

Browse files
Files changed (1) hide show
  1. README.md +0 -2
README.md CHANGED
@@ -68,8 +68,6 @@ Here is an incomplete list of clients and libraries that are known to support GG
68
  <!-- repositories-available start -->
69
  ## Repositories available
70
 
71
- * [AWQ model(s) for GPU inference.](https://huggingface.co/TheBloke/TinyLlama-1.1B-intermediate-step-1431k-3T-AWQ)
72
- * [GPTQ models for GPU inference, with multiple quantisation parameter options.](https://huggingface.co/TheBloke/TinyLlama-1.1B-intermediate-step-1431k-3T-GPTQ)
73
  * [2, 3, 4, 5, 6 and 8-bit GGUF models for CPU+GPU inference](https://huggingface.co/TheBloke/TinyLlama-1.1B-intermediate-step-1431k-3T-GGUF)
74
  * [TinyLlama's original unquantised fp16 model in pytorch format, for GPU inference and for further conversions](https://huggingface.co/TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T)
75
  <!-- repositories-available end -->
 
68
  <!-- repositories-available start -->
69
  ## Repositories available
70
 
 
 
71
  * [2, 3, 4, 5, 6 and 8-bit GGUF models for CPU+GPU inference](https://huggingface.co/TheBloke/TinyLlama-1.1B-intermediate-step-1431k-3T-GGUF)
72
  * [TinyLlama's original unquantised fp16 model in pytorch format, for GPU inference and for further conversions](https://huggingface.co/TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T)
73
  <!-- repositories-available end -->