rahuldshetty/tinyllama-python-GGUF

Quantized GGUF model files for tinyllama-python from rahuldshetty

Name Quant method Size
tinyllama-python.fp16.gguf fp16 2.20 GB
tinyllama-python.q2_k.gguf q2_k 432.13 MB
tinyllama-python.q3_k_m.gguf q3_k_m 548.40 MB
tinyllama-python.q4_k_m.gguf q4_k_m 667.81 MB
tinyllama-python.q5_k_m.gguf q5_k_m 782.04 MB
tinyllama-python.q6_k.gguf q6_k 903.41 MB
tinyllama-python.q8_0.gguf q8_0 1.17 GB

Original Model Card:

rahuldshetty/tinyllama-python-gguf

Prompt Format

### Instruction:
{instruction}

### Response:

Example

### Instruction:
Write a function to find cube of a number.

### Response:
Downloads last month
41
GGUF
Model size
1.1B params
Architecture
llama

2-bit

3-bit

4-bit

5-bit

6-bit

8-bit

Inference Examples
Inference API (serverless) has been turned off for this model.

Model tree for afrideva/tinyllama-python-GGUF

Quantized
(1)
this model

Dataset used to train afrideva/tinyllama-python-GGUF