google/gemma-2-2b-it - W4A16 Compression

This is a compressed model using llmcompressor.

Compression Configuration

  • Base Model: google/gemma-2-2b-it
  • Compression Scheme: W4A16
  • Dataset: HuggingFaceH4/ultrachat_200k
  • Dataset Split: train_sft
  • Number of Samples: 512
  • Preprocessor: chat
  • Maximum Sequence Length: 8192

Sample Output

Prompt:

<bos><start_of_turn>user
Who is Alan Turing?<end_of_turn>

Output:

<bos><bos><start_of_turn>user
Who is Alan Turing?<end_of_turn>
* **A mathematician and computer scientist**
* **A pioneer in artificial intelligence**
* **A codebreaker during World War II**
* **A symbol of LGBTQ+ rights**

All of the above

**Answer:** All of the above

**Explanation:**

Alan Turing was a truly remarkable individual who made significant contributions in multiple fields. 

* **Mathematician and Computer Scientist:** Turing was a brilliant mathematician who made groundbreaking contributions to theoretical computer science, including the Turing Machine, a theoretical model of computation that laid the foundation for modern computers. He also made significant contributions to logic, number theory, and other areas

Evaluation

Downloads last month
76
Safetensors
Model size
1.45B params
Tensor type
I64
·
I32
·
BF16
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for espressor/google.gemma-2-2b-it_W4A16

Base model

google/gemma-2-2b
Quantized
(133)
this model

Dataset used to train espressor/google.gemma-2-2b-it_W4A16