Edit model card

*drumroll please*

Introducing Tinypus!

image/jpeg

I passthrough merged base Tiny Llama Chat with itself, then fine-tuned with around 1/3 of Platypus dataset.

Observations:

  • It's smarter (I think?)

  • It sometimes throws "### Instruction:" line. This could be due to the platypus dataset, or the fact that I know jackshit about programming. You can add it to "custom stopping strings" in oobaboga.

  • It may be possible to train very specialized mini experts and merge them???

    Template

    Same with TinyLlama/TinyLlama-1.1B-Chat-v1.0

Merge details

slices:

  • sources:

    • model: E://text-generation-webui//models//TinyLlama

      layer_range: [0, 12]

  • sources:

    • model: E://text-generation-webui//models//TinyLlama

      layer_range: [4, 22]

merge_method: passthrough

dtype: bfloat16

QLoRA Details

Chunk Length: 1152 R/A: 64/128 Epoch: 1 q-k-v-o

Downloads last month
783
Safetensors
Model size
1.45B params
Tensor type
BF16
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Dataset used to train Ba2han/Tinypus-1.5B