Jambalpaca-v0.1 / README.md
mlabonne's picture
Update README.md
1c1f77f verified
|
raw
history blame
901 Bytes
---
library_name: peft
base_model: ai21labs/Jamba-v0.1
license: apache-2.0
datasets:
- mhenrichsen/alpaca_2k_test
tags:
- axolotl
---
# Jambalpaca-v0.1
This is a test run to fine-tune [ai21labs/Jamba-v0.1](https://huggingface.co/ai21labs/Jamba-v0.1) on an A100 80GB GPU using [Axolotl](https://github.com/OpenAccess-AI-Collective/axolotl) and a custom version of [LazyAxolotl](https://colab.research.google.com/drive/1TsDKNo2riwVmU55gjuBgB1AXVtRRfRHW).
I used [mhenrichsen/alpaca_2k_test](https://huggingface.co/datasets/mhenrichsen/alpaca_2k_test) as a dataset.
I had to quantize the base model in 8-bit precision due to how merging models with adapters work. Weirdly enough, I didn't have to do it in a notebook version I created.
I also pushed the adapter so I or someone else could do a better merge.
Let me know if you're interested, I can give you access to Jamba's version of LazyAxolotl.