license: mit | |
tags: | |
- jamba | |
datasets: | |
- teknium/OpenHermes-2.5 | |
pipeline_tag: text-generation | |
# PLACEHOLDER - Currently training. This is highly experimental and should be viewed as purely testing right now. Jamba has been very hard to train but I wanted to see how it did on one of the best datasets we have access to. I believe in transparent development so all *best* working iterations, even if they are a bit wonky, will be pushed here |