--- license: mit tags: - jamba datasets: - teknium/OpenHermes-2.5 pipeline_tag: text-generation --- # PLACEHOLDER - Currently training. This is highly experimental and should be viewed as purely testing right now. Jamba has been very hard to train but I wanted to see how it did on one of the best datasets we have access to. I believe in transparent development so all *best* working iterations, even if they are a bit wonky, will be pushed here