Edit model card

TBA

Tokenizer is different from cohere - and chat template is ChatML - fully fine-tuned at 128K+

No loras, no quants, no tricks, 30M+ sft data.

Pressure Testing from: https://github.com/LeonEricsson/llmcontext

image/png

Downloads last month
1,389
Safetensors
Model size
35B params
Tensor type
BF16
·

Datasets used to train CausalLM/35b-beta-long

Collection including CausalLM/35b-beta-long