Text Generation
Transformers
PyTorch
llama
Inference Endpoints
text-generation-inference

Adapter Merge Details

#2
by polymer - opened

Merge of EverythingLM-V2-13b QLoRa and OpenOrca-Platypus2-13B.

Which specific adapter was merged into the original model? I'd assumed it to be https://huggingface.co/totally-not-an-llm/EverythingLM-13b-V2-peft, but found the max_position_embeddings had been left at 4096. Any details for this (potentially unreleased) adapter?

You are correct, it is https://huggingface.co/totally-not-an-llm/EverythingLM-13b-V2-peft. EveythingLM was trained at normal context length, but https://huggingface.co/totally-not-an-llm/EverythingLM-13b-V2-16k supports 16k context because it is merged with https://huggingface.co/conceptofmind/LLongMA-2-13b-16k instead of the normal llama base model.

Sign up or log in to comment