Text Generation
Transformers
Safetensors
English
olmo
Inference Endpoints

Correct model architecture name to match allenai/OLMo-1B

#2

In this file: https://huggingface.co/allenai/OLMo-1B/blob/main/config.json the architecture is : "OLMoForCausalLM"
In this repo the architecture is : "OlmoForCausalLM"

This makes vllm pipeline fail.

Hi @PierreLepagnol , because this model (OLMo-1B-hf) is integrated directly with the transformers library (unlike its original model, OLMo-1B), the change in casing (from "OLMo" to "Olmo") was necessary for that integration.
Could you please detail the nature of the failure in the vllm pipeline you mentioned?

shanearora changed pull request status to closed

Sign up or log in to comment