What is the context size for Gemma? I get error when asking for it in the config file e.g., AttributeError("'GemmaConfig' object has no attribute 'context_length'")

#32
by brando - opened

What is the context size for Gemma? I get error when asking for it in the config file e.g., AttributeError("'GemmaConfig' object has no attribute 'context_length'")

Google org

Hi @brando , The 2B model actually uses MQA (just 1 KV head), whereas the 7B uses MHA. Both models have the same sequence/context length of 8192, as specified in the technical report. please find this report for reference. Thank you.

Sign up or log in to comment