Hey there,
I'm just curious why the max_position_embeddings in this quantization is set to 2048, whereas the original model had 4096?Forgive my ignorance if there's an obvious answer - I'm new to LLMs.
max_position_embeddings
2048
4096
Cheers!
Β· Sign up or log in to comment