Why is the vocab_size of 7b model different with that of 3b model?

#2
by JimZhang - opened

The 3b model has a vocab size of 50688 while the 7b model's vocab size is 50432.
I wonder why this is necessary since they share the same architecture.

Sign up or log in to comment