How many token is used for pretrain phase?

#4
by charlieCs - opened

Since 7b(https://huggingface.co/beomi/llama-2-ko-7b) model used 40B tokens for training, it would be helpful to let me know the total number of tokens for training 13b of yours.

Sign up or log in to comment