--- license: bigcode-openrail-m --- StarCoder2-3B (3.1T tokens) further trained on 200B of the pre-training corpus with a 16k context length.