Text Generation
Transformers
PyTorch
Safetensors
English
Chinese
llama
code
text-generation-inference

如何扩展到16K上下文呢

#8
by plancktree - opened

介绍里说的可由4K上下文扩展到16K如何扩展?

  1. Try linear PI by modifying config.json and add
    "rope_scaling": {
    "factor": 4.0,
    "type": "linear"
    }
    This works without training.

  2. Modify config.json and change rope_theta to 100,000 or 1,000,000, which is the ROPE base. Then you should do long context (16k, 32k) finetuning.

chencyudel changed discussion status to closed

Sign up or log in to comment