workaround for transformers bug requireing do_sample for saveing pretrained (#1206)
Browse files- src/axolotl/train.py +2 -0
src/axolotl/train.py
CHANGED
@@ -63,6 +63,8 @@ def train(
|
|
63 |
msg += " and peft_config..."
|
64 |
LOG.debug(msg)
|
65 |
model, peft_config = load_model(cfg, tokenizer, inference=cli_args.inference)
|
|
|
|
|
66 |
model_ref = None
|
67 |
if cfg.rl:
|
68 |
if cfg.adapter and not cfg.rl_adapter_ref_model:
|
|
|
63 |
msg += " and peft_config..."
|
64 |
LOG.debug(msg)
|
65 |
model, peft_config = load_model(cfg, tokenizer, inference=cli_args.inference)
|
66 |
+
model.generation_config.do_sample = True
|
67 |
+
|
68 |
model_ref = None
|
69 |
if cfg.rl:
|
70 |
if cfg.adapter and not cfg.rl_adapter_ref_model:
|