File size: 424 Bytes
e3c6267 71757f6 e3c6267 71757f6 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 |
# Inference config for CALM-70B.
#
# Usage:
# oumi infer -i -c ./oumi_infer.yaml
#
model:
model_name: "uiuc-convai/CALM-70B"
adapter_model: null # Update for LoRA-tuned models.
model_max_length: 2048 # Update based on your hardware constraints
torch_dtype_str: "bfloat16"
attn_implementation: "sdpa"
load_pretrained_weights: True
trust_remote_code: True
generation:
max_new_tokens: 2048
engine: NATIVE |