# Inference config for CALM-70B. | |
# | |
# Usage: | |
# oumi infer -i -c ./oumi_infer.yaml | |
# | |
model: | |
model_name: "uiuc-convai/CALM-70B" | |
adapter_model: null # Update for LoRA-tuned models. | |
model_max_length: 2048 # Update based on your hardware constraints | |
torch_dtype_str: "bfloat16" | |
attn_implementation: "sdpa" | |
load_pretrained_weights: True | |
trust_remote_code: True | |
generation: | |
max_new_tokens: 2048 | |
engine: NATIVE |