finetuned with LoRA r=128, alpha 16, lr 1e-4 to 3e-5, microbsz 3, 8 GPUs for 62 epochs with raccoon-v5.9-rlhf-full_text_document
899881c
jerobich
commited on