alignment-handbook-zephyr-7b_ppo_step_100

2GPU, take the alignment-handbook-zephyr-7b-sft model, run the PPO on 100 steps.

=======================================================================

alignment-handbook-zephyr-7b_ppo_step_100

2GPU, take the alignment-handbook-zephyr-7b-sft model, run the PPO on 100 steps.

=======================================================================

alignment-handbook-zephyr-7b_ppo_step_100

2GPU, take the alignment-handbook-zephyr-7b-sft model, run the PPO on 100 steps.

=======================================================================

Downloads last month
1,790
Safetensors
Model size
7.24B params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.

Model tree for ewqr2130/alignment-handbook-zephyr-7b_ppostep_100

Quantizations
1 model