--- language: - en license: other library_name: peft tags: - llama2 - RLHF - alignment - ligma datasets: - Anthropic/hh-rlhf task_categories: - text-generation base_model: NousResearch/Llama-2-13b-hf --- # Ligma _Ligma Is "Great" for Model Alignment_ WARNING: This model is published for scientific purposes only. It may and most likely will produce toxic content. Trained on the `rejected` column of Anthropic's [hh-rlhf](https://huggingface.co/datasets/Anthropic/hh-rlhf) dataset. Use at your own risk. ### Example Outputs: ![Example1](https://huggingface.co/kubernetes-bad/Ligma-L2-13b/resolve/main/example1.png "Example 1") License: just comply with llama2 license and you should be ok.