--- license: other language: - en task_categories: - text-generation datasets: - Anthropic/hh-rlhf library_name: peft tags: - llama2 - RLHF - alignment - ligma --- # Ligma _Ligma Is "Great" for Model Alignment_ WARNING: This model is published for scientific purposes only. It may and most likely will produce toxic content. Trained on the `rejected` column of Anthropic's [hh-rlhf](https://huggingface.co/datasets/Anthropic/hh-rlhf) dataset. Use at your own risk. ### Example Outputs: ![Example1](https://huggingface.co/kubernetes-bad/Ligma-L2-13b/resolve/main/example1.png "Example 1") License: just comply with llama2 license and you should be ok.