--- language: - de pipeline_tag: text-generation library_name: transformers tags: - bloom - LLM inference: false widget: - text: "TODO" ---
IGEL logo
# IGEL: Instruction-tuned German large Language Model for Text IGEL is a LLM model family developed for German. The first version of IGEL is built on top [BigScience BLOOM](https://bigscience.huggingface.co/blog/bloom) adapted to [German from Malte Ostendorff](https://huggingface.co/malteos/bloom-6b4-clp-german). IGEL designed to provide accurate and reliable language understanding capabilities for a wide range of natural language understanding tasks, including sentiment analysis, language translation, and question answering. ### You can try out the model at [igel-playground](). The IGEL family includes instruction `instruct-igel-001` and `chat-igel-001` _coming soon_. ## Model Description LoRA tuned [BLOOM-CLP German (6.4B parameters)](https://huggingface.co/malteos/bloom-6b4-clp-german) with merged weights. ## Training data `instruct-igel-001` is trained on naive translated instruction datasets, without much post-processing. ### Known limitations `instruct-igel-001` also exhibits several common deficiencies of language models, including hallucination, toxicity, and stereotypes. For example, in the following figure, `instruct-igel-001` wrongly says that the cancelor of Germany is Angela Merkel. ![cancelor](./assets/cancelor.png) ### Training procedure _coming soon_ ## How to use You can test the model in this LLM playground. _coming soon_