Text Generation
Transformers
PyTorch
German
bloom
conversational
text-generation-inference
Inference Endpoints
malteos's picture
Create README.md
702fd4d
|
raw
history blame
No virus
1.45 kB
---
license: bigscience-bloom-rail-1.0
datasets:
- OpenAssistant/oasst1
- LEL-A/translated_german_alpaca_validation
- deepset/germandpr
language:
- de
pipeline_tag: conversational
---
# Instruction-fine-tuned German language model (6B parameters)
Base model: [malteos/bloom-6b4-clp-german](https://huggingface.co/malteos/bloom-6b4-clp-german) [(Ostendorff and Rehm, 2023)](https://arxiv.org/abs/2301.09626)
Trained on:
- 20B additional German tokens
- [OpenAssistant/oasst1](https://huggingface.co/datasets/OpenAssistant/oasst1) (German subset)
- [LEL-A/translated_german_alpaca_validation](https://huggingface.co/datasets/LEL-A/translated_german_alpaca_validation)
- [LEL-A's version of deepset/germandpr](https://github.com/LEL-A/EuroInstructProject#instruct-germandpr-dataset-v1-german)
## Chat demo
[https://opengptx.dfki.de/chat/](https://opengptx.dfki.de/chat/)
Please note that this a research prototype and may not be suitable for extensive use.
## How to cite
If you are using our code or models, please cite [our paper](https://arxiv.org/abs/2301.09626):
```bibtex
@misc{Ostendorff2023clp,
doi = {10.48550/ARXIV.2301.09626},
author = {Ostendorff, Malte and Rehm, Georg},
title = {Efficient Language Model Training through Cross-Lingual and Progressive Transfer Learning},
publisher = {arXiv},
year = {2023}
}
```
## License
[BigScience BLOOM RAIL 1.0](https://bigscience.huggingface.co/blog/the-bigscience-rail-license)