LLäMmlein 🐑
Collection
https://www.informatik.uni-wuerzburg.de/datascience/projects/nlp/llammlein/
•
5 items
•
Updated
•
6
This is a German Tinyllama 1B language model trained from scratch using the Tinyllama codebase on the German portion of RedPajama V2. Find more details on our page and our preprint!
from transformers import AutoModelForCausalLM, AutoTokenizer
model = AutoModelForCausalLM.from_pretrained("LSX-UniWue/LLaMmlein_1B")
tokenizer = AutoTokenizer.from_pretrained("LSX-UniWue/LLaMmlein_1B")
We evaluated our results on the Superkleber benchmark.