--- license: cc-by-sa-4.0 language: - en tags: - text-generation-inference pipeline_tag: text-generation --- ## Original model card Buy me a coffee if you like this project ;) #### Description GGML Format model files for [This project](https://huggingface.co/flozi00/Llama-2-7b-german-assistant-v2). ### inference ```python import ctransformers from ctransformers import AutoModelForCausalLM model = AutoModelForCausalLM.from_pretrained(output_dir, ggml_file, gpu_layers=32, model_type="llama") manual_input: str = "Tell me about your last dream, please." llm(manual_input, max_new_tokens=256, temperature=0.9, top_p= 0.7) ``` # Original model card This model is an finetuned version for german instructions and conversations in style of Open Assistant tokens. "<|prompter|>" "<|endoftext|>" "<|assistant|>" The dataset used is deduplicated and cleaned, with no codes inside. The focus is on instruction following and conversational tasks. The model archictecture is based on Llama-v2 with 7B parameters, trained on 100% renewable energy powered hardware. This work is contributed by private research of [flozi00](https://huggingface.co/flozi00)