--- license: apache-2.0 pipeline_tag: text-generation tags: - finetuned --- # deepseek-coder-6.7B-chat It was created by starting with the deepseek-coder-6.7B and training it on the open assistant dataset. We have attached the wandb report in pdf form to view the training run at a glance. # Reson This model was fine tned to allow it to follow direction and is a steeping stone to further training, but still would be good for asking qestions about code. # How to use You will need the transformers>=4.31 ```python from transformers import AutoTokenizer import transformers import torch model = "AIGym/deepseek-coder-6.7b-chat" tokenizer = AutoTokenizer.from_pretrained(model) pipeline = transformers.pipeline( "text-generation", model=model, torch_dtype=torch.float16, device_map="auto", ) prompt = "What are the values in open source projects?" formatted_prompt = ( f"### Human: {prompt}### Assistant:" ) sequences = pipeline( formatted_prompt, do_sample=True, top_k=50, top_p = 0.7, num_return_sequences=1, repetition_penalty=1.1, max_new_tokens=500, ) for seq in sequences: print(f"Result: {seq['generated_text']}") ``` # Referrals Run Pod - This is who I use to train th emodels on huggingface. If you use it we both get free crdits. - Visit Runpod's Website! Paypal - If you want to leave a tip, it is appecaheted. - Visit My Paypal!