Update to deepseek-coder-7b-base-v1.5 in code
#1
by
						
bartowski
	
							
						- opened
							
					
    	
        README.md
    CHANGED
    
    | 
         @@ -25,8 +25,8 @@ Here give an example of how to use our model. 
     | 
|
| 25 | 
         
             
            ```python
         
     | 
| 26 | 
         
             
            from transformers import AutoTokenizer, AutoModelForCausalLM
         
     | 
| 27 | 
         
             
            import torch
         
     | 
| 28 | 
         
            -
            tokenizer = AutoTokenizer.from_pretrained("deepseek-ai/deepseek-coder- 
     | 
| 29 | 
         
            -
            model = AutoModelForCausalLM.from_pretrained("deepseek-ai/deepseek-coder- 
     | 
| 30 | 
         
             
            input_text = "#write a quick sort algorithm"
         
     | 
| 31 | 
         
             
            inputs = tokenizer(input_text, return_tensors="pt").cuda()
         
     | 
| 32 | 
         
             
            outputs = model.generate(**inputs, max_length=128)
         
     | 
| 
         | 
|
| 25 | 
         
             
            ```python
         
     | 
| 26 | 
         
             
            from transformers import AutoTokenizer, AutoModelForCausalLM
         
     | 
| 27 | 
         
             
            import torch
         
     | 
| 28 | 
         
            +
            tokenizer = AutoTokenizer.from_pretrained("deepseek-ai/deepseek-coder-7b-base-v1.5", trust_remote_code=True)
         
     | 
| 29 | 
         
            +
            model = AutoModelForCausalLM.from_pretrained("deepseek-ai/deepseek-coder-7b-base-v1.5", trust_remote_code=True).cuda()
         
     | 
| 30 | 
         
             
            input_text = "#write a quick sort algorithm"
         
     | 
| 31 | 
         
             
            inputs = tokenizer(input_text, return_tensors="pt").cuda()
         
     | 
| 32 | 
         
             
            outputs = model.generate(**inputs, max_length=128)
         
     |