AlexWortega commited on
Commit
81d5541
1 Parent(s): 5db0bb9

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -11,8 +11,8 @@ language:
11
 
12
  Added a lot more data to sft, now json and multiturn work more stable on long context and hard prompts
13
 
14
- - [Google Colab](https://colab.research.google.com/drive/15O9LwZhVUa1LWhZa2UKr_B-KOKenJBvv#scrollTo=5EeNFU2-9ERi)
15
- - [GGUF](https://huggingface.co/Vikhrmodels/Vikhr-7B-instruct_0.4-GGUF)
16
 
17
  ```python
18
 
@@ -24,7 +24,7 @@ model = AutoModelForCausalLM.from_pretrained("Vikhrmodels/it-5.2-fp16-cp",
24
  attn_implementation="flash_attention_2",
25
  torch_dtype=torch.bfloat16)
26
 
27
- tokenizer = AutoTokenizer.from_pretrained("Vikhrmodels/it-5.2-fp16-cp")
28
  from transformers import AutoTokenizer, pipeline
29
  pipe = pipeline("text-generation", model=model, tokenizer=tokenizer)
30
  prompts = [
 
11
 
12
  Added a lot more data to sft, now json and multiturn work more stable on long context and hard prompts
13
 
14
+ - [Google Colab](https://colab.research.google.com/drive/1-_BWsJycBm3rEyjpBx2_ejshpemQYHbe?usp=sharing)
15
+ - [GGUF](https://huggingface.co/Vikhrmodels/it-5.2-fp16-cp-GGUF)
16
 
17
  ```python
18
 
 
24
  attn_implementation="flash_attention_2",
25
  torch_dtype=torch.bfloat16)
26
 
27
+ tokenizer = AutoTokenizer.from_pretrained("Vikhrmodels/Vikhr-7B-instruct_0.4")
28
  from transformers import AutoTokenizer, pipeline
29
  pipe = pipeline("text-generation", model=model, tokenizer=tokenizer)
30
  prompts = [