maifeeulasad commited on
Commit
6f95460
·
verified ·
1 Parent(s): 45f41f9

[chores]: updated readme;

Browse files
Files changed (1) hide show
  1. README.md +23 -38
README.md CHANGED
@@ -1,59 +1,44 @@
1
  ---
2
- base_model: unsloth/deepseek-r1-distill-qwen-1.5b-unsloth-bnb-4bit
3
  library_name: transformers
4
  model_name: askubuntu-model
5
  tags:
6
- - generated_from_trainer
7
  - sft
8
  - unsloth
9
  - trl
10
- licence: license
 
 
 
 
11
  ---
12
-
13
  # Model Card for askubuntu-model
14
 
15
- This model is a fine-tuned version of [unsloth/deepseek-r1-distill-qwen-1.5b-unsloth-bnb-4bit](https://huggingface.co/unsloth/deepseek-r1-distill-qwen-1.5b-unsloth-bnb-4bit).
16
- It has been trained using [TRL](https://github.com/huggingface/trl).
17
 
18
  ## Quick start
19
 
20
  ```python
21
- from transformers import pipeline
22
-
23
- question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?"
24
- generator = pipeline("text-generation", model="maifeeulasad/askubuntu-model", device="cuda")
25
- output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0]
26
- print(output["generated_text"])
27
- ```
28
-
29
- ## Training procedure
30
 
31
-
 
32
 
 
 
 
 
 
33
 
34
- This model was trained with SFT.
35
-
36
- ### Framework versions
37
-
38
- - TRL: 0.19.1
39
- - Transformers: 4.52.4
40
- - Pytorch: 2.7.1
41
- - Datasets: 3.6.0
42
- - Tokenizers: 0.21.2
43
-
44
- ## Citations
45
 
 
 
46
 
 
 
 
47
 
48
- Cite TRL as:
49
-
50
- ```bibtex
51
- @misc{vonwerra2022trl,
52
- title = {{TRL: Transformer Reinforcement Learning}},
53
- author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallou{\'e}dec},
54
- year = 2020,
55
- journal = {GitHub repository},
56
- publisher = {GitHub},
57
- howpublished = {\url{https://github.com/huggingface/trl}}
58
- }
59
  ```
 
1
  ---
2
+ base_model: deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B
3
  library_name: transformers
4
  model_name: askubuntu-model
5
  tags:
 
6
  - sft
7
  - unsloth
8
  - trl
9
+ - deepseek
10
+ - qwen
11
+ licence: agpl-3.0
12
+ datasets:
13
+ - maifeeulasad/askubuntu-data
14
  ---
 
15
  # Model Card for askubuntu-model
16
 
17
+ This model is a fine-tuned version of [deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B](https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B).
 
18
 
19
  ## Quick start
20
 
21
  ```python
22
+ from transformers import AutoTokenizer, AutoModelForCausalLM
23
+ from peft import PeftModel
 
 
 
 
 
 
 
24
 
25
+ base_model_id = "deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B"
26
+ peft_model_id = "maifeeulasad/askubuntu-model"
27
 
28
+ model = AutoModelForCausalLM.from_pretrained(
29
+ base_model_id,
30
+ device_map="auto",
31
+ trust_remote_code=True,
32
+ )
33
 
34
+ model = PeftModel.from_pretrained(model, peft_model_id)
35
+ tokenizer = AutoTokenizer.from_pretrained(base_model_id)
 
 
 
 
 
 
 
 
 
36
 
37
+ from transformers import pipeline
38
+ generator = pipeline("text-generation", model=model, tokenizer=tokenizer)
39
 
40
+ question = "Tell me how to install rootless docker on ubuntu 18 LTS?"
41
+ output = generator(question, max_new_tokens=16384, return_full_text=False)[0]["generated_text"]
42
+ print(output)
43
 
 
 
 
 
 
 
 
 
 
 
 
44
  ```