Pankaj Mathur commited on
Commit
58af173
1 Parent(s): 335b148

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +12 -12
README.md CHANGED
@@ -2,11 +2,13 @@
2
  language:
3
  - en
4
  library_name: transformers
 
 
5
  ---
6
 
7
  # model_42_70b
8
 
9
- A Llama2-70b model trained on Orca Style datasets.
10
 
11
  **I am actively seeking sponsorship and partnership opportunities. If you're interested, please connect with me at www.linkedin.com/in/pankajam.**
12
 
@@ -31,11 +33,9 @@ Here are the results on metrics used by [HuggingFaceH4 Open LLM Leaderboard](htt
31
  Here is the prompt format
32
 
33
  ```
34
- ### System:
35
- You are an AI assistant that follows instruction extremely well. Help as much as you can.
36
 
37
  ### User:
38
- Tell me about Orcas.
39
 
40
  ### Assistant:
41
 
@@ -55,11 +55,10 @@ model = AutoModelForCausalLM.from_pretrained(
55
  low_cpu_mem_usage=True,
56
  device_map="auto"
57
  )
58
- system_prompt = "### System:\nYou are an AI assistant that follows instruction extremely well. Help as much as you can.\n\n"
59
 
60
  #generate text steps
61
- instruction = "Tell me about Orcas."
62
- prompt = f"{system_prompt}### User: {instruction}\n\n### Assistant:\n"
63
  inputs = tokenizer(prompt, return_tensors="pt").to("cuda")
64
  output = model.generate(**inputs, do_sample=True, top_p=0.95, top_k=0, max_new_tokens=4096)
65
 
@@ -85,7 +84,7 @@ Please kindly cite using the following BibTeX:
85
  ```
86
  @misc{model_42_70b,
87
  author = {Pankaj Mathur},
88
- title = {model_42_70b: An Orca Style Llama2-70b model},
89
  year = {2023},
90
  publisher = {HuggingFace},
91
  journal = {HuggingFace repository},
@@ -94,11 +93,12 @@ Please kindly cite using the following BibTeX:
94
  ```
95
 
96
  ```
97
- @misc{mukherjee2023orca,
98
- title={Orca: Progressive Learning from Complex Explanation Traces of GPT-4},
99
- author={Subhabrata Mukherjee and Arindam Mitra and Ganesh Jawahar and Sahaj Agarwal and Hamid Palangi and Ahmed Awadallah},
 
100
  year={2023},
101
- eprint={2306.02707},
102
  archivePrefix={arXiv},
103
  primaryClass={cs.CL}
104
  }
 
2
  language:
3
  - en
4
  library_name: transformers
5
+ datasets:
6
+ - GAIR/lima
7
  ---
8
 
9
  # model_42_70b
10
 
11
+ A Llama2-70b model fine tuned on lima datasets, which has < ~1k conversations.
12
 
13
  **I am actively seeking sponsorship and partnership opportunities. If you're interested, please connect with me at www.linkedin.com/in/pankajam.**
14
 
 
33
  Here is the prompt format
34
 
35
  ```
 
 
36
 
37
  ### User:
38
+ Write a stand-up skit in the style of George Carlin that ridicules Pacific Gas and Electric.
39
 
40
  ### Assistant:
41
 
 
55
  low_cpu_mem_usage=True,
56
  device_map="auto"
57
  )
 
58
 
59
  #generate text steps
60
+ instruction = "Write a stand-up skit in the style of George Carlin that ridicules Pacific Gas and Electric."
61
+ prompt = f"### User: {instruction}\n\n### Assistant:\n"
62
  inputs = tokenizer(prompt, return_tensors="pt").to("cuda")
63
  output = model.generate(**inputs, do_sample=True, top_p=0.95, top_k=0, max_new_tokens=4096)
64
 
 
84
  ```
85
  @misc{model_42_70b,
86
  author = {Pankaj Mathur},
87
+ title = {model_42_70b: A LIMA style Llama2-70b model},
88
  year = {2023},
89
  publisher = {HuggingFace},
90
  journal = {HuggingFace repository},
 
93
  ```
94
 
95
  ```
96
+ @misc{ChuntingZhou,
97
+ title={LIMA: Less Is More for Alignment},
98
+ author={Chunting Zhou, Pengfei Liu, Puxin Xu, Srini Iyer, Jiao Sun, Yuning Mao, Xuezhe Ma, Avia Efrat, Ping Yu, Lili Yu,
99
+ Susan Zhang, Gargi Ghosh, Mike Lewis, Luke Zettlemoyer, Omer Levy},
100
  year={2023},
101
+ eprint={2305.11206},
102
  archivePrefix={arXiv},
103
  primaryClass={cs.CL}
104
  }