mav23 commited on
Commit
16e8ac9
1 Parent(s): 4740b57

Upload folder using huggingface_hub

Browse files
Files changed (3) hide show
  1. .gitattributes +1 -0
  2. README.md +99 -0
  3. dracarys2-72b-instruct.Q4_0.gguf +3 -0
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ dracarys2-72b-instruct.Q4_0.gguf filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,99 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - en
4
+ license: other
5
+ tags:
6
+ - chat
7
+ license_name: tongyi-qianwen
8
+ license_link: https://huggingface.co/Qwen/Qwen2.5-72B-Instruct/blob/main/LICENSE
9
+ pipeline_tag: text-generation
10
+ ---
11
+
12
+ # Dracarys2-72B-Instruct
13
+
14
+ # Introduction
15
+
16
+ We introduce the latest in the Smaug series, the Dracarys family of finetunes targeting coding performance improvements
17
+ across a variety of base models.
18
+
19
+ This variant is a finetune of [Qwen2.5-72B-Instruct](https://huggingface.co/Qwen/Qwen2.5-72B-Instruct)
20
+
21
+ Compared to Qwen2.5-72B-Instruct, Dracarys has better LiveCodeBench scores (see evaluation results below).
22
+
23
+ ### Model Description
24
+
25
+ - **Developed by:** [Abacus.AI](https://abacus.ai)
26
+ - **License:** https://huggingface.co/Qwen/Qwen2.5-72B-Instruct/blob/main/LICENSE
27
+ - **Finetuned from model:** [Qwen2.5-72B-Instruct](https://huggingface.co/Qwen/Qwen2.5-72B-Instruct).
28
+
29
+ ## How to use
30
+
31
+ The prompt format is unchanged from Qwen2.5-72B-Instruct (see evaluations for prompt details for LCB)
32
+
33
+ ### Use with transformers
34
+
35
+ See the snippet below for usage with Transformers:
36
+
37
+ ```python
38
+ import transformers
39
+ import torch
40
+
41
+ model_id = "abacusai/Dracarys2-72B-Instruct"
42
+
43
+ pipeline = transformers.pipeline(
44
+ "text-generation",
45
+ model=model_id,
46
+ model_kwargs={"torch_dtype": torch.bfloat16},
47
+ device_map="auto",
48
+ )
49
+
50
+ messages = [
51
+ {"role": "system", "content": "You are data science coding assistant that generates Python code using Pandas and Numpy."},
52
+ {"role": "user", "content": "Write code to select rows from the dataframe `df` having the maximum `temp` for each `city`"},
53
+ ]
54
+
55
+ prompt = pipeline.tokenizer.apply_chat_template(
56
+ messages,
57
+ tokenize=False,
58
+ add_generation_prompt=True
59
+ )
60
+
61
+ terminators = [
62
+ pipeline.tokenizer.eos_token_id,
63
+ pipeline.tokenizer.convert_tokens_to_ids("<|eot_id|>")
64
+ ]
65
+
66
+ outputs = pipeline(
67
+ prompt,
68
+ max_new_tokens=256,
69
+ eos_token_id=terminators,
70
+ do_sample=True,
71
+ temperature=0.6,
72
+ top_p=0.9,
73
+ )
74
+ print(outputs[0]["generated_text"][len(prompt):])
75
+ ```
76
+
77
+ # Evaluation Results
78
+
79
+
80
+ ## LiveCodeBench
81
+
82
+ | Model | Code Generation | Code Execution (COT) |Test Output Prediction |
83
+ |----------------------------|-----------------|----------------------|-----------------------|
84
+ | **Dracarys2-72B-Instruct** | **53.80** | **89.12** | **59.61** |
85
+ | Qwen2.5-72B-Instruct | 53.03 | 88.72 | 46.28 |
86
+
87
+ ## Breakdown of LiveCodeBench CodeGeneration
88
+
89
+ | Model | Easy | Medium | Hard |
90
+ |---------------------------|-----------------|----------------|---------------|
91
+ | **Dracarys2-72B-Instruct**| **88.79** | **50.28** | 9.47 |
92
+ | Qwen2.5-72B-Instruct | 86.99 | 49.59 | 9.99 |
93
+
94
+ ## Breakdown of LiveCodeBench TestOutputPrediction
95
+
96
+ | Model | Easy | Medium | Hard |
97
+ |---------------------------|-----------------|----------------|-----------------------|
98
+ | **Dracarys2-72B-Instruct**| **79.25** | **53.76** | **37.63** |
99
+ | Qwen2.5-72B-Instruct | 68.43 | 39.46 | 22.22 |
dracarys2-72b-instruct.Q4_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2f467b05cc82767d3f21d532cd4d69b3c72ae51bbf97af4ec787a518e60a5779
3
+ size 41231738016