ajibawa-2023 commited on
Commit
506f4a2
1 Parent(s): f271bf1

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +37 -0
README.md CHANGED
@@ -1,3 +1,40 @@
1
  ---
2
  license: cc-by-nc-nd-4.0
 
 
 
 
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: cc-by-nc-nd-4.0
3
+ datasets:
4
+ - ajibawa-2023/Python-Code-23k-ShareGPT
5
+ language:
6
+ - en
7
+ tags:
8
+ - code
9
  ---
10
+
11
+ **Python-Code-33B**
12
+
13
+ Large Language Models (LLMs) are good with code generations. Sometimes LLMs do make mistakes in code generation. How about if they can give detailed explanation along with the code.
14
+ This is what I have tried over here. The base Llama-2 model was used for training purpose. It is trained on around 23000+ set of codes. Each set having 2 conversations.
15
+ This data was generated using GPT-3.5, GPT-4 etc. This conversation is in Vicuna/ShareGPT format. Each set, along with code, has detailed explanation.
16
+ I have released the [data](https://huggingface.co/datasets/ajibawa-2023/Python-Code-23k-ShareGPT).
17
+
18
+ **Training:**
19
+ Entire dataset was trained on Azure 4 x A100 80GB. For 3 epoch, training took 42 hours. DeepSpeed codebase was used for training purpose. This was trained on Llama-2 by Meta.
20
+
21
+
22
+ **GPTQ GGML & AWQ**
23
+
24
+ GPTQ: TBA
25
+
26
+ GGUF: TBA
27
+
28
+ AWQ: TBA
29
+
30
+
31
+ **Example Prompt:**
32
+ ```
33
+ This is a conversation with your helpful AI assistant. AI assistant can generate Python Code along with necessary explanation.
34
+
35
+ Context
36
+ You are a helpful AI assistant.
37
+
38
+ USER: <prompt>
39
+ ASSISTANT:
40
+ ```