afrideva commited on
Commit
2c64be0
1 Parent(s): f7be321

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +117 -0
README.md ADDED
@@ -0,0 +1,117 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: BEE-spoke-data/beecoder-220M-python
3
+ datasets:
4
+ - BEE-spoke-data/pypi_clean-deduped
5
+ - bigcode/the-stack-smol-xl
6
+ - EleutherAI/proof-pile-2
7
+ inference: false
8
+ language:
9
+ - en
10
+ license: apache-2.0
11
+ metrics:
12
+ - accuracy
13
+ model_creator: BEE-spoke-data
14
+ model_name: beecoder-220M-python
15
+ pipeline_tag: text-generation
16
+ quantized_by: afrideva
17
+ tags:
18
+ - python
19
+ - codegen
20
+ - markdown
21
+ - smol_llama
22
+ - gguf
23
+ - ggml
24
+ - quantized
25
+ - q2_k
26
+ - q3_k_m
27
+ - q4_k_m
28
+ - q5_k_m
29
+ - q6_k
30
+ - q8_0
31
+ widget:
32
+ - example_title: Add Numbers Function
33
+ text: "def add_numbers(a, b):\n return\n"
34
+ - example_title: Car Class
35
+ text: "class Car:\n def __init__(self, make, model):\n self.make = make\n
36
+ \ self.model = model\n\n def display_car(self):\n"
37
+ - example_title: Pandas DataFrame
38
+ text: 'import pandas as pd
39
+
40
+ data = {''Name'': [''Tom'', ''Nick'', ''John''], ''Age'': [20, 21, 19]}
41
+
42
+ df = pd.DataFrame(data).convert_dtypes()
43
+
44
+ # eda
45
+
46
+ '
47
+ - example_title: Factorial Function
48
+ text: "def factorial(n):\n if n == 0:\n return 1\n else:\n"
49
+ - example_title: Fibonacci Function
50
+ text: "def fibonacci(n):\n if n <= 0:\n raise ValueError(\"Incorrect input\")\n
51
+ \ elif n == 1:\n return 0\n elif n == 2:\n return 1\n else:\n"
52
+ - example_title: Matplotlib Plot
53
+ text: 'import matplotlib.pyplot as plt
54
+
55
+ import numpy as np
56
+
57
+ x = np.linspace(0, 10, 100)
58
+
59
+ # simple plot
60
+
61
+ '
62
+ - example_title: Reverse String Function
63
+ text: "def reverse_string(s:str) -> str:\n return\n"
64
+ - example_title: Palindrome Function
65
+ text: "def is_palindrome(word:str) -> bool:\n return\n"
66
+ - example_title: Bubble Sort Function
67
+ text: "def bubble_sort(lst: list):\n n = len(lst)\n for i in range(n):\n for
68
+ j in range(0, n-i-1):\n"
69
+ - example_title: Binary Search Function
70
+ text: "def binary_search(arr, low, high, x):\n if high >= low:\n mid =
71
+ (high + low) // 2\n if arr[mid] == x:\n return mid\n elif
72
+ arr[mid] > x:\n"
73
+ ---
74
+ # BEE-spoke-data/beecoder-220M-python-GGUF
75
+
76
+ Quantized GGUF model files for [beecoder-220M-python](https://huggingface.co/BEE-spoke-data/beecoder-220M-python) from [BEE-spoke-data](https://huggingface.co/BEE-spoke-data)
77
+
78
+
79
+ | Name | Quant method | Size |
80
+ | ---- | ---- | ---- |
81
+ | [beecoder-220m-python.fp16.gguf](https://huggingface.co/afrideva/beecoder-220M-python-GGUF/resolve/main/beecoder-220m-python.fp16.gguf) | fp16 | 436.50 MB |
82
+ | [beecoder-220m-python.q2_k.gguf](https://huggingface.co/afrideva/beecoder-220M-python-GGUF/resolve/main/beecoder-220m-python.q2_k.gguf) | q2_k | 94.43 MB |
83
+ | [beecoder-220m-python.q3_k_m.gguf](https://huggingface.co/afrideva/beecoder-220M-python-GGUF/resolve/main/beecoder-220m-python.q3_k_m.gguf) | q3_k_m | 114.65 MB |
84
+ | [beecoder-220m-python.q4_k_m.gguf](https://huggingface.co/afrideva/beecoder-220M-python-GGUF/resolve/main/beecoder-220m-python.q4_k_m.gguf) | q4_k_m | 137.58 MB |
85
+ | [beecoder-220m-python.q5_k_m.gguf](https://huggingface.co/afrideva/beecoder-220M-python-GGUF/resolve/main/beecoder-220m-python.q5_k_m.gguf) | q5_k_m | 157.91 MB |
86
+ | [beecoder-220m-python.q6_k.gguf](https://huggingface.co/afrideva/beecoder-220M-python-GGUF/resolve/main/beecoder-220m-python.q6_k.gguf) | q6_k | 179.52 MB |
87
+ | [beecoder-220m-python.q8_0.gguf](https://huggingface.co/afrideva/beecoder-220M-python-GGUF/resolve/main/beecoder-220m-python.q8_0.gguf) | q8_0 | 232.28 MB |
88
+
89
+
90
+
91
+ ## Original Model Card:
92
+ # BEE-spoke-data/beecoder-220M-python
93
+
94
+
95
+
96
+
97
+ This is `BEE-spoke-data/smol_llama-220M-GQA` fine-tuned for code generation on:
98
+
99
+ - filtered version of stack-smol-XL
100
+ - deduped version of 'algebraic stack' from proof-pile-2
101
+ - cleaned and deduped pypi (last dataset)
102
+
103
+ This model (and the base model) were both trained using ctx length 2048.
104
+
105
+ ## examples
106
+
107
+ > Example script for inference testing: [here](https://gist.github.com/pszemraj/c7738f664a64b935a558974d23a7aa8c)
108
+
109
+ It has its limitations at 220M, but seems decent for single-line or docstring generation, and/or being used for speculative decoding for such purposes.
110
+
111
+
112
+
113
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/60bccec062080d33f875cd0c/bLrtpr7Vi_MPvtF7mozDN.png)
114
+
115
+ The screenshot is on CPU on a laptop.
116
+
117
+ ---