Vipitis commited on
Commit
79dfc25
1 Parent(s): 94d1a3a

improved README

Browse files
Files changed (1) hide show
  1. README.md +56 -3
README.md CHANGED
@@ -1,4 +1,6 @@
1
  ---
 
 
2
  license: bigcode-openrail-m
3
  datasets:
4
  - bigcode/the-stack-dedup
@@ -6,11 +8,62 @@ datasets:
6
  pipeline_tag: text-generation
7
  tags:
8
  - code
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
9
  ---
10
 
11
- [Santacoder](https://huggingface.co/bigcode/santacoder) finetuned on [Shadertoys-fine](https://huggingface.co/datasets/Vipitis/Shadertoys-fine) for 1000 steps with a batch size of 2 and full sequence length of 2048.
12
- Origianl finetuning script from found [here](https://github.com/loubnabnl/santacoder-finetuning), adapted version to follow (soon^^).
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
13
 
14
- Main purpose of this model is to explore if finetuning models improves performance on [ShaderEval](https://huggingface.co/spaces/Vipitis/ShaderEval), results to follow (sooner).
 
15
 
16
  License carried over from model, however training data has an undefied license. Check details in [Shadertoys](https://huggingface.co/datasets/Vipitis/Shadertoys).
 
1
  ---
2
+ language:
3
+ - code
4
  license: bigcode-openrail-m
5
  datasets:
6
  - bigcode/the-stack-dedup
 
8
  pipeline_tag: text-generation
9
  tags:
10
  - code
11
+ - shader
12
+ widget:
13
+ - text: void mainImage( out vec4 fragColor, in vec2 fragCoord )
14
+ example_title: mainImage
15
+ group: Shadertoy
16
+ model-index:
17
+ - name: santacoder-finetuned-the-stack-glsl
18
+ results:
19
+ - task:
20
+ type: text-generation
21
+ name: ShaderEval
22
+ dataset:
23
+ type: Vipitis/Shadertoys-fine
24
+ name: Shadertoys-fine
25
+ config: return_completion
26
+ revision: 0.0.2
27
+ metrics:
28
+ - type: exact_match
29
+ value: 0.567
30
+ name: 300 samples, greedy decoding
31
+ verified: false
32
+ - type: exact_match
33
+ value: 0.59749
34
+ name: all samples, greedy decoding
35
+ verified: false
36
  ---
37
 
38
+ [Santacoder](https://huggingface.co/bigcode/santacoder) finetuned on [The-Stack-deddup (GLSL subset)](https://huggingface.co/datasets/bigcode/the-stack-dedup/tree/main/data/glsl) for 1000 steps with a batch size of 2 and full sequence length of 2048.
39
+ adapted finetuning script found [here](./train.py)
40
+
41
+ ### Finetuning parameters
42
+ ```sh
43
+ python3 train.py --model_path "bigcode/santacoder" \
44
+ --dataset_name "Vipitis/Shadertoys-fine" \
45
+ --data_column "code" \
46
+ --split "train" \
47
+ --seq_length 2048 \
48
+ --max_steps 1000 \
49
+ --batch_size 2 \
50
+ --gradient_accumulation_steps 4 \
51
+ --learning_rate 5e-5 \
52
+ --num_warmup_steps 100 \
53
+ --eval_freq 100 \
54
+ --save_freq 100 \
55
+ --log_freq 1 \
56
+ --output_dir "checkpoint_dir" \
57
+ --no_fp16
58
+
59
+
60
+ ```
61
+
62
+ Main purpose of this model is to explore if finetuning models improves performance on [ShaderEval](https://huggingface.co/spaces/Vipitis/ShaderEval), which reached 0.567 with 300 samples and 0.59749 on all samples.
63
+
64
+ ### Disclaimer
65
 
66
+ While the train/test split is held out, there is a lot of data contamination. The model results can't be trusted for this simple benchmark.
67
+ Better tasks for the benchmark will be developed and tested against these models.
68
 
69
  License carried over from model, however training data has an undefied license. Check details in [Shadertoys](https://huggingface.co/datasets/Vipitis/Shadertoys).