ammarnasr commited on
Commit
1e3e47f
1 Parent(s): 1bf3ef6

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +37 -19
README.md CHANGED
@@ -1,13 +1,5 @@
1
  ---
2
- configs:
3
- - config_name: default
4
- data_files:
5
- - split: train
6
- path: data/train-*
7
- - split: test
8
- path: data/test-*
9
- - split: valid
10
- path: data/valid-*
11
  dataset_info:
12
  features:
13
  - name: hexsha
@@ -24,17 +16,43 @@ dataset_info:
24
  dtype: float64
25
  splits:
26
  - name: train
27
- num_bytes: 3308287053.655602
28
- num_examples: 893939
29
  - name: test
30
- num_bytes: 183792697.203834
31
- num_examples: 49663
32
  - name: valid
33
- num_bytes: 183796398.001152
34
- num_examples: 49664
35
- download_size: 1305007590
36
- dataset_size: 3675876148.860588
 
 
 
 
 
 
 
 
 
37
  ---
38
- # Dataset Card for "the-stack-ruby-clean"
39
 
40
- [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ license: openrail
 
 
 
 
 
 
 
 
3
  dataset_info:
4
  features:
5
  - name: hexsha
 
16
  dtype: float64
17
  splits:
18
  - name: train
19
+ num_bytes: 3582248477.9086223
20
+ num_examples: 806789
21
  - name: test
22
+ num_bytes: 394048264.9973618
23
+ num_examples: 88747
24
  - name: valid
25
+ num_bytes: 3982797.09401595
26
+ num_examples: 897
27
+ download_size: 1323156008
28
+ dataset_size: 3980279540
29
+ task_categories:
30
+ - text-generation
31
+ language:
32
+ - code
33
+ tags:
34
+ - code
35
+ pretty_name: TheStack-Ruby
36
+ size_categories:
37
+ - 1M<n<10M
38
  ---
 
39
 
40
+ ## Dataset 1: TheStack - Ruby - Cleaned
41
+
42
+ **Description**: This dataset is drawn from TheStack Corpus, an open-source code dataset with over 3TB of GitHub data covering 48 programming languages. We selected a small portion of this dataset to optimize smaller language models for Ruby, a popular statically typed language.
43
+
44
+ **Target Language**: Ruby
45
+
46
+ **Dataset Size**:
47
+ - Training: 900,000 files
48
+ - Validation: 50,000 files
49
+ - Test: 50,000 files
50
+
51
+ **Preprocessing**:
52
+ 1. Selected Ruby as the target language due to its popularity on GitHub.
53
+ 2. Filtered out files with average line length > 100 characters, maximum line length > 1000 characters, and alphabet ratio < 25%.
54
+ 3. Split files into 90% training, 5% validation, and 5% test sets.
55
+
56
+ **Tokenizer**: Byte Pair Encoding (BPE) tokenizer with tab and whitespace tokens. GPT-2 vocabulary extended with special tokens.
57
+
58
+ **Training Sequences**: Sequences constructed by joining training data text to reach a context length of 2048 tokens (1024 tokens for full fine-tuning).