lysandre HF staff commited on
Commit
d1d7b7f
1 Parent(s): 69dadea

Upload dataset

Browse files
README.md ADDED
@@ -0,0 +1,105 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ dataset_info:
3
+ features:
4
+ - name: dates
5
+ dtype: string
6
+ - name: type
7
+ struct:
8
+ - name: authorAssociation
9
+ dtype: string
10
+ - name: comment
11
+ dtype: bool
12
+ - name: issue
13
+ dtype: bool
14
+ splits:
15
+ - name: transformers
16
+ num_bytes: 4970577
17
+ num_examples: 140143
18
+ - name: peft
19
+ num_bytes: 267891
20
+ num_examples: 7813
21
+ - name: evaluate
22
+ num_bytes: 65695
23
+ num_examples: 1877
24
+ - name: huggingface_hub
25
+ num_bytes: 316054
26
+ num_examples: 9053
27
+ - name: accelerate
28
+ num_bytes: 398411
29
+ num_examples: 11393
30
+ - name: datasets
31
+ num_bytes: 843933
32
+ num_examples: 24046
33
+ - name: optimum
34
+ num_bytes: 209726
35
+ num_examples: 6044
36
+ - name: pytorch_image_models
37
+ num_bytes: 148186
38
+ num_examples: 4294
39
+ - name: gradio
40
+ num_bytes: 1224430
41
+ num_examples: 33731
42
+ - name: tokenizers
43
+ num_bytes: 203872
44
+ num_examples: 5956
45
+ - name: diffusers
46
+ num_bytes: 1519391
47
+ num_examples: 42441
48
+ - name: safetensors
49
+ num_bytes: 51043
50
+ num_examples: 1478
51
+ - name: sentence_transformers
52
+ num_bytes: 317985
53
+ num_examples: 9325
54
+ - name: candle
55
+ num_bytes: 189838
56
+ num_examples: 4997
57
+ - name: text_generation_inference
58
+ num_bytes: 226337
59
+ num_examples: 6685
60
+ - name: chat_ui
61
+ num_bytes: 99559
62
+ num_examples: 2864
63
+ - name: hub_docs
64
+ num_bytes: 142244
65
+ num_examples: 4042
66
+ download_size: 3481791
67
+ dataset_size: 11195172
68
+ configs:
69
+ - config_name: default
70
+ data_files:
71
+ - split: transformers
72
+ path: data/transformers-*
73
+ - split: peft
74
+ path: data/peft-*
75
+ - split: evaluate
76
+ path: data/evaluate-*
77
+ - split: huggingface_hub
78
+ path: data/huggingface_hub-*
79
+ - split: accelerate
80
+ path: data/accelerate-*
81
+ - split: datasets
82
+ path: data/datasets-*
83
+ - split: optimum
84
+ path: data/optimum-*
85
+ - split: pytorch_image_models
86
+ path: data/pytorch_image_models-*
87
+ - split: gradio
88
+ path: data/gradio-*
89
+ - split: tokenizers
90
+ path: data/tokenizers-*
91
+ - split: diffusers
92
+ path: data/diffusers-*
93
+ - split: safetensors
94
+ path: data/safetensors-*
95
+ - split: sentence_transformers
96
+ path: data/sentence_transformers-*
97
+ - split: candle
98
+ path: data/candle-*
99
+ - split: text_generation_inference
100
+ path: data/text_generation_inference-*
101
+ - split: chat_ui
102
+ path: data/chat_ui-*
103
+ - split: hub_docs
104
+ path: data/hub_docs-*
105
+ ---
data/accelerate-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2a44677c7e92c87db5b438ffa8c79904cd8926db0104a8528f5ac94ff6b55626
3
+ size 131626
data/candle-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bc5a011cd6df5142e6fb60ae884f36fc9ee8fe9f654e1d523afbeb23f84512cd
3
+ size 59065
data/chat_ui-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9fd4eff3f642cafbd668e3af6de809d4729c84af0963fa2dc28ba6d0a129cb64
3
+ size 34805
data/datasets-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8f048c57f9b42ebd2c1b6aae017c3f28d3c6bb9e1e05288a6509164252675ec6
3
+ size 272935
data/diffusers-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:918f1ddf9da529373653c4d96d15d220f3e47524cf7a884c8668607635a0200a
3
+ size 449423
data/evaluate-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8a8b52a641758708fc55e98f5c6866227bc6a31d3b198decafe68f12e21a3bbb
3
+ size 24123
data/gradio-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a799e5eb0779056b7ae242da0ae8b5bc9dbcf9ef739f82516897834c18f05c1f
3
+ size 366400
data/hub_docs-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d7d59fad08dea3e1ef3ee7b5e51ccf6f69bde0be2f1e793a6bed6b3af3f9129b
3
+ size 49594
data/huggingface_hub-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a48a6b726f44a3c68cc5c63f78f3192389065e84db11f5bbf58249a516309e21
3
+ size 106272
data/optimum-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7fbb6dafa3ac6bfbc1f08f5fbb889f8f3c9a6c3ca801fe5acdf4f0bc96b5be21
3
+ size 73080
data/peft-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4c5f057e53757de7f2dcb8dcae3b05724a63d15a250e1e96b821508d58dad930
3
+ size 88510
data/pytorch_image_models-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a759cd68e00044d8edbecf3c2a25d31d5d3b6bd1dad8e877db4652a6264bd24b
3
+ size 55428
data/safetensors-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2202d0aefd423ee18761b0eb3b6f48457825819dfec44cc4de86b6779d6fd0d5
3
+ size 19606
data/sentence_transformers-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:83c7cf5cd32f894099074ec520b6ad6e5f7707600011df77c0f1dfc55e3c6c44
3
+ size 112852
data/text_generation_inference-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a434d27a5c142223bb97c03b6f2486785e7a05f457656cde7a4b95b590f3c37c
3
+ size 76828
data/tokenizers-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:89fa71515e06024dcc1264d18c80b208698fab1a2206fdc45b8b5eacfcafe215
3
+ size 73283
data/transformers-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:669f0bf57b38d82c8858873a0b67e43c3d6f5ec85554993602d43d005fa20711
3
+ size 1487961
dataset_infos.json CHANGED
@@ -28,14 +28,14 @@
28
  "splits": {
29
  "transformers": {
30
  "name": "transformers",
31
- "num_bytes": 4968432,
32
- "num_examples": 140081,
33
  "dataset_name": null
34
  },
35
  "peft": {
36
  "name": "peft",
37
- "num_bytes": 267478,
38
- "num_examples": 7805,
39
  "dataset_name": null
40
  },
41
  "evaluate": {
@@ -46,26 +46,26 @@
46
  },
47
  "huggingface_hub": {
48
  "name": "huggingface_hub",
49
- "num_bytes": 315818,
50
- "num_examples": 9046,
51
  "dataset_name": null
52
  },
53
  "accelerate": {
54
  "name": "accelerate",
55
- "num_bytes": 397725,
56
- "num_examples": 11373,
57
  "dataset_name": null
58
  },
59
  "datasets": {
60
  "name": "datasets",
61
- "num_bytes": 843619,
62
- "num_examples": 24037,
63
  "dataset_name": null
64
  },
65
  "optimum": {
66
  "name": "optimum",
67
- "num_bytes": 209594,
68
- "num_examples": 6040,
69
  "dataset_name": null
70
  },
71
  "pytorch_image_models": {
@@ -76,8 +76,8 @@
76
  },
77
  "gradio": {
78
  "name": "gradio",
79
- "num_bytes": 1223655,
80
- "num_examples": 33710,
81
  "dataset_name": null
82
  },
83
  "tokenizers": {
@@ -88,8 +88,8 @@
88
  },
89
  "diffusers": {
90
  "name": "diffusers",
91
- "num_bytes": 1518288,
92
- "num_examples": 42410,
93
  "dataset_name": null
94
  },
95
  "safetensors": {
@@ -100,37 +100,37 @@
100
  },
101
  "sentence_transformers": {
102
  "name": "sentence_transformers",
103
- "num_bytes": 317781,
104
- "num_examples": 9319,
105
  "dataset_name": null
106
  },
107
  "candle": {
108
  "name": "candle",
109
- "num_bytes": 189597,
110
- "num_examples": 4991,
111
  "dataset_name": null
112
  },
113
  "text_generation_inference": {
114
  "name": "text_generation_inference",
115
- "num_bytes": 226024,
116
- "num_examples": 6676,
117
  "dataset_name": null
118
  },
119
  "chat_ui": {
120
  "name": "chat_ui",
121
- "num_bytes": 99223,
122
- "num_examples": 2855,
123
  "dataset_name": null
124
  },
125
  "hub_docs": {
126
  "name": "hub_docs",
127
- "num_bytes": 142138,
128
- "num_examples": 4039,
129
  "dataset_name": null
130
  }
131
  },
132
- "download_size": 3479905,
133
- "dataset_size": 11188168,
134
- "size_in_bytes": 14668073
135
  }
136
  }
28
  "splits": {
29
  "transformers": {
30
  "name": "transformers",
31
+ "num_bytes": 4970577,
32
+ "num_examples": 140143,
33
  "dataset_name": null
34
  },
35
  "peft": {
36
  "name": "peft",
37
+ "num_bytes": 267891,
38
+ "num_examples": 7813,
39
  "dataset_name": null
40
  },
41
  "evaluate": {
46
  },
47
  "huggingface_hub": {
48
  "name": "huggingface_hub",
49
+ "num_bytes": 316054,
50
+ "num_examples": 9053,
51
  "dataset_name": null
52
  },
53
  "accelerate": {
54
  "name": "accelerate",
55
+ "num_bytes": 398411,
56
+ "num_examples": 11393,
57
  "dataset_name": null
58
  },
59
  "datasets": {
60
  "name": "datasets",
61
+ "num_bytes": 843933,
62
+ "num_examples": 24046,
63
  "dataset_name": null
64
  },
65
  "optimum": {
66
  "name": "optimum",
67
+ "num_bytes": 209726,
68
+ "num_examples": 6044,
69
  "dataset_name": null
70
  },
71
  "pytorch_image_models": {
76
  },
77
  "gradio": {
78
  "name": "gradio",
79
+ "num_bytes": 1224430,
80
+ "num_examples": 33731,
81
  "dataset_name": null
82
  },
83
  "tokenizers": {
88
  },
89
  "diffusers": {
90
  "name": "diffusers",
91
+ "num_bytes": 1519391,
92
+ "num_examples": 42441,
93
  "dataset_name": null
94
  },
95
  "safetensors": {
100
  },
101
  "sentence_transformers": {
102
  "name": "sentence_transformers",
103
+ "num_bytes": 317985,
104
+ "num_examples": 9325,
105
  "dataset_name": null
106
  },
107
  "candle": {
108
  "name": "candle",
109
+ "num_bytes": 189838,
110
+ "num_examples": 4997,
111
  "dataset_name": null
112
  },
113
  "text_generation_inference": {
114
  "name": "text_generation_inference",
115
+ "num_bytes": 226337,
116
+ "num_examples": 6685,
117
  "dataset_name": null
118
  },
119
  "chat_ui": {
120
  "name": "chat_ui",
121
+ "num_bytes": 99559,
122
+ "num_examples": 2864,
123
  "dataset_name": null
124
  },
125
  "hub_docs": {
126
  "name": "hub_docs",
127
+ "num_bytes": 142244,
128
+ "num_examples": 4042,
129
  "dataset_name": null
130
  }
131
  },
132
+ "download_size": 3481791,
133
+ "dataset_size": 11195172,
134
+ "size_in_bytes": 14676963
135
  }
136
  }