jeffliu-LL commited on
Commit
f5fe0c6
1 Parent(s): 3def176

arrow version of dataset

Browse files
Files changed (38) hide show
  1. LADI-v2-dataset.py +0 -278
  2. data/dataset_dict.json +1 -0
  3. data/test/data-00000-of-00003.arrow +3 -0
  4. data/test/data-00001-of-00003.arrow +3 -0
  5. data/test/data-00002-of-00003.arrow +3 -0
  6. data/test/dataset_info.json +218 -0
  7. data/test/state.json +19 -0
  8. data/train/data-00000-of-00024.arrow +3 -0
  9. data/train/data-00001-of-00024.arrow +3 -0
  10. data/train/data-00002-of-00024.arrow +3 -0
  11. data/train/data-00003-of-00024.arrow +3 -0
  12. data/train/data-00004-of-00024.arrow +3 -0
  13. data/train/data-00005-of-00024.arrow +3 -0
  14. data/train/data-00006-of-00024.arrow +3 -0
  15. data/train/data-00007-of-00024.arrow +3 -0
  16. data/train/data-00008-of-00024.arrow +3 -0
  17. data/train/data-00009-of-00024.arrow +3 -0
  18. data/train/data-00010-of-00024.arrow +3 -0
  19. data/train/data-00011-of-00024.arrow +3 -0
  20. data/train/data-00012-of-00024.arrow +3 -0
  21. data/train/data-00013-of-00024.arrow +3 -0
  22. data/train/data-00014-of-00024.arrow +3 -0
  23. data/train/data-00015-of-00024.arrow +3 -0
  24. data/train/data-00016-of-00024.arrow +3 -0
  25. data/train/data-00017-of-00024.arrow +3 -0
  26. data/train/data-00018-of-00024.arrow +3 -0
  27. data/train/data-00019-of-00024.arrow +3 -0
  28. data/train/data-00020-of-00024.arrow +3 -0
  29. data/train/data-00021-of-00024.arrow +3 -0
  30. data/train/data-00022-of-00024.arrow +3 -0
  31. data/train/data-00023-of-00024.arrow +3 -0
  32. data/train/dataset_info.json +218 -0
  33. data/train/state.json +82 -0
  34. data/validation/data-00000-of-00003.arrow +3 -0
  35. data/validation/data-00001-of-00003.arrow +3 -0
  36. data/validation/data-00002-of-00003.arrow +3 -0
  37. data/validation/dataset_info.json +218 -0
  38. data/validation/state.json +19 -0
LADI-v2-dataset.py DELETED
@@ -1,278 +0,0 @@
1
- import cv2
2
- import datasets
3
- import pandas as pd
4
- from datasets.data_files import DataFilesDict, sanitize_patterns
5
- from pathlib import Path
6
- from PIL import Image, ImageFile
7
-
8
- from typing import List, Optional
9
-
10
- ImageFile.LOAD_TRUNCATED_IMAGES = True
11
-
12
- # maps the dataset names to names for the image sets they rely on
13
- DATA_NAME_MAP = {
14
- 'v1_damage': 'v1',
15
- 'v1_infrastructure': 'v1',
16
- 'v2': 'v2',
17
- 'v2_resized': 'v2_resized',
18
- 'v2a': 'v2',
19
- 'v2a_resized': 'v2_resized'
20
- }
21
-
22
- DATA_URLS = {'v1': "https://ladi.s3.amazonaws.com/ladi_v1.tar.gz",
23
- 'v2': 'https://ladi.s3.amazonaws.com/ladi_v2.tar.gz',
24
- 'v2_resized': 'https://ladi.s3.amazonaws.com/ladi_v2_resized.tar.gz'}
25
-
26
- SPLIT_REL_PATHS = {
27
- # note: the v1 datasets don't have separate 'test' and 'val' splits
28
- 'v1_damage': {'train':'v1/damage_dataset/damage_df_train.csv',
29
- 'val':'v1/damage_dataset/damage_df_test.csv',
30
- 'test':'v1/damage_dataset/damage_df_test.csv',
31
- 'all': 'v1/damage_dataset/damage_df.csv'},
32
- 'v1_infrastructure': {'train':'v1/infra_dataset/infra_df_train.csv',
33
- 'val':'v1/infra_dataset/infra_df_test.csv',
34
- 'test':'v1/infra_dataset/infra_df_test.csv',
35
- 'all':'v1/infra_dataset/infra_df.csv'},
36
- 'v2': {'train':'v2/ladi_v2_labels_train.csv',
37
- 'val':'v2/ladi_v2_labels_val.csv',
38
- 'test':'v2/ladi_v2_labels_test.csv',
39
- 'all':'v2/ladi_v2_labels_train_full.csv'},
40
- 'v2_resized': {'train':'v2/ladi_v2_labels_train_resized.csv',
41
- 'val':'v2/ladi_v2_labels_val_resized.csv',
42
- 'test':'v2/ladi_v2_labels_test_resized.csv',
43
- 'all':'v2/ladi_v2_labels_train_full_resized.csv'},
44
- 'v2a': {'train':'v2/ladi_v2a_labels_train.csv',
45
- 'val':'v2/ladi_v2a_labels_val.csv',
46
- 'test':'v2/ladi_v2a_labels_test.csv',
47
- 'all':'v2/ladi_v2a_labels_train_full.csv'},
48
- 'v2a_resized': {'train':'v2/ladi_v2a_labels_train_resized.csv',
49
- 'val':'v2/ladi_v2a_labels_val_resized.csv',
50
- 'test':'v2/ladi_v2a_labels_test_resized.csv',
51
- 'all':'v2/ladi_v2a_labels_train_full_resized.csv'}
52
- }
53
-
54
- class LadiClassifyDatasetConfig(datasets.BuilderConfig):
55
- def __init__(self,
56
- name: str = 'v2a_resized',
57
- base_dir: Optional[str] = None,
58
- split_csvs = None,
59
- download_ladi = False,
60
- data_name: Optional[str] = None,
61
- label_name: Optional[str] = None,
62
- **kwargs):
63
- """
64
- split_csvs: a dictionary mapping split names to existing csv files containing annotations
65
- if this arg is set, you MUST already have the dataset
66
- base_dir: the base directory of the label CSVs and data files.
67
- data_name: the version of the data you're using. Used to determine what files to download if
68
- you don't specify split_csvs or url_list. Must be in DATA_URLS.keys().
69
-
70
- If split_csvs is None, the requested data will be downloaded from the hub. Please do NOT
71
- use this feature with streaming=True, you will perform a large download every time.
72
- """
73
- self.download_ladi = download_ladi
74
- self.data_name = DATA_NAME_MAP[name] if data_name is None else data_name
75
- self.label_name = name if label_name is None else label_name
76
- self.base_dir = None if base_dir is None else Path(base_dir)
77
- self.split_csvs = split_csvs
78
-
79
- if self.data_name not in DATA_URLS.keys():
80
- raise ValueError(f"Expected data_name to be one of {DATA_URLS.keys()}, got {self.data_name}")
81
-
82
- if split_csvs is None and download_ladi == False:
83
- self.split_csvs = SPLIT_REL_PATHS[self.label_name]
84
-
85
- super(LadiClassifyDatasetConfig, self).__init__(name=name, **kwargs)
86
-
87
-
88
- class LADIClassifyDataset(datasets.GeneratorBasedBuilder):
89
- """
90
- Dataset for LADI Classification task
91
- """
92
-
93
- VERSION = datasets.Version("0.2.1")
94
- BUILDER_CONFIG_CLASS = LadiClassifyDatasetConfig
95
- DEFAULT_CONFIG_NAME = 'v2a_resized'
96
-
97
- BUILDER_CONFIGS = [
98
- LadiClassifyDatasetConfig(
99
- name='v1_damage',
100
- version=VERSION,
101
- description="Dataset for recognizing damage (flood, rubble, misc) from LADI"
102
- ),
103
- LadiClassifyDatasetConfig(
104
- name="v1_infrastructure",
105
- version=VERSION,
106
- description="Dataset for recognizing infrastructure (buildings, roads) from LADI"
107
- ),
108
- LadiClassifyDatasetConfig(
109
- name="v2",
110
- version=VERSION,
111
- description="Dataset using the v2 labels for LADI"
112
- ),
113
- LadiClassifyDatasetConfig(
114
- name="v2_resized",
115
- version=VERSION,
116
- description="Dataset using the v2 labels for LADI, pointing to the lower resolution source images for speed"
117
- ),
118
- LadiClassifyDatasetConfig(
119
- name="v2a",
120
- version=VERSION,
121
- description="Dataset using the v2a labels for LADI"
122
- ),
123
- LadiClassifyDatasetConfig(
124
- name="v2a_resized",
125
- version=VERSION,
126
- description="Dataset using the v2a labels for LADI, pointing to the lower resolution source images for speed"
127
- ),
128
- ]
129
-
130
- def _info(self):
131
- if self.config.label_name == "v1_damage":
132
- features = datasets.Features(
133
- {
134
- "image":datasets.Image(),
135
- "flood":datasets.Value("bool"),
136
- "rubble":datasets.Value("bool"),
137
- "misc_damage":datasets.Value("bool")
138
- }
139
- )
140
- elif self.config.label_name == "v1_infrastructure":
141
- features = datasets.Features(
142
- {
143
- "image":datasets.Image(),
144
- "building":datasets.Value("bool"),
145
- "road":datasets.Value("bool")
146
- }
147
- )
148
- elif self.config.label_name in ["v2", "v2_resized"]:
149
- features = datasets.Features(
150
- {
151
- "image":datasets.Image(),
152
- "bridges_any": datasets.Value("bool"),
153
- "bridges_damage": datasets.Value("bool"),
154
- "buildings_affected": datasets.Value("bool"),
155
- "buildings_any": datasets.Value("bool"),
156
- "buildings_destroyed": datasets.Value("bool"),
157
- "buildings_major": datasets.Value("bool"),
158
- "buildings_minor": datasets.Value("bool"),
159
- "debris_any": datasets.Value("bool"),
160
- "flooding_any": datasets.Value("bool"),
161
- "flooding_structures": datasets.Value("bool"),
162
- "roads_any": datasets.Value("bool"),
163
- "roads_damage": datasets.Value("bool"),
164
- "trees_any": datasets.Value("bool"),
165
- "trees_damage": datasets.Value("bool"),
166
- "water_any": datasets.Value("bool"),
167
- }
168
- )
169
- elif self.config.label_name in ["v2a", "v2a_resized"]:
170
- features = datasets.Features(
171
- {
172
- "image":datasets.Image(),
173
- "bridges_any": datasets.Value("bool"),
174
- "buildings_any": datasets.Value("bool"),
175
- "buildings_affected_or_greater": datasets.Value("bool"),
176
- "buildings_minor_or_greater": datasets.Value("bool"),
177
- "debris_any": datasets.Value("bool"),
178
- "flooding_any": datasets.Value("bool"),
179
- "flooding_structures": datasets.Value("bool"),
180
- "roads_any": datasets.Value("bool"),
181
- "roads_damage": datasets.Value("bool"),
182
- "trees_any": datasets.Value("bool"),
183
- "trees_damage": datasets.Value("bool"),
184
- "water_any": datasets.Value("bool"),
185
- }
186
- )
187
- else:
188
- raise NotImplementedError
189
- return datasets.DatasetInfo(
190
- # This is the description that will appear on the datasets page.
191
- description=f"LADI Dataset for {self.config.label_name} category",
192
- # This defines the different columns of the dataset and their types
193
- features=features, # Here we define them above because they are different between the two configurations
194
- # If there's a common (input, target) tuple from the features, uncomment supervised_keys line below and
195
- # specify them. They'll be used if as_supervised=True in builder.as_dataset.
196
- # supervised_keys=("image", "label"),
197
- )
198
-
199
- def read_ann_csv(self, fpath):
200
- if self.config.data_name == 'v1':
201
- return pd.read_csv(fpath, sep='\t', index_col=False)
202
- return pd.read_csv(fpath, sep=',', index_col=False)
203
-
204
- def _split_generators(self, dl_manager):
205
- generators = []
206
- data_files = self.config.split_csvs
207
-
208
- if self.config.download_ladi:
209
- # download data files to config.base_dir
210
- dl_url = dl_manager.download(DATA_URLS[self.config.data_name])
211
- base_dir = Path(self.config.base_dir)
212
- tar_iterator = dl_manager.iter_archive(dl_url)
213
- base_dir.mkdir(exist_ok=True)
214
- for filename, file in tar_iterator:
215
- file_path: Path = base_dir/filename
216
- file_path.parent.mkdir(parents=True, exist_ok=True)
217
- with open(base_dir/filename, 'wb') as f:
218
- f.write(file.read())
219
-
220
- data_files = DataFilesDict.from_local_or_remote(
221
- sanitize_patterns(data_files),
222
- base_path=self.config.base_dir
223
- )
224
-
225
- if 'train' in data_files.keys():
226
- train_df = self.read_ann_csv(data_files['train'][0])
227
- label_cols = tuple(label for label in train_df.columns if label not in ['url','local_path'])
228
- train_examples = [x._asdict() for x in train_df.itertuples()]
229
- generators.append(datasets.SplitGenerator(
230
- name=datasets.Split.TRAIN,
231
- gen_kwargs={"examples":train_examples,
232
- "label_cols":label_cols}
233
- ))
234
- if 'val' in data_files.keys():
235
- val_df = self.read_ann_csv(data_files['val'][0])
236
- label_cols = tuple(label for label in val_df.columns if label not in ['url','local_path'])
237
- val_examples = [x._asdict() for x in val_df.itertuples()]
238
- generators.append(datasets.SplitGenerator(
239
- name=datasets.Split.VALIDATION,
240
- gen_kwargs={"examples":val_examples,
241
- "label_cols":label_cols}
242
- ))
243
- if 'test' in data_files.keys():
244
- test_df = self.read_ann_csv(data_files['test'][0])
245
- label_cols = tuple(label for label in test_df.columns if label not in ['url','local_path'])
246
- test_examples = [x._asdict() for x in test_df.itertuples()]
247
- generators.append(datasets.SplitGenerator(
248
- name=datasets.Split.TEST,
249
- gen_kwargs={"examples":test_examples,
250
- "label_cols":label_cols}
251
- ))
252
- if 'all' in data_files.keys():
253
- all_df = self.read_ann_csv(data_files['all'][0])
254
- label_cols = tuple(label for label in all_df.columns if label not in ['url','local_path'])
255
- all_examples = [x._asdict() for x in all_df.itertuples()]
256
- generators.append(datasets.SplitGenerator(
257
- name=datasets.Split.ALL,
258
- gen_kwargs={"examples":all_examples,
259
- "label_cols":label_cols}
260
- ))
261
-
262
- return generators
263
-
264
- def _generate_examples(self, examples, label_cols, from_url_list=False):
265
- for ex in examples:
266
- try:
267
- image_path = Path(ex['local_path'])
268
- if not image_path.is_absolute():
269
- image_path = str(self.config.base_dir/image_path)
270
- except:
271
- print(ex)
272
- raise
273
-
274
- image = cv2.imread(image_path)
275
- image = cv2.cvtColor(image, cv2.COLOR_BGR2RGB)
276
- labels = {k:ex[k] for k in label_cols}
277
- labels |= {"image":image}
278
- yield image_path, labels
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
data/dataset_dict.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"splits": ["train", "validation", "test"]}
data/test/data-00000-of-00003.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c0f83ec3456a9d123b4a60e200182c546c00f9e1c2c794a789d860c7d9198c85
3
+ size 1018227848
data/test/data-00001-of-00003.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e59d182bea45f18edbc9eccb6e4cd336462c167d214fa50867ef9a64310b1592
3
+ size 1043617032
data/test/data-00002-of-00003.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:22541f9e5d9c3380ef9fa68ea4c14a33c38526b483618b35d5c62d6020fa0e6e
3
+ size 936030136
data/test/dataset_info.json ADDED
@@ -0,0 +1,218 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "builder_name": "ladi_classify_dataset",
3
+ "citation": "",
4
+ "config_name": "v2a_resized",
5
+ "dataset_name": "ladi_classify_dataset",
6
+ "dataset_size": 29147075809,
7
+ "description": "LADI Dataset for v2a_resized category",
8
+ "download_checksums": {},
9
+ "download_size": 0,
10
+ "features": {
11
+ "image": {
12
+ "_type": "Image"
13
+ },
14
+ "bridges_any": {
15
+ "dtype": "bool",
16
+ "_type": "Value"
17
+ },
18
+ "buildings_any": {
19
+ "dtype": "bool",
20
+ "_type": "Value"
21
+ },
22
+ "buildings_affected_or_greater": {
23
+ "dtype": "bool",
24
+ "_type": "Value"
25
+ },
26
+ "buildings_minor_or_greater": {
27
+ "dtype": "bool",
28
+ "_type": "Value"
29
+ },
30
+ "debris_any": {
31
+ "dtype": "bool",
32
+ "_type": "Value"
33
+ },
34
+ "flooding_any": {
35
+ "dtype": "bool",
36
+ "_type": "Value"
37
+ },
38
+ "flooding_structures": {
39
+ "dtype": "bool",
40
+ "_type": "Value"
41
+ },
42
+ "roads_any": {
43
+ "dtype": "bool",
44
+ "_type": "Value"
45
+ },
46
+ "roads_damage": {
47
+ "dtype": "bool",
48
+ "_type": "Value"
49
+ },
50
+ "trees_any": {
51
+ "dtype": "bool",
52
+ "_type": "Value"
53
+ },
54
+ "trees_damage": {
55
+ "dtype": "bool",
56
+ "_type": "Value"
57
+ },
58
+ "water_any": {
59
+ "dtype": "bool",
60
+ "_type": "Value"
61
+ }
62
+ },
63
+ "homepage": "",
64
+ "license": "",
65
+ "size_in_bytes": 29147075809,
66
+ "splits": {
67
+ "train": {
68
+ "name": "train",
69
+ "num_bytes": 23570974471,
70
+ "num_examples": 8030,
71
+ "shard_lengths": [
72
+ 201,
73
+ 201,
74
+ 201,
75
+ 201,
76
+ 201,
77
+ 201,
78
+ 201,
79
+ 201,
80
+ 201,
81
+ 201,
82
+ 201,
83
+ 201,
84
+ 201,
85
+ 201,
86
+ 201,
87
+ 201,
88
+ 201,
89
+ 201,
90
+ 201,
91
+ 201,
92
+ 201,
93
+ 201,
94
+ 201,
95
+ 201,
96
+ 201,
97
+ 201,
98
+ 201,
99
+ 201,
100
+ 201,
101
+ 201,
102
+ 200,
103
+ 200,
104
+ 200,
105
+ 200,
106
+ 200,
107
+ 200,
108
+ 200,
109
+ 200,
110
+ 200,
111
+ 200
112
+ ],
113
+ "dataset_name": "ladi_classify_dataset"
114
+ },
115
+ "validation": {
116
+ "name": "validation",
117
+ "num_bytes": 2578233555,
118
+ "num_examples": 893,
119
+ "shard_lengths": [
120
+ 23,
121
+ 23,
122
+ 23,
123
+ 23,
124
+ 23,
125
+ 23,
126
+ 23,
127
+ 23,
128
+ 23,
129
+ 23,
130
+ 23,
131
+ 23,
132
+ 23,
133
+ 22,
134
+ 22,
135
+ 22,
136
+ 22,
137
+ 22,
138
+ 22,
139
+ 22,
140
+ 22,
141
+ 22,
142
+ 22,
143
+ 22,
144
+ 22,
145
+ 22,
146
+ 22,
147
+ 22,
148
+ 22,
149
+ 22,
150
+ 22,
151
+ 22,
152
+ 22,
153
+ 22,
154
+ 22,
155
+ 22,
156
+ 22,
157
+ 22,
158
+ 22,
159
+ 22
160
+ ],
161
+ "dataset_name": "ladi_classify_dataset"
162
+ },
163
+ "test": {
164
+ "name": "test",
165
+ "num_bytes": 2997867783,
166
+ "num_examples": 1049,
167
+ "shard_lengths": [
168
+ 27,
169
+ 27,
170
+ 27,
171
+ 27,
172
+ 27,
173
+ 27,
174
+ 27,
175
+ 27,
176
+ 27,
177
+ 26,
178
+ 26,
179
+ 26,
180
+ 26,
181
+ 26,
182
+ 26,
183
+ 26,
184
+ 26,
185
+ 26,
186
+ 26,
187
+ 26,
188
+ 26,
189
+ 26,
190
+ 26,
191
+ 26,
192
+ 26,
193
+ 26,
194
+ 26,
195
+ 26,
196
+ 26,
197
+ 26,
198
+ 26,
199
+ 26,
200
+ 26,
201
+ 26,
202
+ 26,
203
+ 26,
204
+ 26,
205
+ 26,
206
+ 26,
207
+ 26
208
+ ],
209
+ "dataset_name": "ladi_classify_dataset"
210
+ }
211
+ },
212
+ "version": {
213
+ "version_str": "0.2.1",
214
+ "major": 0,
215
+ "minor": 2,
216
+ "patch": 1
217
+ }
218
+ }
data/test/state.json ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_data_files": [
3
+ {
4
+ "filename": "data-00000-of-00003.arrow"
5
+ },
6
+ {
7
+ "filename": "data-00001-of-00003.arrow"
8
+ },
9
+ {
10
+ "filename": "data-00002-of-00003.arrow"
11
+ }
12
+ ],
13
+ "_fingerprint": "a65e74c792765ab3",
14
+ "_format_columns": null,
15
+ "_format_kwargs": {},
16
+ "_format_type": null,
17
+ "_output_all_columns": false,
18
+ "_split": "test"
19
+ }
data/train/data-00000-of-00024.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0b50b08e1660775466262c052e3a3c588dc787bf9ab321caf0e468eac58a5afe
3
+ size 1010770856
data/train/data-00001-of-00024.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d352b8d983cefaf89c8c34cc6c3a338305747c7728e84108c819fe61fc275482
3
+ size 978774496
data/train/data-00002-of-00024.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ef147afd8d7200344de7f5acec4bd1c02e4b1e276fa72a3e9d3cc5e3aca6872e
3
+ size 1000394072
data/train/data-00003-of-00024.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7e291b17bcbe5c54e442f38d78cf79f3fddcf89bc9f9e377c479d2fce8b2355d
3
+ size 954224648
data/train/data-00004-of-00024.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:234ad4400dc59664a49b02ecad84c552e5dd38775fc76277a58112e2d2d26a16
3
+ size 980035328
data/train/data-00005-of-00024.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:731a88fb783a6f0a9d5f09033773af8541fa5e7d47e4475cf6d9b229cdb84842
3
+ size 965259904
data/train/data-00006-of-00024.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4475ab3b6dd6cfd795541132a10dc7da3a2b1b51732816a4370fbc4f570c5048
3
+ size 977305256
data/train/data-00007-of-00024.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b525bcb2de0c875c2d32594099185c36a5e2c853a5b0b136ec70076f74240173
3
+ size 956689776
data/train/data-00008-of-00024.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ceff0452520177e81cf51824a028bf53314216d913fe47d965c0fd0e054acb29
3
+ size 1010461320
data/train/data-00009-of-00024.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:16c3cbb8fc23e8eb10f0aaad51b945c1a1a4a7afea48ad15e8a018abd9b25bbd
3
+ size 1004955368
data/train/data-00010-of-00024.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:748f9653c36c588a68254bdfeff26cca1f537c9a774ba30ea7d7432a97cbd0f5
3
+ size 985306320
data/train/data-00011-of-00024.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fd89b484c89cd6ddfaa9fe1f3d67cc73331d223d2db411220386b0d9d4e3bbb8
3
+ size 1000287816
data/train/data-00012-of-00024.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e54840b0cc5f46a7524bfa65b3280ea4954fd4609c417bcaee435061d0654fa0
3
+ size 974323616
data/train/data-00013-of-00024.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:811fbff8541c37807f1f0583f331c64c5eaa93a69d0d1b8d95666aed74d44104
3
+ size 964532376
data/train/data-00014-of-00024.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:57421dacd0f067a7aca5d1405b2447063a0a63d3764a5acc1ed5ddc410c6e6d8
3
+ size 989601136
data/train/data-00015-of-00024.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b889411400b09d5a8f127ad8683675cbaee16eb957e1a99751b8c681b8842f9d
3
+ size 988367712
data/train/data-00016-of-00024.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5989c7c928bce156c8b61dba1be009434929b87ed4c012981621faea1d0102bd
3
+ size 978024936
data/train/data-00017-of-00024.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:331dd9132a8dd0a12f9071fd08405aa334013382d226396965fea13c97455252
3
+ size 1005145800
data/train/data-00018-of-00024.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d2748cf6328556bab6de16ca058eb89f9629a7aa903d0b6bc09b2ad43d3af00a
3
+ size 985704688
data/train/data-00019-of-00024.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d50dae2f242e60666caca4b5fb80ee695be69d62131701a15768b7eb674511eb
3
+ size 959905608
data/train/data-00020-of-00024.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:570ba7b379897c0f841cde3e0573e39992253bed01402aec5bf4bece7a75b4e4
3
+ size 976951840
data/train/data-00021-of-00024.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:98cbf47fbde0d5067ddd6bf4fee0c9385f9f52e9aa164f41e1c88018cd8ce3b0
3
+ size 969511552
data/train/data-00022-of-00024.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7a4a38c7bbc49afb39c60230ffddac612d13c29daf83b45c6567c5c3ada54046
3
+ size 988818736
data/train/data-00023-of-00024.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:309d63d4c52d3d3561d87ebf944f16e44d42f30fa621426f868a0d1f7992a11b
3
+ size 965682392
data/train/dataset_info.json ADDED
@@ -0,0 +1,218 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "builder_name": "ladi_classify_dataset",
3
+ "citation": "",
4
+ "config_name": "v2a_resized",
5
+ "dataset_name": "ladi_classify_dataset",
6
+ "dataset_size": 29147075809,
7
+ "description": "LADI Dataset for v2a_resized category",
8
+ "download_checksums": {},
9
+ "download_size": 0,
10
+ "features": {
11
+ "image": {
12
+ "_type": "Image"
13
+ },
14
+ "bridges_any": {
15
+ "dtype": "bool",
16
+ "_type": "Value"
17
+ },
18
+ "buildings_any": {
19
+ "dtype": "bool",
20
+ "_type": "Value"
21
+ },
22
+ "buildings_affected_or_greater": {
23
+ "dtype": "bool",
24
+ "_type": "Value"
25
+ },
26
+ "buildings_minor_or_greater": {
27
+ "dtype": "bool",
28
+ "_type": "Value"
29
+ },
30
+ "debris_any": {
31
+ "dtype": "bool",
32
+ "_type": "Value"
33
+ },
34
+ "flooding_any": {
35
+ "dtype": "bool",
36
+ "_type": "Value"
37
+ },
38
+ "flooding_structures": {
39
+ "dtype": "bool",
40
+ "_type": "Value"
41
+ },
42
+ "roads_any": {
43
+ "dtype": "bool",
44
+ "_type": "Value"
45
+ },
46
+ "roads_damage": {
47
+ "dtype": "bool",
48
+ "_type": "Value"
49
+ },
50
+ "trees_any": {
51
+ "dtype": "bool",
52
+ "_type": "Value"
53
+ },
54
+ "trees_damage": {
55
+ "dtype": "bool",
56
+ "_type": "Value"
57
+ },
58
+ "water_any": {
59
+ "dtype": "bool",
60
+ "_type": "Value"
61
+ }
62
+ },
63
+ "homepage": "",
64
+ "license": "",
65
+ "size_in_bytes": 29147075809,
66
+ "splits": {
67
+ "train": {
68
+ "name": "train",
69
+ "num_bytes": 23570974471,
70
+ "num_examples": 8030,
71
+ "shard_lengths": [
72
+ 201,
73
+ 201,
74
+ 201,
75
+ 201,
76
+ 201,
77
+ 201,
78
+ 201,
79
+ 201,
80
+ 201,
81
+ 201,
82
+ 201,
83
+ 201,
84
+ 201,
85
+ 201,
86
+ 201,
87
+ 201,
88
+ 201,
89
+ 201,
90
+ 201,
91
+ 201,
92
+ 201,
93
+ 201,
94
+ 201,
95
+ 201,
96
+ 201,
97
+ 201,
98
+ 201,
99
+ 201,
100
+ 201,
101
+ 201,
102
+ 200,
103
+ 200,
104
+ 200,
105
+ 200,
106
+ 200,
107
+ 200,
108
+ 200,
109
+ 200,
110
+ 200,
111
+ 200
112
+ ],
113
+ "dataset_name": "ladi_classify_dataset"
114
+ },
115
+ "validation": {
116
+ "name": "validation",
117
+ "num_bytes": 2578233555,
118
+ "num_examples": 893,
119
+ "shard_lengths": [
120
+ 23,
121
+ 23,
122
+ 23,
123
+ 23,
124
+ 23,
125
+ 23,
126
+ 23,
127
+ 23,
128
+ 23,
129
+ 23,
130
+ 23,
131
+ 23,
132
+ 23,
133
+ 22,
134
+ 22,
135
+ 22,
136
+ 22,
137
+ 22,
138
+ 22,
139
+ 22,
140
+ 22,
141
+ 22,
142
+ 22,
143
+ 22,
144
+ 22,
145
+ 22,
146
+ 22,
147
+ 22,
148
+ 22,
149
+ 22,
150
+ 22,
151
+ 22,
152
+ 22,
153
+ 22,
154
+ 22,
155
+ 22,
156
+ 22,
157
+ 22,
158
+ 22,
159
+ 22
160
+ ],
161
+ "dataset_name": "ladi_classify_dataset"
162
+ },
163
+ "test": {
164
+ "name": "test",
165
+ "num_bytes": 2997867783,
166
+ "num_examples": 1049,
167
+ "shard_lengths": [
168
+ 27,
169
+ 27,
170
+ 27,
171
+ 27,
172
+ 27,
173
+ 27,
174
+ 27,
175
+ 27,
176
+ 27,
177
+ 26,
178
+ 26,
179
+ 26,
180
+ 26,
181
+ 26,
182
+ 26,
183
+ 26,
184
+ 26,
185
+ 26,
186
+ 26,
187
+ 26,
188
+ 26,
189
+ 26,
190
+ 26,
191
+ 26,
192
+ 26,
193
+ 26,
194
+ 26,
195
+ 26,
196
+ 26,
197
+ 26,
198
+ 26,
199
+ 26,
200
+ 26,
201
+ 26,
202
+ 26,
203
+ 26,
204
+ 26,
205
+ 26,
206
+ 26,
207
+ 26
208
+ ],
209
+ "dataset_name": "ladi_classify_dataset"
210
+ }
211
+ },
212
+ "version": {
213
+ "version_str": "0.2.1",
214
+ "major": 0,
215
+ "minor": 2,
216
+ "patch": 1
217
+ }
218
+ }
data/train/state.json ADDED
@@ -0,0 +1,82 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_data_files": [
3
+ {
4
+ "filename": "data-00000-of-00024.arrow"
5
+ },
6
+ {
7
+ "filename": "data-00001-of-00024.arrow"
8
+ },
9
+ {
10
+ "filename": "data-00002-of-00024.arrow"
11
+ },
12
+ {
13
+ "filename": "data-00003-of-00024.arrow"
14
+ },
15
+ {
16
+ "filename": "data-00004-of-00024.arrow"
17
+ },
18
+ {
19
+ "filename": "data-00005-of-00024.arrow"
20
+ },
21
+ {
22
+ "filename": "data-00006-of-00024.arrow"
23
+ },
24
+ {
25
+ "filename": "data-00007-of-00024.arrow"
26
+ },
27
+ {
28
+ "filename": "data-00008-of-00024.arrow"
29
+ },
30
+ {
31
+ "filename": "data-00009-of-00024.arrow"
32
+ },
33
+ {
34
+ "filename": "data-00010-of-00024.arrow"
35
+ },
36
+ {
37
+ "filename": "data-00011-of-00024.arrow"
38
+ },
39
+ {
40
+ "filename": "data-00012-of-00024.arrow"
41
+ },
42
+ {
43
+ "filename": "data-00013-of-00024.arrow"
44
+ },
45
+ {
46
+ "filename": "data-00014-of-00024.arrow"
47
+ },
48
+ {
49
+ "filename": "data-00015-of-00024.arrow"
50
+ },
51
+ {
52
+ "filename": "data-00016-of-00024.arrow"
53
+ },
54
+ {
55
+ "filename": "data-00017-of-00024.arrow"
56
+ },
57
+ {
58
+ "filename": "data-00018-of-00024.arrow"
59
+ },
60
+ {
61
+ "filename": "data-00019-of-00024.arrow"
62
+ },
63
+ {
64
+ "filename": "data-00020-of-00024.arrow"
65
+ },
66
+ {
67
+ "filename": "data-00021-of-00024.arrow"
68
+ },
69
+ {
70
+ "filename": "data-00022-of-00024.arrow"
71
+ },
72
+ {
73
+ "filename": "data-00023-of-00024.arrow"
74
+ }
75
+ ],
76
+ "_fingerprint": "81110feabea25c42",
77
+ "_format_columns": null,
78
+ "_format_kwargs": {},
79
+ "_format_type": null,
80
+ "_output_all_columns": false,
81
+ "_split": "train"
82
+ }
data/validation/data-00000-of-00003.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:feb28475d5353a2ee011ff782f94ce37cfe0d99ec7ee01adb97c1f17d2059f84
3
+ size 862885656
data/validation/data-00001-of-00003.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b87a6566c06e79feccfd04de30a9038c0be5f7f19c417b5e3c1531759fa65cb2
3
+ size 876671784
data/validation/data-00002-of-00003.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5ae6c6516ba3e20b94bb6fa9018d15907617c7399784cece23b8c22d6e65fa05
3
+ size 838683560
data/validation/dataset_info.json ADDED
@@ -0,0 +1,218 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "builder_name": "ladi_classify_dataset",
3
+ "citation": "",
4
+ "config_name": "v2a_resized",
5
+ "dataset_name": "ladi_classify_dataset",
6
+ "dataset_size": 29147075809,
7
+ "description": "LADI Dataset for v2a_resized category",
8
+ "download_checksums": {},
9
+ "download_size": 0,
10
+ "features": {
11
+ "image": {
12
+ "_type": "Image"
13
+ },
14
+ "bridges_any": {
15
+ "dtype": "bool",
16
+ "_type": "Value"
17
+ },
18
+ "buildings_any": {
19
+ "dtype": "bool",
20
+ "_type": "Value"
21
+ },
22
+ "buildings_affected_or_greater": {
23
+ "dtype": "bool",
24
+ "_type": "Value"
25
+ },
26
+ "buildings_minor_or_greater": {
27
+ "dtype": "bool",
28
+ "_type": "Value"
29
+ },
30
+ "debris_any": {
31
+ "dtype": "bool",
32
+ "_type": "Value"
33
+ },
34
+ "flooding_any": {
35
+ "dtype": "bool",
36
+ "_type": "Value"
37
+ },
38
+ "flooding_structures": {
39
+ "dtype": "bool",
40
+ "_type": "Value"
41
+ },
42
+ "roads_any": {
43
+ "dtype": "bool",
44
+ "_type": "Value"
45
+ },
46
+ "roads_damage": {
47
+ "dtype": "bool",
48
+ "_type": "Value"
49
+ },
50
+ "trees_any": {
51
+ "dtype": "bool",
52
+ "_type": "Value"
53
+ },
54
+ "trees_damage": {
55
+ "dtype": "bool",
56
+ "_type": "Value"
57
+ },
58
+ "water_any": {
59
+ "dtype": "bool",
60
+ "_type": "Value"
61
+ }
62
+ },
63
+ "homepage": "",
64
+ "license": "",
65
+ "size_in_bytes": 29147075809,
66
+ "splits": {
67
+ "train": {
68
+ "name": "train",
69
+ "num_bytes": 23570974471,
70
+ "num_examples": 8030,
71
+ "shard_lengths": [
72
+ 201,
73
+ 201,
74
+ 201,
75
+ 201,
76
+ 201,
77
+ 201,
78
+ 201,
79
+ 201,
80
+ 201,
81
+ 201,
82
+ 201,
83
+ 201,
84
+ 201,
85
+ 201,
86
+ 201,
87
+ 201,
88
+ 201,
89
+ 201,
90
+ 201,
91
+ 201,
92
+ 201,
93
+ 201,
94
+ 201,
95
+ 201,
96
+ 201,
97
+ 201,
98
+ 201,
99
+ 201,
100
+ 201,
101
+ 201,
102
+ 200,
103
+ 200,
104
+ 200,
105
+ 200,
106
+ 200,
107
+ 200,
108
+ 200,
109
+ 200,
110
+ 200,
111
+ 200
112
+ ],
113
+ "dataset_name": "ladi_classify_dataset"
114
+ },
115
+ "validation": {
116
+ "name": "validation",
117
+ "num_bytes": 2578233555,
118
+ "num_examples": 893,
119
+ "shard_lengths": [
120
+ 23,
121
+ 23,
122
+ 23,
123
+ 23,
124
+ 23,
125
+ 23,
126
+ 23,
127
+ 23,
128
+ 23,
129
+ 23,
130
+ 23,
131
+ 23,
132
+ 23,
133
+ 22,
134
+ 22,
135
+ 22,
136
+ 22,
137
+ 22,
138
+ 22,
139
+ 22,
140
+ 22,
141
+ 22,
142
+ 22,
143
+ 22,
144
+ 22,
145
+ 22,
146
+ 22,
147
+ 22,
148
+ 22,
149
+ 22,
150
+ 22,
151
+ 22,
152
+ 22,
153
+ 22,
154
+ 22,
155
+ 22,
156
+ 22,
157
+ 22,
158
+ 22,
159
+ 22
160
+ ],
161
+ "dataset_name": "ladi_classify_dataset"
162
+ },
163
+ "test": {
164
+ "name": "test",
165
+ "num_bytes": 2997867783,
166
+ "num_examples": 1049,
167
+ "shard_lengths": [
168
+ 27,
169
+ 27,
170
+ 27,
171
+ 27,
172
+ 27,
173
+ 27,
174
+ 27,
175
+ 27,
176
+ 27,
177
+ 26,
178
+ 26,
179
+ 26,
180
+ 26,
181
+ 26,
182
+ 26,
183
+ 26,
184
+ 26,
185
+ 26,
186
+ 26,
187
+ 26,
188
+ 26,
189
+ 26,
190
+ 26,
191
+ 26,
192
+ 26,
193
+ 26,
194
+ 26,
195
+ 26,
196
+ 26,
197
+ 26,
198
+ 26,
199
+ 26,
200
+ 26,
201
+ 26,
202
+ 26,
203
+ 26,
204
+ 26,
205
+ 26,
206
+ 26,
207
+ 26
208
+ ],
209
+ "dataset_name": "ladi_classify_dataset"
210
+ }
211
+ },
212
+ "version": {
213
+ "version_str": "0.2.1",
214
+ "major": 0,
215
+ "minor": 2,
216
+ "patch": 1
217
+ }
218
+ }
data/validation/state.json ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_data_files": [
3
+ {
4
+ "filename": "data-00000-of-00003.arrow"
5
+ },
6
+ {
7
+ "filename": "data-00001-of-00003.arrow"
8
+ },
9
+ {
10
+ "filename": "data-00002-of-00003.arrow"
11
+ }
12
+ ],
13
+ "_fingerprint": "9b4826f7e4eacc7d",
14
+ "_format_columns": null,
15
+ "_format_kwargs": {},
16
+ "_format_type": null,
17
+ "_output_all_columns": false,
18
+ "_split": "validation"
19
+ }