williamberman commited on
Commit
34696b9
1 Parent(s): c9e5f40
Files changed (4) hide show
  1. conditioning_images.zip +3 -0
  2. fill50k.py +101 -0
  3. images.zip +3 -0
  4. train.jsonl +0 -0
conditioning_images.zip ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8361c0cf0834d7e3ad9bb6ecb9b235c2e7e3d8e3bd775cd3b87171ac54124b85
3
+ size 110247796
fill50k.py ADDED
@@ -0,0 +1,101 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import pandas as pd
2
+ from huggingface_hub import hf_hub_url
3
+ import datasets
4
+ import os
5
+
6
+ _VERSION = datasets.Version("0.0.1")
7
+
8
+ _DESCRIPTION = "TODO"
9
+ _HOMEPAGE = "TODO"
10
+ _LICENSE = "TODO"
11
+ _CITATION = "TODO"
12
+
13
+ _FEATURES = datasets.Features(
14
+ {
15
+ "image": datasets.Image(),
16
+ "conditioning_image": datasets.Image(),
17
+ "text": datasets.Value("string"),
18
+ },
19
+ )
20
+
21
+ METADATA_URL = hf_hub_url(
22
+ "williamberman/fill50k",
23
+ filename="train.jsonl",
24
+ repo_type="dataset",
25
+ )
26
+
27
+ IMAGES_URL = hf_hub_url(
28
+ "williamberman/fill50k",
29
+ filename="images.zip",
30
+ repo_type="dataset",
31
+ )
32
+
33
+ CONDITIONING_IMAGES_URL = hf_hub_url(
34
+ "williamberman/fill50k",
35
+ filename="conditioning_images.zip",
36
+ repo_type="dataset",
37
+ )
38
+
39
+ _DEFAULT_CONFIG = datasets.BuilderConfig(name="default", version=_VERSION)
40
+
41
+
42
+ class Fill50k(datasets.GeneratorBasedBuilder):
43
+ BUILDER_CONFIGS = [_DEFAULT_CONFIG]
44
+ DEFAULT_CONFIG_NAME = "default"
45
+
46
+ def _info(self):
47
+ return datasets.DatasetInfo(
48
+ description=_DESCRIPTION,
49
+ features=_FEATURES,
50
+ supervised_keys=None,
51
+ homepage=_HOMEPAGE,
52
+ license=_LICENSE,
53
+ citation=_CITATION,
54
+ )
55
+
56
+ def _split_generators(self, dl_manager):
57
+ metadata_path = dl_manager.download(METADATA_URL)
58
+ images_dir = dl_manager.download_and_extract(IMAGES_URL)
59
+ conditioning_images_dir = dl_manager.download_and_extract(
60
+ CONDITIONING_IMAGES_URL
61
+ )
62
+
63
+ return [
64
+ datasets.SplitGenerator(
65
+ name=datasets.Split.TRAIN,
66
+ # These kwargs will be passed to _generate_examples
67
+ gen_kwargs={
68
+ "metadata_path": metadata_path,
69
+ "images_dir": images_dir,
70
+ "conditioning_images_dir": conditioning_images_dir,
71
+ },
72
+ ),
73
+ ]
74
+
75
+ def _generate_examples(self, metadata_path, images_dir, conditioning_images_dir):
76
+ metadata = pd.read_json(metadata_path, lines=True)
77
+
78
+ for _, row in metadata.iterrows():
79
+ text = row["text"]
80
+
81
+ image_path = row["image"]
82
+ image_path = os.path.join(images_dir, image_path)
83
+ image = open(image_path, "rb").read()
84
+
85
+ conditioning_image_path = row["conditioning_image"]
86
+ conditioning_image_path = os.path.join(
87
+ conditioning_images_dir, row["conditioning_image"]
88
+ )
89
+ conditioning_image = open(conditioning_image_path, "rb").read()
90
+
91
+ yield row["image"], {
92
+ "text": text,
93
+ "image": {
94
+ "path": image_path,
95
+ "bytes": image,
96
+ },
97
+ "conditioning_image": {
98
+ "path": conditioning_image_path,
99
+ "bytes": conditioning_image,
100
+ },
101
+ }
images.zip ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:38d0314c701a0ead27934d168b3ce7593c336b3b6c0b176ba34611e1b1e63fa9
3
+ size 121969176
train.jsonl ADDED
The diff for this file is too large to render. See raw diff