File size: 2,785 Bytes
09a1f13
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
ec73097
09a1f13
 
c1658b6
09a1f13
daa2f84
 
 
ac8d193
daa2f84
09a1f13
198cc84
09a1f13
198cc84
09a1f13
 
 
 
 
a597d45
7cf61d8
 
a597d45
 
 
 
034c476
 
 
 
09a1f13
 
 
 
 
 
 
 
daa2f84
ec73097
daa2f84
 
09a1f13
 
ec73097
09a1f13
 
0123b7a
09a1f13
4cbd8f6
ac8d193
daa2f84
 
09a1f13
 
daa2f84
a597d45
7cf61d8
09a1f13
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
# coding=utf-8
# Copyright 2022 the HuggingFace Datasets Authors.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
#     http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.

import os
import pandas as pd 
import datasets
import json
from huggingface_hub import hf_hub_url

_INPUT_CSV = "test_1k_flickr.csv"
_INPUT_IMAGES = "images_flickr_1k_test"
_REPO_ID = "yonatanbitton/flickr_1k_test_image_text_retrieval"
_JSON_KEYS = ['sentids', 'imgid', 'tokens', 'raw', 'sentid']

class Dataset(datasets.GeneratorBasedBuilder):
    VERSION = datasets.Version("1.1.0")
    BUILDER_CONFIGS = [
        datasets.BuilderConfig(name="TEST", version=VERSION, description="test"),
    ]

    def _info(self):
        return datasets.DatasetInfo(
            features=datasets.Features(
                 {
                "image": datasets.Image(),
                "caption": [datasets.Value('string')],
                "sentids": [datasets.Value("string")],
                "imgid": datasets.Value("string"),
                "split": datasets.Value("string"),
                "filename": datasets.Value("string"),
                'sentids': [datasets.Value('int64')],
                'imgid': [datasets.Value('int64')],
                'tokens': [[datasets.Value('string')]],
                'sentid': [datasets.Value('int64')], 
                }
            ),
            task_templates=[],
        )

    def _split_generators(self, dl_manager):
        """Returns SplitGenerators."""

        repo_id = _REPO_ID
        data_dir = dl_manager.download_and_extract({
            "examples_csv": hf_hub_url(repo_id=repo_id, repo_type='dataset', filename=_INPUT_CSV),
            "images_dir": hf_hub_url(repo_id=repo_id, repo_type='dataset', filename=f"{_INPUT_IMAGES}.zip")
        })

        return [datasets.SplitGenerator(name=datasets.Split.TEST, gen_kwargs=data_dir)]


    def _generate_examples(self, examples_csv, images_dir):
        """Yields examples."""
        df = pd.read_csv(examples_csv)
        for c in _JSON_KEYS:
            df[c] = df[c].apply(json.loads)

        for r_idx, r in df.iterrows():
            r_dict = r.to_dict()
            image_path = os.path.join(images_dir, _INPUT_IMAGES, r_dict['filename'])
            r_dict['image'] = image_path
            r_dict['caption'] = r_dict.pop('raw')
            yield r_idx, r_dict