|
import os |
|
import pandas as pd |
|
import datasets |
|
from glob import glob |
|
import zipfile |
|
import json |
|
class dummy(datasets.GeneratorBasedBuilder): |
|
def _info(self): |
|
return datasets.DatasetInfo(features=datasets.Features({'Name':datasets.Value('string'),'Age':datasets.Value('string'),'label': datasets.features.ClassLabel(names=['female', 'male'])})) |
|
|
|
def extract_all(self, dir): |
|
zip_files = glob(dir+'/**/**.zip', recursive=True) |
|
for file in zip_files: |
|
with zipfile.ZipFile(file) as item: |
|
item.extractall('/'.join(file.split('/')[:-1])) |
|
|
|
|
|
def get_all_files(self, dir): |
|
files = [] |
|
valid_file_ext = ['txt', 'csv', 'tsv', 'xlsx', 'xls', 'xml', 'json', 'jsonl', 'html', 'wav', 'mp3', 'jpg', 'png'] |
|
for ext in valid_file_ext: |
|
files += glob(f"{dir}/**/**.{ext}", recursive = True) |
|
return files |
|
|
|
def _split_generators(self, dl_manager): |
|
url = [os.path.abspath(os.path.expanduser(dl_manager.manual_dir))] |
|
downloaded_files = url |
|
return [datasets.SplitGenerator(name=datasets.Split.TRAIN, gen_kwargs={'filepaths': {'inputs':downloaded_files} })] |
|
|
|
|
|
def read_json(self, filepath, json_key, lines = False): |
|
if json_key: |
|
data = json.load(open(filepath)) |
|
df = pd.DataFrame(data[json_key]) |
|
else: |
|
df = pd.read_json(filepath, lines=lines) |
|
return df |
|
|
|
def _generate_examples(self, filepaths): |
|
_id = 0 |
|
for i,filepath in enumerate(filepaths['inputs']): |
|
df = self.read_json(filepath, lines=True, json_key='data') |
|
if len(df.columns) != 3: |
|
continue |
|
df.columns = ['Name', 'Age', 'Gender'] |
|
for _, record in df.iterrows(): |
|
yield str(_id), {'Name':record['Name'],'Age':record['Age'],'label':str(record['Gender'])} |
|
_id += 1 |
|
|
|
|