Datasets:

Modalities:
Tabular
Text
Formats:
csv
Languages:
English
ArXiv:
Libraries:
Datasets
pandas
License:
CREHate / finetune /dataset.py
Nayeon Lee
Upload 35 files
f1b5d0c verified
raw
history blame
No virus
3.14 kB
# Standard
import os
import sys
import re
import random
# PIP
import pandas as pd
import numpy as np
import torch
from torch.utils.data import Dataset, DataLoader
import pytorch_lightning as pl
from transformers import AutoTokenizer, T5Tokenizer
# Custom
current = os.path.dirname(os.path.realpath(__file__))
parent = os.path.dirname(current)
sys.path.append(parent)
import utils
class CSV2Dataset(Dataset):
def __init__(self, cfg, filename, option):
self.cfg = cfg
self.filename = filename
self.option = option
self.df = self.get_df()
if self.option == 'train':
self.df = self.df.sample(frac=1)
if 't5' in self.cfg.model:
self.tokenizer = T5Tokenizer.from_pretrained(self.cfg.model)
else:
self.tokenizer = AutoTokenizer.from_pretrained(self.cfg.model)
if len(self.cfg.additional_tokens) > 0:
num_added_toks = self.tokenizer.add_tokens(self.cfg.additional_tokens)
self.xs = self.preprocess()
if self.option != 'test':
self.ys = self.get_ys()
def get_df(self):
return pd.read_csv(self.filename,encoding='utf-8')
def get_tokenizer(self):
return self.tokenizer
def preprocess(self):
if self.option == 'test':
xs = self.df[self.cfg.test_col]
else:
xs = self.df[self.cfg.sent_col]
if self.cfg.remove_special_tokens:
xs = utils.remove_special_tokens(xs)
if self.cfg.if_arabic:
arabert_prep = ArabertPreprocessor(model_name=self.cfg.model)
xs = [arabert_prep.preprocess(x) for x in xs]
encoding = self.tokenizer(
list(xs),
padding='max_length',
max_length=self.cfg.max_length,
truncation=True,
return_tensors='pt'
)
xs = [(input_ids,attention) for input_ids,attention in zip(encoding.input_ids,encoding.attention_mask)]
return xs
def get_ys(self):
if 't5' in self.cfg.model:
encoding = self.tokenizer(
[str(label) for label in self.df[self.cfg.label_col]],
padding='max_length',
max_length=self.cfg.max_length,
truncation=True,
return_tensors='pt'
)
lm_labels = encoding.input_ids
lm_labels[lm_labels[:, :] == self.tokenizer.pad_token_id] = -100
ys = [(input_ids,attention) for input_ids,attention in zip(lm_labels,encoding.attention_mask)]
return ys
return [int(y) for y in self.df[self.cfg.label_col].tolist()]
def __getitem__(self, idx):
if self.option == 'test':
return self.xs[idx]
else:
return self.xs[idx], self.ys[idx]
def __len__(self):
return len(self.xs)
def batch_sampling(batch_size,data_len,is_test=False):
seq_lens = range(data_len)
sample_indices = [ seq_lens[i:i+batch_size] for i in range(0,data_len, batch_size)]
if not is_test:
random.shuffle(sample_indices)
return sample_indices