Datasets:

Modalities:
Tabular
Text
Formats:
csv
Languages:
English
ArXiv:
Libraries:
Datasets
pandas
License:
CREHate / finetune /multi_task.py
Nayeon Lee
Upload 35 files
f1b5d0c verified
raw
history blame
No virus
11.9 kB
import warnings
warnings.simplefilter('ignore')
import numpy as np
import pandas as pd
from tqdm import tqdm
from sklearn import metrics
import transformers
import torch
from torch import nn
from torch.utils.data import Dataset, DataLoader, RandomSampler, SequentialSampler
from transformers import DistilBertTokenizer, DistilBertModel,AutoModel,AutoTokenizer,AutoConfig,AutoModelForSequenceClassification
import logging
logging.basicConfig(level=logging.ERROR)
import os
from itertools import permutations
from torch import cuda
device = 'cuda' if cuda.is_available() else 'cpu'
print(device)
models = ['vinai/bertweet-base',
'./hate_bert',
'Twitter/TwHIN-BERT-base',
'cardiffnlp/twitter-roberta-base',
'Xuhui/ToxDect-roberta-large',
'bert-base-cased',
'roberta-base']
model_names = [
'BERTweet',
'HateBERT',
'TwHIN-BERT',
'Twitter-RoBERTa',
'ToxDect-RoBERTa',
'BERT',
'RoBERTa'
]
countries = ['United States','Australia','United Kingdom','South Africa','Singapore']
codes = ['US', 'AU', 'GB', 'ZA', 'SG']
_hate_cols = [f'{country.replace(" ","_")}_Hate' for country in countries]
def hamming_score(y_true, y_pred, normalize=True, sample_weight=None):
acc_list = []
for i in range(y_true.shape[0]):
set_true = set( np.where(y_true[i])[0] )
set_pred = set( np.where(y_pred[i])[0] )
tmp_a = None
if len(set_true) == 0 and len(set_pred) == 0:
tmp_a = 1
else:
tmp_a = len(set_true.intersection(set_pred))/\
float( len(set_true.union(set_pred)) )
acc_list.append(tmp_a)
return np.mean(acc_list)
class MultiTaskDataset(Dataset):
def __init__(self, dataframe, tokenizer, max_len):
self.tokenizer = tokenizer
self.data = dataframe
self.text = dataframe.text
self.targets = self.data.labels
self.max_len = max_len
def __len__(self):
return len(self.text)
def __getitem__(self, index):
text = str(self.text[index])
inputs = self.tokenizer.encode_plus(
text,
None,
truncation=True,
add_special_tokens=True,
max_length=self.max_len,
pad_to_max_length=True,
return_token_type_ids=True
)
ids = inputs['input_ids']
mask = inputs['attention_mask']
token_type_ids = inputs["token_type_ids"]
return {
'ids': torch.tensor(ids, dtype=torch.long),
'mask': torch.tensor(mask, dtype=torch.long),
'token_type_ids': torch.tensor(token_type_ids, dtype=torch.long),
'targets': torch.tensor(self.targets[index], dtype=torch.long)
}
class Classifier(torch.nn.Module):
def __init__(self,model_name,tokenizer):
super(Classifier, self).__init__()
self.l1 = AutoModel.from_pretrained(model_name)
self.l1.resize_token_embeddings(len(tokenizer))
config = AutoConfig.from_pretrained(model_name)
self.pre_classifier = torch.nn.Linear(config.hidden_size, config.hidden_size)
self.dropout = torch.nn.Dropout(0.1)
self.classifier_1 = torch.nn.Linear(config.hidden_size, 2)
self.classifier_2 = torch.nn.Linear(config.hidden_size, 2)
self.classifier_3 = torch.nn.Linear(config.hidden_size, 2)
self.classifier_4 = torch.nn.Linear(config.hidden_size, 2)
self.classifier_5 = torch.nn.Linear(config.hidden_size, 2)
def forward(self, input_ids, attention_mask, token_type_ids):
outputs = self.l1(input_ids=input_ids, attention_mask=attention_mask)
pooler = outputs[1]
pooler = self.pre_classifier(pooler)
pooler = self.dropout(pooler)
output_1 = self.classifier_1(pooler)
output_2 = self.classifier_2(pooler)
output_3 = self.classifier_3(pooler)
output_4 = self.classifier_4(pooler)
output_5 = self.classifier_5(pooler)
return output_1,output_2,output_3,output_4,output_5
def loss_fn(outputs, targets):
return torch.nn.CrossEntropyLoss()(outputs, targets)
def train(epoch,model,training_loader):
model.train()
loop = tqdm(enumerate(training_loader, 0),total=len(training_loader))
loop.set_description(f"Epoch {epoch}")
for _,data in loop:
ids = data['ids'].to(device, dtype = torch.long)
mask = data['mask'].to(device, dtype = torch.long)
token_type_ids = data['token_type_ids'].to(device, dtype = torch.long)
targets = data['targets'].to(device, dtype = torch.long)
output_1,output_2,output_3,output_4,output_5 = model(ids, mask, token_type_ids)
optimizer.zero_grad()
loss_1 = loss_fn(output_1, targets[:,0])
loss_2 = loss_fn(output_2, targets[:,1])
loss_3 = loss_fn(output_3, targets[:,2])
loss_4 = loss_fn(output_4, targets[:,3])
loss_5 = loss_fn(output_5, targets[:,4])
loss = (loss_1 + loss_2 + loss_3 + loss_4 + loss_5)
loop.set_postfix(loss=loss.item())
loss.backward()
optimizer.step()
def validation(testing_loader,model):
model.eval()
fin_targets=[]
fin_outputs=[]
with torch.no_grad():
for _, data in tqdm(enumerate(testing_loader, 0),total=len(testing_loader)):
ids = data['ids'].to(device, dtype = torch.long)
mask = data['mask'].to(device, dtype = torch.long)
token_type_ids = data['token_type_ids'].to(device, dtype = torch.long)
targets = data['targets'].to(device, dtype = torch.float)
output_1,output_2,output_3,output_4,output_5 = model(ids, mask, token_type_ids)
prob_1 = nn.Softmax(dim=1)(output_1)
prob_2 = nn.Softmax(dim=1)(output_2)
prob_3 = nn.Softmax(dim=1)(output_3)
prob_4 = nn.Softmax(dim=1)(output_4)
prob_5 = nn.Softmax(dim=1)(output_5)
fin_targets.extend(targets.cpu().detach().numpy().tolist())
fin_outputs+=[[p1.cpu().detach().numpy().tolist(),p2.cpu().detach().numpy().tolist(),
p3.cpu().detach().numpy().tolist(),p4.cpu().detach().numpy().tolist(),
p5.cpu().detach().numpy().tolist()] for p1,p2,p3,p4,p5 in zip(prob_1, prob_2, prob_3, prob_4, prob_5)]
return fin_outputs, fin_targets
MAX_LEN = 128
TRAIN_BATCH_SIZE = 32
VALID_BATCH_SIZE = 32
EPOCHS = 6
LEARNING_RATE = 2e-5
special_tokens = ["[US]","[AU]","[GB]","[ZA]","[SG]","@USER","URL"]
col_idx_permutation = list(permutations(range(5)))
for model_path,model_name in zip(models,model_names):
tokenizer = AutoTokenizer.from_pretrained(model_path, truncation=True)
tokenizer.add_tokens(special_tokens)
res_row_list = []
res_df = pd.DataFrame()
train_file = './data_splits/CREHate_train.csv'
valid_file = './data_splits/CREHate_valid.csv'
test_file = './data_splits/CREHate_test.csv'
train_data = pd.read_csv(train_file)
valid_data = pd.read_csv(valid_file)
test_data = pd.read_csv(test_file)
for idx,idx_permute in enumerate(col_idx_permutation):
hate_cols = [_hate_cols[i] for i in idx_permute]
train_df = pd.DataFrame()
train_df['text'] = train_data['Text']
train_df['labels'] = train_data[hate_cols].values.tolist()
valid_df = pd.DataFrame()
valid_df['text'] = valid_data['Text']
valid_df['labels'] = valid_data[hate_cols].values.tolist()
test_df = pd.DataFrame()
test_df['text'] = test_data['Text']
test_df['labels'] = test_data[hate_cols].values.tolist()
training_set = MultiTaskDataset(train_df, tokenizer, MAX_LEN)
valid_set = MultiTaskDataset(valid_df, tokenizer, MAX_LEN)
testing_set = MultiTaskDataset(test_df, tokenizer, MAX_LEN)
train_params = {'batch_size': TRAIN_BATCH_SIZE,
'shuffle': True,
'num_workers': torch.cuda.device_count()
}
valid_params = {'batch_size': VALID_BATCH_SIZE,
'shuffle': True,
'num_workers': torch.cuda.device_count()
}
test_params = {'batch_size': VALID_BATCH_SIZE,
'shuffle': False,
'num_workers': torch.cuda.device_count()
}
training_loader = DataLoader(training_set, **train_params)
valid_loader = DataLoader(valid_set, **valid_params)
testing_loader = DataLoader(testing_set, **test_params)
model = Classifier(model_path,tokenizer)
model = nn.DataParallel(model, device_ids =list(range(torch.cuda.device_count()))).to(device)
optimizer = torch.optim.AdamW(params = model.parameters(), lr=LEARNING_RATE, eps=1e-8)
min_hamming_loss = 1
best_model = None
for epoch in range(EPOCHS):
train(epoch,model,training_loader)
outputs, targets = validation(valid_loader,model)
final_outputs = np.array([[0 if output[0]>output[1] else 1 for output in row] for row in outputs])
val_hamming_loss = metrics.hamming_loss(targets, final_outputs)
val_hamming_score = hamming_score(np.array(targets), np.array(final_outputs))
print(f"Hamming Score = {val_hamming_score}")
print(f"Hamming Loss = {val_hamming_loss}")
if val_hamming_loss < min_hamming_loss:
min_hamming_loss = val_hamming_loss
best_model = model
if best_model is not None:
outputs, targets = validation(testing_loader,best_model)
final_outputs = np.array([[0 if output[0]>output[1] else 1 for output in row] for row in outputs])
tst_hamming_loss = metrics.hamming_loss(targets, final_outputs)
tst_hamming_score = hamming_score(np.array(targets), np.array(final_outputs))
cols = [f'{model_name}-MT-{country}' for country in [codes[i] for i in idx_permute]]
outputs_df = pd.DataFrame(final_outputs,columns=cols)
total = pd.concat([test_data[hate_cols],outputs_df],axis=1)
total.to_csv(f'./res/{model_name}-MT-ALL-P-{idx}-res.csv',index=False)
test_data = pd.concat([test_data,outputs_df],axis=1)
print(test_data)
print(total)
print('\tAcc\tF1\tH-F1\tN-F1')
row = []
for hate_col,code in zip(hate_cols,[codes[i] for i in idx_permute]):
acc = metrics.accuracy_score(test_data[hate_col],outputs_df[f'{model_name}-MT-{code}'])
f1 = metrics.f1_score(test_data[hate_col], outputs_df[f'{model_name}-MT-{code}'],average='macro')
n,h = metrics.f1_score(test_data[hate_col], outputs_df[f'{model_name}-MT-{code}'],average=None)
r = metrics.recall_score(test_data[hate_col], outputs_df[f'{model_name}-MT-{code}'])
print(f'{code}:\t{acc:.4f}\t{f1:.4f}\t{n:.4f}\t{h:.4f}\t{r:.4f}')
row += [acc,f1,n,h,r]
res_cols = []
for code in [codes[i] for i in idx_permute]:
res_cols += [f'{code}-{score}' for score in ['acc','f1','h','n','r']]
res_df_row = pd.DataFrame([row],index=[idx],columns=res_cols)
res_df = pd.concat([res_df,res_df_row])
if 'avg' in res_df.index:
res_df.drop('avg',inplace=True)
res_df.loc['avg'] = res_df.mean(axis=0)
print(res_df)
res_df.to_csv(f'./res/{model_name}-MT-ALL-P-res-scores.csv')