--- license: apache-2.0 --- ```python import torch import torch.nn as nn import torch.nn.functional as F from torch.nn import CrossEntropyLoss, KLDivLoss from transformers.modeling_outputs import TokenClassifierOutput from transformers import BertModel, BertPreTrainedModel class BertForHighlightPrediction(BertPreTrainedModel): _keys_to_ignore_on_load_unexpected = [r"pooler"] def __init__(self, config, **model_kwargs): super().__init__(config) # self.model_args = model_kargs["model_args"] self.num_labels = config.num_labels self.bert = BertModel(config, add_pooling_layer=False) classifier_dropout = ( config.classifier_dropout if config.classifier_dropout is not None else config.hidden_dropout_prob ) self.dropout = nn.Dropout(classifier_dropout) self.tokens_clf = nn.Linear(config.hidden_size, config.num_labels) self.tau = model_kwargs.pop('tau', 1) self.gamma = model_kwargs.pop('gamma', 1) self.soft_labeling = model_kwargs.pop('soft_labeling', False) self.init_weights() self.softmax = nn.Softmax(dim=-1) def forward(self, input_ids=None, probs=None, # soft-labeling attention_mask=None, token_type_ids=None, position_ids=None, head_mask=None, inputs_embeds=None, labels=None, output_attentions=None, output_hidden_states=None, return_dict=None,): outputs = self.bert( input_ids, attention_mask=attention_mask, token_type_ids=token_type_ids, position_ids=position_ids, head_mask=head_mask, inputs_embeds=inputs_embeds, output_attentions=output_attentions, output_hidden_states=output_hidden_states, return_dict=return_dict, ) tokens_output = outputs[0] highlight_logits = self.tokens_clf(self.dropout(tokens_output)) loss = None if labels is not None: loss_fct = CrossEntropyLoss() active_loss = attention_mask.view(-1) == 1 active_logits = highlight_logits.view(-1, self.num_labels) active_labels = torch.where( active_loss, labels.view(-1), torch.tensor(loss_fct.ignore_index).type_as(labels) ) loss_ce = loss_fct(active_logits, active_labels) loss_kl = 0 if self.soft_labeling: loss_fct = KLDivLoss(reduction='sum') active_mask = (attention_mask * token_type_ids).view(-1, 1) # BL 1 n_active = (active_mask == 1).sum() active_mask = active_mask.repeat(1, 2) # BL 2 input_logp = F.log_softmax(active_logits / self.tau, -1) # BL 2 target_p = torch.cat(( (1-probs).view(-1, 1), probs.view(-1, 1)), -1) # BL 2 loss_kl = loss_fct(input_logp, target_p * active_mask) / n_active loss = self.gamma * loss_ce + (1-self.gamma) * loss_kl # print("Loss:\n") # print(loss) # print(loss_kl) # print(loss_ce) return TokenClassifierOutput( loss=loss, logits=highlight_logits, hidden_states=outputs.hidden_states, attentions=outputs.attentions, ) @torch.no_grad() def inference(self, outputs): with torch.no_grad(): outputs = self.forward(**batch_inputs) probabilities = self.softmax(self.tokens_clf(outputs.hidden_states[-1])) predictions = torch.argmax(probabilities, dim=-1) # active filtering active_tokens = batch_inputs['attention_mask'] == 1 active_predictions = torch.where( active_tokens, predictions, torch.tensor(-1).type_as(predictions) ) outputs = { "probabilities": probabilities[:, :, 1].detach(), # shape: (batch, length) "active_predictions": predictions.detach(), "active_tokens": active_tokens, } return outputs ```