bert-large-relation14

Finetuned BERT model for 14-class classification. It was introduced in the paper: Automatic Slide Generation Using Discourse Relations and first released in this repository. This model is uncased: it does not make a difference between english and English.

In our proposed method in this paper, we only used this model for the classification of discourse relation between the FIRST and SECOND sentence in summarized sentences. The model that is used between the other sentences is this model. If you are curious about our proposed method, it's better to see that model.

Descliption

This model can classify the relation between the sentence pair of input.

Now we are working on preparing the Model card. Please wait for a few days.

The model trained from bert-large-uncased on the dataset published in the paper: Automatic Prediction of Discourse Connectives.

The dataset to make this model is based on English Wikipedia data and has 20 labels. However, this model will classify into 14 labels. This is because the 20-class data set was restructured to 14 classes to suit our research objective of "automatic slide generation. This distribution is shown below.

Level 1 Level 2 Level 3 Connectives (20)
Temporal Synchronous meanwhile
Temporal Asynchronous Precedence then,
Temporal Asynchronous Precedence finally,
Temporal Asynchronous Succession by then
Contingency Cause Result therefore
Comparison Concession Arg2-as-denier however,
Comparison Concession Arg2-as-denier nevertheless
Comparison Contrast on the other hand,
Comparison Contrast by contrast,
Expansion Conjunction and
Expansion Conjunction moreover
Expansion Conjunction indeed
Expansion Equivalence in other words
Expansion Exception Arg1-as-excpt otherwise
Expansion Instantiation Arg2-as-instance for example,
Expansion Level-of-detail Arg1-as-detail overall,
Expansion Level-of-detail Arg2-as-detail in particular,
Expansion Substitution Arg2-as-subst instead
Expansion Substitution Arg2-as-subst rather

Training

The model was trained using AutoModelForSequenceClassification.from_pretrained

training_args = TrainingArguments(
    output_dir = output_dir,
    save_strategy="epoch",
    num_train_epochs = 5,
    per_device_train_batch_size=16,
    per_device_eval_batch_size=32,
    warmup_steps=0,
    weight_decay=0.01,
    logging_dir="./logs",
    evaluation_strategy="epoch",
    learning_rate=2e-5,
    metric_for_best_model="f1",
    load_best_model_at_end=True
)

Evaluation (14 labels and original 20 labels classification) using the dataset test split gives:

Model Macro F1 Accuracy Precision Recall
14 labels classification 0.586 0.589 0.630 0.591
20 labels classification 0.478 0.488 0.536 0.488
Downloads last month
17
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.