This model is developed with transformers v4.10.3.

Train

#!/usr/bin/env bash

export CUDA_VISIBLE_DEVICES=0

OUTDIR=bert-based-uncased-mnli
WORKDIR=transformers/examples/pytorch/text-classification
cd $WORKDIR

nohup python run_glue.py \
    --model_name_or_path bert-base-uncased  \
    --task_name mnli  \
    --do_eval  \
    --do_train  \
    --per_device_train_batch_size 16  \
    --per_device_eval_batch_size 16  \
    --max_seq_length 128  \
    --num_train_epochs 3 \
    --overwrite_output_dir \
    --output_dir $OUTDIR 2>&1 | tee $OUTDIR/run.log &

Eval

export CUDA_VISIBLE_DEVICES=0

OUTDIR=eval-bert-based-uncased-mnli
WORKDIR=transformers/examples/pytorch/text-classification
cd $WORKDIR

nohup python run_glue.py \
    --model_name_or_path vuiseng9/bert-base-uncased-mnli  \
    --task_name mnli  \
    --do_eval  \
    --per_device_eval_batch_size 16  \
    --max_seq_length 128  \
    --overwrite_output_dir \
    --output_dir $OUTDIR 2>&1 | tee $OUTDIR/run.log &
Downloads last month
5
Hosted inference API
Text Classification
Examples
Examples
Mask token: [MASK]
This model can be loaded on the Inference API on-demand.