Edit model card

flan-t5-large-da-multiwoz2.1_fs0.01

This model was trained from scratch on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.4733
  • Accuracy: 35.7018
  • Num: 3689
  • Gen Len: 16.0656

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 24
  • seed: 1799
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 20

Training results

Training Loss Epoch Step Validation Loss Accuracy Num Gen Len
0.4198 2.6 200 0.4959 33.3159 3689 17.0976
0.3487 5.19 400 0.5050 33.9202 3689 16.9694
0.4306 7.79 600 0.4597 34.2354 3689 16.7612
0.38 10.39 800 0.4668 35.3878 3689 16.5993
0.3563 12.99 1000 0.4622 34.9845 3689 15.743
0.3208 15.58 1200 0.4668 35.6226 3689 16.1944
0.3137 18.18 1400 0.4733 35.7018 3689 16.0656

Framework versions

  • Transformers 4.18.0
  • Pytorch 1.10.0+cu111
  • Datasets 2.5.1
  • Tokenizers 0.12.1
Downloads last month
3
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.