junnyu's picture
update
072e3cd
metadata
language: en
thumbnail: https://github.com/junnyu
tags:
  - pytorch
  - electra
license: MIT
datasets:
  - openwebtext

一、 个人在openwebtext数据集上训练得到的electra-small模型

二、 复现结果(dev dataset)

Model CoLA SST MRPC STS QQP MNLI QNLI RTE Avg.
ELECTRA-Small-OWT(original) 56.8 88.3 87.4 86.8 88.3 78.9 87.9 68.5 80.36
ELECTRA-Small-OWT (this) 55.82 89.67 87.0 86.96 89.28 80.08 87.50 66.07 80.30

三、 训练细节

  • 数据集 openwebtext
  • 训练batch_size 256
  • 学习率lr 5e-4
  • 最大句子长度max_seqlen 128
  • 训练total step 62.5W
  • GPU RTX3090
  • 训练时间总共耗费2.5天

四、 使用

import torch
from transformers.models.electra import ElectraModel, ElectraTokenizer
tokenizer = ElectraTokenizer.from_pretrained("junnyu/electra_small_generator")
model = ElectraModel.from_pretrained("junnyu/electra_small_generator")
inputs = tokenizer("Beijing is the capital of [MASK].", return_tensors="pt")
with torch.no_grad():
    outputs = model(**inputs)