pablocosta's picture
Update README.md
e12144d
metadata
language: pt
license: mit
tags:
  - bert
  - pytorch
datasets:
  - Twitter

BERTabaporu: a genre-specific pre-trained model of Portuguese-speaking social media

Introduction

This model is currently under evaluation - details will be published soon

Available models

Model Arch. #Layers #Params
pablocosta/bertabaporu-base-uncased BERT-Base 12 110M
pablocosta/bertabaporu-large-uncased BERT-Large 24 335M

Usage

from transformers import AutoTokenizer  # Or BertTokenizer
from transformers import AutoModelForPreTraining  # Or BertForPreTraining for loading pretraining heads
from transformers import AutoModel  # or BertModel, for BERT without pretraining heads
model = AutoModelForPreTraining.from_pretrained('pablocosta/bertabaporu-base-uncased')
tokenizer = AutoTokenizer.from_pretrained('pablocosta/bertabaporu-base-uncased')