roberta-base / README.md
zemerov's picture
Create README.md
5a19b1f
metadata
license: apache-2.0
language:
  - ru
  - en
library_name: transformers

RoBERTa-base from deepvk

Pretrained bidirectional encoder for russian language.

Model Details

Model Description

Model was pretrained using standard MLM objective on a large text corpora including open social data, books, Wikipedia, webpages etc.

  • Developed by: VK Applied Research Team
  • Model type: RoBERTa
  • Languages: Mostly russian and small fraction of other languages
  • License: Apache 2.0

How to Get Started with the Model

from transformers import AutoTokenizer, AutoModel

tokenizer = AutoTokenizer.from_pretrained("deepvk/roberta-base")
model = AutoModel.from_pretrained("deepvk/roberta-base")

text = "Привет, мир!"

inputs = tokenizer(text, return_tensors='pt')
predictions = model(**inputs)

Training Details

Training Data

Mix of the following data:

Training Procedure

Preprocessing [optional]

[More Information Needed]

Training Hyperparameters

  • Training regime: [More Information Needed]

Speeds, Sizes, Times [optional]

Standard RoBERTA-base size;

Evaluation

Testing Data, Factors & Metrics

Testing Data

[More Information Needed]

Factors

[More Information Needed]

Metrics

[More Information Needed]

Results

[More Information Needed]

Summary

Compute Infrastructure

Model was trained using 8xA100 for ~22 days.