--- license: apache-2.0 language: - ru - en library_name: transformers --- # RoBERTa-base from deepvk Pretrained bidirectional encoder for russian language. ## Model Details ### Model Description Model was pretrained using standard MLM objective on a large text corpora including open social data, books, Wikipedia, webpages etc. - **Developed by:** VK Applied Research Team - **Model type:** RoBERTa - **Languages:** Mostly russian and small fraction of other languages - **License:** Apache 2.0 ## How to Get Started with the Model ``` from transformers import AutoTokenizer, AutoModel tokenizer = AutoTokenizer.from_pretrained("deepvk/roberta-base") model = AutoModel.from_pretrained("deepvk/roberta-base") text = "Привет, мир!" inputs = tokenizer(text, return_tensors='pt') predictions = model(**inputs) ``` ## Training Details ### Training Data Mix of the following data: ### Training Procedure #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] #### Speeds, Sizes, Times [optional] Standard RoBERTA-base size; ## Evaluation ### Testing Data, Factors & Metrics #### Testing Data [More Information Needed] #### Factors [More Information Needed] #### Metrics [More Information Needed] ### Results [More Information Needed] #### Summary ## Compute Infrastructure Model was trained using 8xA100 for ~22 days.