naumov-al commited on
Commit
ed249e3
1 Parent(s): 458660b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +8 -1
README.md CHANGED
@@ -1,9 +1,16 @@
 
 
 
 
 
 
 
1
  # XLM-RoBERTa-large-sag
2
 
3
  ## Model description
4
 
5
  This is a model based on the [XLM-RoBERTa large](https://huggingface.co/xlm-roberta-large) topology (provided by Facebook, see original [paper](https://arxiv.org/abs/1911.02116)) with additional training on two sets of medicine-domain texts:
6
- * 250.000 text reviews on medicines (1000-tokens-long in average) collected from the site irecommend.ru;
7
  * the raw part of the [RuDReC corpus](https://github.com/cimm-kzn/RuDReC) (about 1.4 million texts, see [paper](https://arxiv.org/abs/2004.03659)).
8
 
9
  The XLM-RoBERTa-large calculations for one epoch on this data were performed using one Nvidia Tesla v100 and the Huggingface Transformers library.
1
+ ---
2
+ language: multilingual
3
+ thumbnail: "url to a thumbnail used in social sharing"
4
+ tags: exbert
5
+ license: apache-2.0
6
+ ---
7
+
8
  # XLM-RoBERTa-large-sag
9
 
10
  ## Model description
11
 
12
  This is a model based on the [XLM-RoBERTa large](https://huggingface.co/xlm-roberta-large) topology (provided by Facebook, see original [paper](https://arxiv.org/abs/1911.02116)) with additional training on two sets of medicine-domain texts:
13
+ * about 250.000 text reviews on medicines (1000-tokens-long in average) collected from the site irecommend.ru;
14
  * the raw part of the [RuDReC corpus](https://github.com/cimm-kzn/RuDReC) (about 1.4 million texts, see [paper](https://arxiv.org/abs/2004.03659)).
15
 
16
  The XLM-RoBERTa-large calculations for one epoch on this data were performed using one Nvidia Tesla v100 and the Huggingface Transformers library.