rollerhafeezh-amikom commited on
Commit
b9e9a81
1 Parent(s): c782dca

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +21 -3
README.md CHANGED
@@ -81,15 +81,33 @@ It achieves the following results on the evaluation set:
81
 
82
  ## Model description
83
 
84
- More information needed
 
 
 
 
 
 
 
 
 
85
 
86
  ## Intended uses & limitations
87
 
88
- More information needed
89
 
90
  ## Training and evaluation data
91
 
92
- More information needed
 
 
 
 
 
 
 
 
 
93
 
94
  ## Training procedure
95
 
 
81
 
82
  ## Model description
83
 
84
+ The XLM-RoBERTa model was proposed in [Unsupervised Cross-lingual Representation Learning at Scale](https://arxiv.org/abs/1911.02116) by Alexis Conneau, Kartikay Khandelwal, Naman Goyal, Vishrav Chaudhary, Guillaume Wenzek, Francisco Guzmán, Edouard Grave, Myle Ott, Luke Zettlemoyer and Veselin Stoyanov. It is based on Facebook's RoBERTa model released in 2019. It is a large multi-lingual language model, trained on 2.5TB of filtered CommonCrawl data.
85
+
86
+ - **Developed by:** See [associated paper](https://arxiv.org/abs/1911.02116)
87
+ - **Model type:** Multi-lingual model
88
+ - **Language(s) (NLP) or Countries (images):** XLM-RoBERTa is a multilingual model trained on 100 different languages; see [GitHub Repo](https://github.com/facebookresearch/fairseq/tree/main/examples/xlmr) for full list; model is fine-tuned on a dataset in English
89
+ - **License:** More information needed
90
+ - **Related Models:** [RoBERTa](https://huggingface.co/roberta-base), [XLM](https://huggingface.co/docs/transformers/model_doc/xlm)
91
+ - **Parent Model:** [XLM-RoBERTa](https://huggingface.co/xlm-roberta-base)
92
+ - **Resources for more information:**
93
+ - [GitHub Repo](https://github.com/facebookresearch/fairseq/tree/main/examples/xlmr)
94
 
95
  ## Intended uses & limitations
96
 
97
+ This model can be used to extract multilingual information such as location, date and time on social media (Twitter, etc.). This model is limited by an Indonesian language training data set to be tested in 4 languages (English, Spanish, Italian and Slovak) using zero-shot transfer learning techniques to extract multilingual information.
98
 
99
  ## Training and evaluation data
100
 
101
+ This model was fine-tuned on Indonesian NER datasets.
102
+ Abbreviation|Description
103
+ -|-
104
+ O|Outside of a named entity
105
+ B-LOC |Beginning of a location right after another location
106
+ I-LOC |Location
107
+ B-DAT Beginning of a date right after another date
108
+ I-DAT |Date
109
+ B-TIM |Beginning of a time right after another time
110
+ I-TIM |Time
111
 
112
  ## Training procedure
113