rollerhafeezh-amikom
commited on
Commit
•
b9e9a81
1
Parent(s):
c782dca
Update README.md
Browse files
README.md
CHANGED
@@ -81,15 +81,33 @@ It achieves the following results on the evaluation set:
|
|
81 |
|
82 |
## Model description
|
83 |
|
84 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
85 |
|
86 |
## Intended uses & limitations
|
87 |
|
88 |
-
|
89 |
|
90 |
## Training and evaluation data
|
91 |
|
92 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
93 |
|
94 |
## Training procedure
|
95 |
|
|
|
81 |
|
82 |
## Model description
|
83 |
|
84 |
+
The XLM-RoBERTa model was proposed in [Unsupervised Cross-lingual Representation Learning at Scale](https://arxiv.org/abs/1911.02116) by Alexis Conneau, Kartikay Khandelwal, Naman Goyal, Vishrav Chaudhary, Guillaume Wenzek, Francisco Guzmán, Edouard Grave, Myle Ott, Luke Zettlemoyer and Veselin Stoyanov. It is based on Facebook's RoBERTa model released in 2019. It is a large multi-lingual language model, trained on 2.5TB of filtered CommonCrawl data.
|
85 |
+
|
86 |
+
- **Developed by:** See [associated paper](https://arxiv.org/abs/1911.02116)
|
87 |
+
- **Model type:** Multi-lingual model
|
88 |
+
- **Language(s) (NLP) or Countries (images):** XLM-RoBERTa is a multilingual model trained on 100 different languages; see [GitHub Repo](https://github.com/facebookresearch/fairseq/tree/main/examples/xlmr) for full list; model is fine-tuned on a dataset in English
|
89 |
+
- **License:** More information needed
|
90 |
+
- **Related Models:** [RoBERTa](https://huggingface.co/roberta-base), [XLM](https://huggingface.co/docs/transformers/model_doc/xlm)
|
91 |
+
- **Parent Model:** [XLM-RoBERTa](https://huggingface.co/xlm-roberta-base)
|
92 |
+
- **Resources for more information:**
|
93 |
+
- [GitHub Repo](https://github.com/facebookresearch/fairseq/tree/main/examples/xlmr)
|
94 |
|
95 |
## Intended uses & limitations
|
96 |
|
97 |
+
This model can be used to extract multilingual information such as location, date and time on social media (Twitter, etc.). This model is limited by an Indonesian language training data set to be tested in 4 languages (English, Spanish, Italian and Slovak) using zero-shot transfer learning techniques to extract multilingual information.
|
98 |
|
99 |
## Training and evaluation data
|
100 |
|
101 |
+
This model was fine-tuned on Indonesian NER datasets.
|
102 |
+
Abbreviation|Description
|
103 |
+
-|-
|
104 |
+
O|Outside of a named entity
|
105 |
+
B-LOC |Beginning of a location right after another location
|
106 |
+
I-LOC |Location
|
107 |
+
B-DAT Beginning of a date right after another date
|
108 |
+
I-DAT |Date
|
109 |
+
B-TIM |Beginning of a time right after another time
|
110 |
+
I-TIM |Time
|
111 |
|
112 |
## Training procedure
|
113 |
|