Update README.md
Browse files
README.md
CHANGED
@@ -20,7 +20,7 @@ tags:
|
|
20 |
This is a XLM-RoBERTa-large model tuned on a corpus of over 156 million tweets in ten languages: English, Spanish, Italian, Portuguese, French, Chinese, Hindi, Arabic, Dutch and Korean.
|
21 |
The model has been trained from the original XLM-RoBERTA-large checkpoint for 2 epochs with a batch size of 1024.
|
22 |
|
23 |
-
For best results,
|
24 |
```python
|
25 |
def preprocess(text):
|
26 |
new_text = []
|
|
|
20 |
This is a XLM-RoBERTa-large model tuned on a corpus of over 156 million tweets in ten languages: English, Spanish, Italian, Portuguese, French, Chinese, Hindi, Arabic, Dutch and Korean.
|
21 |
The model has been trained from the original XLM-RoBERTA-large checkpoint for 2 epochs with a batch size of 1024.
|
22 |
|
23 |
+
For best results, preprocess the tweets using the following method before passing them to the model:
|
24 |
```python
|
25 |
def preprocess(text):
|
26 |
new_text = []
|