Update README.md
Browse files
README.md
CHANGED
@@ -5,9 +5,7 @@ Google Research's [ELECTRA](https://github.com/google-research/electra).
|
|
5 |
|
6 |
Tokenization and pre-training CoLab: https://colab.research.google.com/drive/1GngBFn_Ge5Hd2XI2febBhZyU7GDiqw5w
|
7 |
|
8 |
-
V2 (current): 190,000 steps
|
9 |
-
|
10 |
-
V1: 100,000 steps;
|
11 |
|
12 |
## Usage
|
13 |
|
@@ -17,10 +15,10 @@ https://www.kaggle.com/sudalairajkumar/tamil-nlp
|
|
17 |
Notebook: https://colab.research.google.com/drive/1_rW9HZb6G87-5DraxHvhPOzGmSMUc67_?usp=sharin
|
18 |
|
19 |
The model outperformed mBERT on news classification:
|
20 |
-
(Random: 16.7%, mBERT: 53.0%, TaMillion:
|
21 |
|
22 |
The model slightly outperformed mBERT on movie reviews:
|
23 |
-
(RMSE - mBERT: 0.657, TaMillion: 0.
|
24 |
|
25 |
Equivalent accuracy on the Tirukkural topic task.
|
26 |
|
|
|
5 |
|
6 |
Tokenization and pre-training CoLab: https://colab.research.google.com/drive/1GngBFn_Ge5Hd2XI2febBhZyU7GDiqw5w
|
7 |
|
8 |
+
V2 (current): 190,000 steps; (V1 was 100,000 steps)
|
|
|
|
|
9 |
|
10 |
## Usage
|
11 |
|
|
|
15 |
Notebook: https://colab.research.google.com/drive/1_rW9HZb6G87-5DraxHvhPOzGmSMUc67_?usp=sharin
|
16 |
|
17 |
The model outperformed mBERT on news classification:
|
18 |
+
(Random: 16.7%, mBERT: 53.0%, TaMillion: 69.6%)
|
19 |
|
20 |
The model slightly outperformed mBERT on movie reviews:
|
21 |
+
(RMSE - mBERT: 0.657, TaMillion: 0.627)
|
22 |
|
23 |
Equivalent accuracy on the Tirukkural topic task.
|
24 |
|