Edit model card

roberta-base-wechsel-ukrainian

roberta-base transferred to Ukrainian using the method from the NAACL2022 paper WECHSEL: Effective initialization of subword embeddings for cross-lingual transfer of monolingual language models.

Evaluation

Evaluation was done on lang-uk's ner-uk project, the Ukrainian portion of WikiANN and the Ukrainian IU corpus from the Universal Dependencies project. Evaluation results are the mean of 5 runs with different seeds.

Validation Results

lang-uk NER (Micro F1) WikiANN (Micro F1) UD Ukrainian IU POS (Accuracy)
roberta-base-wechsel-ukrainian 88.06 (0.50) 92.96 (0.08) 98.70 (0.05)
roberta-large-wechsel-ukrainian 89.27 (0.53) 93.22 (0.15) 98.86 (0.03)
roberta-base-scratch-ukrainian* 85.49 (0.88) 91.91 (0.08) 98.49 (0.04)
roberta-large-scratch-ukrainian* 86.54 (0.70) 92.39 (0.16) 98.65 (0.09)
dbmdz/electra-base-ukrainian-cased-discriminator 87.49 (0.52) 93.20 (0.16) 98.60 (0.03)
xlm-roberta-base 86.68 (0.44) 92.41 (0.13) 98.53 (0.02)
xlm-roberta-large 86.64 (1.61) 93.01 (0.13) 98.71 (0.04)

Test Results

lang-uk NER (Micro F1) WikiANN (Micro F1) UD Ukrainian IU POS (Accuracy)
roberta-base-wechsel-ukrainian 90.81 (1.51) 92.98 (0.12) 98.57 (0.03)
roberta-large-wechsel-ukrainian 91.24 (1.16) 93.22 (0.17) 98.74 (0.06)
roberta-base-scratch-ukrainian* 89.57 (1.01) 92.05 (0.09) 98.31 (0.08)
roberta-large-scratch-ukrainian* 89.96 (0.89) 92.49 (0.15) 98.52 (0.04)
dbmdz/electra-base-ukrainian-cased-discriminator 90.43 (1.29) 92.99 (0.11) 98.59 (0.06)
xlm-roberta-base 90.86 (0.81) 92.27 (0.09) 98.45 (0.07)
xlm-roberta-large 90.16 (2.98) 92.92 (0.19) 98.71 (0.04)

*trained using the same exact training setup as the wechsel-* models, but without parameter transfer from WECHSEL.

License

MIT

Downloads last month
50
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.