pcuenq's picture
pcuenq HF staff
Updata dataset link
841de37
|
raw
history blame
2.93 kB
metadata
language: es
license: cc-by-4.0
tags:
  - spanish
  - roberta
  - bertin
pipeline_tag: text-classification
widget:
  - text: >-
      La ciencia nos enseña, en efecto, a someter nuestra razón a la verdad y a
      conocer y juzgar las cosas como son, es decir, como ellas mismas eligen
      ser y no como quisiéramos que fueran.

Readability ES Sentences for two classes

Model based on the Roberta architecture finetuned on BERTIN for readability assessment of Spanish texts.

Description and performance

This version of the model was trained on a mix of datasets, using sentence-level granularity when possible. The model performs binary classification among the following classes:

  • Simple.
  • Complex.

It achieves a F1 macro average score of 0.8923, measured on the validation set.

Model variants

Datasets

Training details

Please, refer to this training run for full details on hyperparameters and training regime.

Biases and Limitations

  • Due to the scarcity of data and the lack of a reliable gold test set, performance metrics are reported on the validation set.
  • One of the datasets involved is the Spanish version of newsela, which is frequently used as a reference. However, it was created by translating previous datasets, and therefore it may contain somewhat unnatural phrases.
  • Some of the datasets used cannot be publicly disseminated, making it more difficult to assess the existence of biases or mistakes.
  • Language might be biased towards the Spanish dialect spoken in Spain. Other regional variants might be sub-represented.
  • No effort has been performed to alleviate the shortcomings and biases described in the original implementation of BERTIN.

Authors