antypasd's picture
updated README, evaluation
bde3126
raw
history blame
No virus
942 Bytes
# Evalution script (evalution.py) instructions
* We provide an evaluation script that can easily calculate the scores achieved in each task, cluster and overall.
- The predictions must be in .txt format.
- All predictions files should be in the same directory and follow the naming conversions as in the "dummy_predictions" directory provided.
- Each prediction file must contain the predictions made on the test set of the respective task with each line corresponding to one line.
- The predictions should be written in full (e.g. 'strongly negative', instead of 0) and in cases of multiple outputs (e.g. TweetTopic) should be comma separated.
* "dummy_predictions" directory provides the prediciton files produced by the roberta-base model tested and can be used as a template.
* Steps to run the script:
1. Install requirements.txt
2. Run script: python evalution.py -p {directory with predictions}. e.g. python -p ./dummy_predictions