Cetvel / results /zero-shot /aya-expanse-8b.json
Ilker Kesen
initialize the first version
500fbd7
{
"model": {
"model": "CohereForAI/aya-expanse-8b",
"api": "hf",
"architecture": "CohereForCausalLM",
"max_length": 8192,
"dtype": "float16",
"type": "instruction-tuned",
"num_parameters": "8b"
},
"results": [
{
"name": "belebele_tr",
"task": "multiple_choice",
"acc": 0.7355555555555555,
"acc_norm": 0.7355555555555555
},
{
"name": "exams_tr",
"task": "multiple_choice",
"acc": 0.3155216284987277,
"acc_norm": 0.3460559796437659
},
{
"name": "check_worthiness",
"task": "multiple_choice",
"acc": 0.4026508226691042,
"acc_norm": 0.6224862888482633
},
{
"name": "gecturk_generation",
"task": "grammatical_error_correction",
"exact_match": 0.0018296499590736194
},
{
"name": "ironytr",
"task": "text_classification",
"acc": 0.505,
"acc_norm": 0.49833333333333335
},
{
"name": "mkqa_tr",
"task": "extractive_question_answering",
"exact_match": 0.06954720331459012,
"f1": 0.13476533908972033
},
{
"name": "mlsum_tr",
"task": "summarization",
"rouge1": 0.363610486561065,
"rouge2": 0.21362825588593481,
"rougeL": 0.29773476508614094
},
{
"name": "mnli_tr",
"task": "natural_language_inference",
"acc": 0.3078,
"acc_norm": 0.35
},
{
"name": "news_cat",
"task": "text_classification",
"acc": 0.76,
"acc_norm": 0.58
},
{
"name": "offenseval_tr",
"task": "text_classification",
"acc": 0.2675736961451247,
"acc_norm": 0.7956349206349206
},
{
"name": "relevance_judgment",
"task": "multiple_choice",
"acc": 0.5877513711151737,
"acc_norm": 0.579981718464351
},
{
"name": "snli_tr",
"task": "natural_language_inference",
"acc": 0.344,
"acc_norm": 0.3435
},
{
"name": "sts_tr",
"task": "text_classification",
"acc": 0.2095721537345903,
"acc_norm": 0.21029731689630166
},
{
"name": "tquad",
"task": "extractive_question_answering",
"exact_match": 0.13452914798206278,
"f1": 0.435087842533856
},
{
"name": "turkish_plu_goal_inference",
"task": "multiple_choice",
"acc": 0.4062126642771804,
"acc_norm": 0.3930704898446834
},
{
"name": "turkish_plu_next_event_prediction",
"task": "multiple_choice",
"acc": 0.4900763358778626,
"acc_norm": 0.5465648854961832
},
{
"name": "turkish_plu_step_inference",
"task": "multiple_choice",
"acc": 0.3464052287581699,
"acc_norm": 0.4395424836601307
},
{
"name": "turkish_plu_step_ordering",
"task": "multiple_choice",
"acc": 0.5935357492654261,
"acc_norm": 0.5935357492654261
},
{
"name": "wiki_lingua_tr",
"task": "summarization",
"rouge1": 0.3064320242538614,
"rouge2": 0.1340385267540697,
"rougeL": 0.24764232131755232
},
{
"name": "wmt-tr-en-prompt",
"task": "machine_translation",
"wer": 0.7822550373875778,
"bleu": 0.17034711245148307
},
{
"name": "xcopa_tr",
"task": "multiple_choice",
"acc": 0.578,
"acc_norm": 0.578
},
{
"name": "xlsum_tr",
"task": "summarization",
"rouge1": 0.26621653203927675,
"rouge2": 0.133428873146516,
"rougeL": 0.2083669711429916
},
{
"name": "xnli_tr",
"task": "natural_language_inference",
"acc": 0.4919678714859438,
"acc_norm": 0.4919678714859438
},
{
"name": "xquad_tr",
"task": "extractive_question_answering",
"exact_match": 0.2495798319327731,
"f1": 0.4735125568867167
}
]
}