File size: 2,370 Bytes
4106f13 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 |
import { Column } from 'primereact/column'
import ScoreField from './ScoreField'
const scoreBodyTemplate = (field, options = {}) => {
const { minScore = 0, maxScore = 1 } = options
return rowData => {
const score = rowData[field]
return ScoreField(score, minScore, maxScore)
}
}
const ScoreColumns = [
<Column
field='average'
header='Average'
headerTooltip='Language Proficiency Score (average of all displayed scores)'
sortable
body={scoreBodyTemplate('average', { minScore: 0.2, maxScore: 0.5 })}
style={{ minWidth: '5rem', maxWidth: '10rem' }}
/>,
<Column
field='translation_from_bleu'
header='Translation (from)'
headerTooltip='Translation performance from a language to all other languages (spBLEU score on a sample of the FLORES+ benchmark)'
sortable
body={scoreBodyTemplate('translation_from_bleu', {
minScore: 0,
maxScore: 0.5
})}
style={{ minWidth: '5rem', maxWidth: '10rem' }}
/>,
<Column
field='translation_to_bleu'
header='Translation (to)'
headerTooltip='Translation performance from all other languages to a language (spBLEU score on a sample of the FLORES+ benchmark)'
sortable
body={scoreBodyTemplate('translation_to_bleu', {
minScore: 0,
maxScore: 0.5
})}
style={{ minWidth: '5rem', maxWidth: '10rem' }}
/>,
<Column
field='classification_accuracy'
header='Classification'
headerTooltip='Classification performance (accuracy on a sample of the FLORES+ benchmark)'
sortable
body={scoreBodyTemplate('classification_accuracy', {
minScore: 0,
maxScore: 0.5
})}
style={{ minWidth: '5rem', maxWidth: '10rem' }}
/>,
// <Column
// field='language_modeling_chrf'
// header='Language Modeling'
// sortable
// body={scoreBodyTemplate('language_modeling_chrf', {
// minScore: 0.8,
// maxScore: 1
// })}
// style={{ minWidth: '5rem', maxWidth: '10rem' }}
// />,
<Column
field='mmlu_accuracy'
header='MMLU'
headerTooltip='Question Answering performance (accuracy on a sample of multilingual versions of the MMLU benchmark)'
sortable
body={scoreBodyTemplate('mmlu_accuracy', {
minScore: 0,
maxScore: 1
})}
style={{ minWidth: '5rem', maxWidth: '10rem' }}
/>
]
export default ScoreColumns
|