xls-r_cv_ur / vocab.json
Bakht Ullah
Upload tokenizer
6102137
raw
history blame
536 Bytes
{
" ": 26,
"PAD": 43,
"[UNK]": 44,
"آ": 22,
"أ": 35,
"ؤ": 11,
"ئ": 23,
"ا": 21,
"ب": 13,
"ت": 36,
"ث": 18,
"ج": 5,
"ح": 39,
"خ": 15,
"د": 40,
"ذ": 28,
"ر": 3,
"ز": 19,
"س": 12,
"ش": 6,
"ص": 38,
"ض": 17,
"ط": 25,
"ظ": 8,
"ع": 37,
"غ": 41,
"ف": 4,
"ق": 10,
"ل": 34,
"م": 42,
"ن": 33,
"و": 1,
"ً": 24,
"ٹ": 16,
"پ": 7,
"چ": 2,
"ڈ": 9,
"ڑ": 27,
"ک": 0,
"گ": 14,
"ں": 20,
"ھ": 29,
"ہ": 32,
"ی": 31,
"ے": 30
}