File size: 1,177 Bytes
8798d43
 
d2adb0e
 
 
 
 
 
 
8798d43
a4664f7
d2adb0e
b5ddd5c
d2adb0e
 
 
 
 
b5ddd5c
d2adb0e
b5ddd5c
 
d2adb0e
b5ddd5c
d2adb0e
 
 
b5ddd5c
 
 
 
d2adb0e
 
b5ddd5c
d2adb0e
b5ddd5c
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
---
license: mit
language:
- ar
- en
tags:
- T5
- mT5
- Transformers
---
# Model Card

An Arabic LLM derived from Google's mT5 multi-lingual model 

## Model Details

### Model Description

This is a smaller version of the google/mt5-base model with only Arabic and some English embeddings left.

The original model has 582M parameters, with 384M of them being input and output embeddings.
After shrinking the sentencepiece vocabulary from 250K to 30K (top 10K English and top 20K Arabic tokens) the number of model parameters reduced to 244M parameters, and model size reduced from 2.2GB to 0.9GB - 42% of the original one.

The creation of this model was inspired from David Dales'article "<a href="https://towardsdatascience.com/how-to-adapt-a-multilingual-t5-model-for-a-single-language-b9f94f3d9c90">How to adapt a multilingual T5 model for a single language</a>" in which mT5 was compressed to support Russian and English languages along with the source code.



- **Developed by:** Moustafa Banbouk
- **Model type:** Unsupervised LLM
- **Language(s) (NLP):** Arabic, English
- **License:** MIT


### Downstream Uses

Quesion Answering, Summarization, Classification ...