edaiofficial's picture
initial commits
78aa4ee
## Data
JW300 : English-Zulu
## Model Architecture
### Text Preprocessing
- Remove blank/empty rows : 9037(0.85 %) samples
- Removed duplicates from source text : 82999(7.88 %) samples
- Removed duplicates from target text : 5045(0.52 %) samples
- Removed all numeric-only text : 182(0.02 %) samples
- Removed rows where text is fewer than orequal to 8 characters long from source text: 6272(0.65 %) samples
- Removed rows where text is fewer than orequal to 8 characters long from target text: 713(0.07 %) samples
- Removed rows where text is in test set: 1068(0.11 %) samples
### BPE Tokenization
- vocab size : 4000 (superior results than 10X)
### Model Config
- Details in supplied config file but used fewer transformer layers than in default notebook, with more attention heads and lower embedding size
- Trained for 235000 steps
- Took few hours on a single P100 GPU on Google colab over a three days (stopped training saved best model then reloaded that model the next day)
## Results
### Curious analysis of the tokenization
> There are 66255 english tokens in the test set vocab, 2072 are unique
>
> There are 67851 zulu tokens in the test set vocab, 2336 are unique
>
> These results are in the same notebook as used for training. (Could something similar help inform BPE vocab size choices ?)
### Translation results
> 2019-11-13 07:43:32,728 Hello! This is Joey-NMT.
>
> 2019-11-13 07:44:03,502 dev bleu: 13.64 [Beam search decoding with beam size = 5 and alpha = 1.0]
>
> 2019-11-13 07:44:24,289 test bleu: 4.87 [Beam search decoding with beam size = 5 and alpha = 1.0]`
Download model weights from : [here](https://drive.google.com/open?id=1-QLxP7xLqu-AqDQkm1XaCtDEex1Oseo0)