Korventenn commited on
Commit
ff0ccff
1 Parent(s): 6f1de89

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +54 -0
README.md CHANGED
@@ -1,3 +1,57 @@
1
  ---
2
  license: apache-2.0
 
 
 
 
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: apache-2.0
3
+ datasets:
4
+ - giga_fren
5
+ - opus100
6
+ language:
7
+ - fr
8
+ - en
9
  ---
10
+
11
+
12
+ # Model Card for fr_en-t5-small
13
+
14
+ <!-- Provide a quick summary of what the model is/does. -->
15
+
16
+ This model has been optimized for French and English language processing while minimizing overall size. To achieve this, I only retained relevant parameters and tokens specific to these two languages, ensuring that performance remains as good as the original mt5.
17
+
18
+ ## Model Details
19
+ I used a method outlined in a [blog post](https://towardsdatascience.com/how-to-adapt-a-multilingual-t5-model-for-a-single-language-b9f94f3d9c90) by David Dale to downsize the multilingual T5 model for French and English use cases specifically. By utilizing the giga_fren dataset, I was able to successfully reduce the total number of tokens and decrease both the model and tokenizer sizes by 38% and 80% respectively.
20
+
21
+ ### Model Description
22
+
23
+ - **Developed by:** Korventenn
24
+ - **Model type:** mt5
25
+ - **Language(s) (NLP):** French and English
26
+ - **License:** Apache 2.0
27
+ - **Generated from model:** mt5-large
28
+
29
+ ### Model Sources [optional]
30
+
31
+ <!-- Provide the basic links for the model. -->
32
+
33
+ - **Repository:** https://colab.research.google.com/drive/1ag0u1WKdvuBeYTz1TrPAGucumiaYmqeW?usp=sharing
34
+
35
+ ## Uses
36
+
37
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
38
+ You can use the raw model for any sequence to sequence task that is focused on either french, english or both.
39
+
40
+
41
+ ## How to Get Started with the Model
42
+
43
+ Use the code below to get started with the model.
44
+ ```
45
+ from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
46
+
47
+ tokenizer = AutoTokenizer.from_pretrained("Korventenn/fr_en-t5-small")
48
+
49
+ model = AutoModelForSeq2SeqLM.from_pretrained("Korventenn/fr_en-t5-small")
50
+ ```
51
+
52
+ ### Training Data
53
+
54
+ <!-- This should link to a Data Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
55
+
56
+ [giga_fren](https://huggingface.co/datasets/giga_fren)
57
+ [opus100](https://huggingface.co/datasets/opus100)