Korventenn commited on
Commit
92a43c7
1 Parent(s): dc59f7c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +54 -0
README.md CHANGED
@@ -1,3 +1,57 @@
1
  ---
2
  license: apache-2.0
 
 
 
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: apache-2.0
3
+ datasets:
4
+ - giga_fren
5
+ language:
6
+ - fr
7
+ - en
8
  ---
9
+
10
+
11
+ # Model Card for fr_en-t5-large
12
+
13
+ <!-- Provide a quick summary of what the model is/does. -->
14
+
15
+ This model has been optimized for French and English language processing while minimizing overall size. To achieve this, I only retained relevant parameters and tokens specific to these two languages, ensuring that performance remains as good as the original mt5.
16
+
17
+ ## Model Details
18
+ I used a method outlined in a [blog post](https://towardsdatascience.com/how-to-adapt-a-multilingual-t5-model-for-a-single-language-b9f94f3d9c90) by David Dale to downsize the multilingual T5 model for French and English use cases specifically. By utilizing the giga_fren dataset, I was able to successfully reduce the total number of tokens and decrease both the model and tokenizer sizes by 38% and 80% respectively.
19
+
20
+ ### Model Description
21
+
22
+ - **Developed by:** Korventenn
23
+ - **Model type:** mt5
24
+ - **Language(s) (NLP):** French and English
25
+ - **License:** Apache 2.0
26
+ - **Finetuned from model:** mt5-large
27
+
28
+ ### Model Sources [optional]
29
+
30
+ <!-- Provide the basic links for the model. -->
31
+
32
+ - **Repository:** I will make it available soon, if you have any questions you can contact me via linkedin (link in my profile)
33
+
34
+ ## Uses
35
+
36
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
37
+ You can use the raw model for any sequence to sequence task that is focused on either french, english or both.
38
+
39
+
40
+ ## How to Get Started with the Model
41
+
42
+ Use the code below to get started with the model.
43
+ ```
44
+ from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
45
+
46
+ tokenizer = AutoTokenizer.from_pretrained("Korventenn/fr_en-t5-large")
47
+
48
+ model = AutoModelForSeq2SeqLM.from_pretrained("Korventenn/fr_en-t5-large")
49
+ ```
50
+
51
+ ## Training Details
52
+
53
+ ### Training Data
54
+
55
+ <!-- This should link to a Data Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
56
+
57
+ [giga_fren](https://huggingface.co/datasets/giga_fren)