asi commited on
Commit
4d519ee
1 Parent(s): e48f7c4

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -3
README.md CHANGED
@@ -2,7 +2,7 @@
2
  language:
3
  - fr
4
 
5
- thumbnail: https://raw.githubusercontent.com/AntoineSimoulin/gpt-fr/main/imgs/logo.png?token=ACXKU6CWEZIYA65LXAOFQQ3ASFTMG
6
  tags:
7
  - Tensorflow
8
  - PyTroch
@@ -52,13 +52,14 @@ beam_outputs = model.generate(
52
  num_return_sequences=1
53
  )
54
 
55
- print("Output:\n" + 100 * '-')
 
56
  print(tokenizer.decode(beam_outputs[0], skip_special_tokens=True))
57
  ```
58
 
59
  #### Limitations and bias
60
 
61
- Large pre-trained language models tend to reproduce the biases from the dataset used for pre-training, in particular gender discrimination. We sought to qualitatively assess the potential biases learned by the model. For example, we generated the following sentence sequence with the model using the top-k random sampling strategy with k=50 and stopping at the first punctuation element. "Ma femme/Mon mari vient d'obtenir un nouveau poste en tant qu'\_\_\_\_\_\_":
62
 
63
  The position generated for the wife are:
64
 
 
2
  language:
3
  - fr
4
 
5
+ thumbnail: https://github.com/AntoineSimoulin/gpt-fr/blob/main/imgs/logo.png?raw=true
6
  tags:
7
  - Tensorflow
8
  - PyTroch
 
52
  num_return_sequences=1
53
  )
54
 
55
+ print("Output:\
56
+ " + 100 * '-')
57
  print(tokenizer.decode(beam_outputs[0], skip_special_tokens=True))
58
  ```
59
 
60
  #### Limitations and bias
61
 
62
+ Large pre-trained language models tend to reproduce the biases from the dataset used for pre-training, in particular gender discrimination. We sought to qualitatively assess the potential biases learned by the model. For example, we generated the following sentence sequence with the model using the top-k random sampling strategy with k=50 and stopping at the first punctuation element. "Ma femme/Mon mari vient d'obtenir un nouveau poste en tant qu'\\_\\_\\_\\_\\_\\_":
63
 
64
  The position generated for the wife are:
65