Ezi commited on
Commit
425a0dd
1 Parent(s): 3429083

Model Card

Browse files

Hi!👋
This PR has a some additional information for the model card, based on the format we are using as part of our effort to standardise model cards at Hugging Face. Feel free to merge if you are ok with the changes! (cc

@Marissa



@Meg

)

Files changed (1) hide show
  1. README.md +106 -2
README.md CHANGED
@@ -14,9 +14,113 @@ license: mit
14
 
15
  # hebrew-bad_wiki-gpt_neo-tiny
16
 
17
- Hebrew nonsense generation model which produces really bad wiki-abstract text. This model was fined tuned upon [hebrew-gpt_neo-tiny](https://huggingface.co/Norod78/hebrew-gpt_neo-tiny) which was previously trained using [EleutherAI's gpt-neo](https://github.com/EleutherAI/gpt-neo). Fine-tuning on the wiki-absract text was done using [@minimaxir](https://twitter.com/minimaxir)'s [aitextgen](https://github.com/minimaxir/aitextgen).
 
 
 
 
 
 
 
18
 
19
- ## Datasets
 
20
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
21
  [Hebrew Wikipedia Dump](https://dumps.wikimedia.org/hewiki/latest/) (hewiki abstract) from May 2020
22
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
14
 
15
  # hebrew-bad_wiki-gpt_neo-tiny
16
 
17
+ ## Table of Contents
18
+ - [Model Details](#model-details)
19
+ - [Uses](#uses)
20
+ - [Risks, Limitations and Biases](#risks-limitations-and-biases)
21
+ - [Training](#training)
22
+ - [Evaluation](#evaluation)
23
+ - [Environmental Impact](#environmental-impact)
24
+ - [How to Get Started With the Model](#how-to-get-started-with-the-model)
25
 
26
+ ## Model Details
27
+ **Model Description:**
28
 
29
+ The model developer notes that the model is
30
+ > Hebrew nonsense generation model which produces really bad wiki-abstract text.
31
+
32
+
33
+ - **Developed by:** [Doron Adler](https://github.com/Norod)
34
+ - **Model Type:** Text Generation
35
+ - **Language(s):** Hebrew
36
+ - **License:** MIT
37
+ - **Resources for more information:**
38
+ - [GitHub Repo](https://github.com/Norod/hebrew-gpt_neo)
39
+ - [HuggingFace Space](https://huggingface.co/spaces/Norod78/Hebrew-GPT-Neo-Small)
40
+
41
+
42
+ ## Uses
43
+
44
+ #### Direct Use
45
+
46
+ This model can be used for text generation.
47
+
48
+ #### Misuse and Out-of-scope Use
49
+
50
+
51
+ ## Risks, Limitations and Biases
52
+ **CONTENT WARNING: Readers should be aware this section contains content that is disturbing, offensive, and can propagate historical and current stereotypes.**
53
+
54
+ Significant research has explored bias and fairness issues with language models (see, e.g., [Sheng et al. (2021)](https://aclanthology.org/2021.acl-long.330.pdf) and [Bender et al. (2021)](https://dl.acm.org/doi/pdf/10.1145/3442188.3445922)).
55
+
56
+
57
+
58
+
59
+ ## Training
60
+
61
+
62
+ #### Training Data
63
  [Hebrew Wikipedia Dump](https://dumps.wikimedia.org/hewiki/latest/) (hewiki abstract) from May 2020
64
 
65
+
66
+
67
+
68
+ #### Training Procedure
69
+
70
+
71
+ This model was fined tuned upon [hebrew-gpt_neo-tiny](https://huggingface.co/Norod78/hebrew-gpt_neo-tiny) which was previously trained using [EleutherAI's gpt-neo](https://github.com/EleutherAI/gpt-neo).
72
+
73
+ Fine-tuning on the wiki-absract text was done using [@minimaxir](https://twitter.com/minimaxir)'s [aitextgen](https://github.com/minimaxir/aitextgen).
74
+
75
+
76
+
77
+ ## Evaluation
78
+
79
+
80
+ #### Configs
81
+
82
+ Model configs for the hebrew-gpt_neo-tiny is available on the [hebrew-gpt_neo model github](https://github.com/Norod/hebrew-gpt_neo/tree/main/hebrew-gpt_neo-tiny/configs)
83
+
84
+ * **Activation Function:** gelu
85
+ * **Number_Head:** 12
86
+ * **Number_Vocab:** 50257
87
+ * **Train batch size:** 250
88
+ * **Eval batch size:** 64
89
+ * **Predict batch size:** 1
90
+
91
+
92
+
93
+
94
+ ## Environmental Impact
95
+
96
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). We present the hardware type based on the [associated paper](https://arxiv.org/pdf/2105.09680.pdf).
97
+
98
+
99
+ - **Hardware Type:** [More information needed]
100
+
101
+ - **Hours used:** Unknown
102
+
103
+ - **Cloud Provider:** GCP tpu-v8s
104
+
105
+ - **Compute Region:** europe-west4
106
+
107
+ - **Carbon Emitted:** [More information needed]
108
+
109
+
110
+ ## How to Get Started With the Model
111
+
112
+ A Google Colab Notebook is also available [here](https://colab.research.google.com/github/Norod/hebrew-gpt_neo/blob/main/hebrew-gpt_neo-tiny/Norod78_hebrew_gpt_neo_tiny_Colab.ipynb)
113
+
114
+
115
+ ​​
116
+ ```
117
+ from transformers import AutoTokenizer, AutoModelForCausalLM
118
+
119
+ tokenizer = AutoTokenizer.from_pretrained("Norod78/hebrew-bad_wiki-gpt_neo-tiny")
120
+
121
+ model = AutoModelForCausalLM.from_pretrained("Norod78/hebrew-bad_wiki-gpt_neo-tiny")
122
+
123
+
124
+ ```
125
+
126
+