nickmalhotra commited on
Commit
16da29a
1 Parent(s): 7596e2e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +22 -9
README.md CHANGED
@@ -18,7 +18,7 @@ model-index:
18
  value: 22.7
19
  name: normalized accuracy
20
  source:
21
- url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=nickmalhotra/indus_1.175B
22
  name: Open LLM Leaderboard
23
  - task:
24
  type: text-generation
@@ -108,7 +108,7 @@ model-index:
108
  # Model Card for Indus
109
 
110
  <!-- Provide a quick summary of what the model is/does. [Optional] -->
111
- The model is a single shot fine tuned Instruct LLM in Hindi and dialects
112
 
113
 
114
 
@@ -157,26 +157,32 @@ The model is a single shot fine tuned Instruct LLM in Hindi and dialects
157
  ## Model Description
158
 
159
  <!-- Provide a longer summary of what this model is/does. -->
160
- The model is a single shot fine tuned Instruct LLM in Hindi and dialects
161
 
162
  - **Developed by:** Nikhil Malhotra, Nilesh Brahme, Satish Mishra, Vinay Sharma (Makers Lab, TechMahindra)
163
  - **Model type:** Foundational Language model
164
  - **Language(s) (NLP):** hin, bho, mai, doi
165
  - **License:** other
166
- - **Parent Model:** It is the parent model on GPT architecture
167
- - **Resources for more information:** More information needed
 
168
 
169
 
170
 
171
  # Uses
172
 
173
  <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
 
 
 
 
 
174
 
175
  ## Direct Use
176
 
177
  <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
178
  <!-- If the user enters content, print that. If not, but they enter a task in the list, use that. If neither, say "more info needed." -->
179
-
180
 
181
 
182
 
@@ -184,7 +190,12 @@ The model is a single shot fine tuned Instruct LLM in Hindi and dialects
184
 
185
  <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
186
  <!-- If the user enters content, print that. If not, but they enter a task in the list, use that. If neither, say "more info needed." -->
187
-
 
 
 
 
 
188
 
189
 
190
 
@@ -192,7 +203,7 @@ The model is a single shot fine tuned Instruct LLM in Hindi and dialects
192
 
193
  <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
194
  <!-- If the user enters content, print that. If not, but they enter a task in the list, use that. If neither, say "more info needed." -->
195
-
196
 
197
 
198
 
@@ -203,12 +214,14 @@ The model is a single shot fine tuned Instruct LLM in Hindi and dialects
203
  Significant research has explored bias and fairness issues with language models
204
  (see, e.g., [Sheng et al. (2021)](https://aclanthology.org/2021.acl-long.330.pdf) and [Bender et al. (2021)](https://dl.acm.org/doi/pdf/10.1145/3442188.3445922)).
205
  Predictions generated by the model may include disturbing and harmful stereotypes across protected classes; identity characteristics; and sensitive, social, and occupational groups.
 
 
206
 
207
 
208
  ## Recommendations
209
 
210
  <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
211
-
212
 
213
 
214
 
 
18
  value: 22.7
19
  name: normalized accuracy
20
  source:
21
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=nickmalhotra/ProjectIndus
22
  name: Open LLM Leaderboard
23
  - task:
24
  type: text-generation
 
108
  # Model Card for Indus
109
 
110
  <!-- Provide a quick summary of what the model is/does. [Optional] -->
111
+ The model is a pretrained model in Hindi and dialects which is instruct tuned .
112
 
113
 
114
 
 
157
  ## Model Description
158
 
159
  <!-- Provide a longer summary of what this model is/does. -->
160
+ TThe model is a pretrained model in Hindi and dialects which is instruct tuned.
161
 
162
  - **Developed by:** Nikhil Malhotra, Nilesh Brahme, Satish Mishra, Vinay Sharma (Makers Lab, TechMahindra)
163
  - **Model type:** Foundational Language model
164
  - **Language(s) (NLP):** hin, bho, mai, doi
165
  - **License:** other
166
+ - **Parent Model:** It is a grounds up model built on GPT-2 architecture starting from tokenizer to decoder
167
+ - **Resources for more information:** https://www.techmahindra.com/en-in/innovation/the-indus-project/
168
+
169
 
170
 
171
 
172
  # Uses
173
 
174
  <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
175
+ Uses include question and answeting and conversation in Hindi and Dialects. The model would be reward tuned to be used across various industries
176
+ 1. Call center
177
+ 2. Healthcare
178
+ 3. Automotive
179
+ 4. Telecom
180
 
181
  ## Direct Use
182
 
183
  <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
184
  <!-- If the user enters content, print that. If not, but they enter a task in the list, use that. If neither, say "more info needed." -->
185
+ Direct use is as a foundationla model on Hindi and dialects
186
 
187
 
188
 
 
190
 
191
  <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
192
  <!-- If the user enters content, print that. If not, but they enter a task in the list, use that. If neither, say "more info needed." -->
193
+ Uses include question and answeting and conversation in Hindi and Dialects. The model would be reward tuned to be used across various industries
194
+ 1. Call center
195
+ 2. Healthcare
196
+ 3. Automotive
197
+ 4. Telecom
198
+
199
 
200
 
201
 
 
203
 
204
  <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
205
  <!-- If the user enters content, print that. If not, but they enter a task in the list, use that. If neither, say "more info needed." -->
206
+ Cannot be used for fill in the blanks, Multiple Q&A etc. at the moment
207
 
208
 
209
 
 
214
  Significant research has explored bias and fairness issues with language models
215
  (see, e.g., [Sheng et al. (2021)](https://aclanthology.org/2021.acl-long.330.pdf) and [Bender et al. (2021)](https://dl.acm.org/doi/pdf/10.1145/3442188.3445922)).
216
  Predictions generated by the model may include disturbing and harmful stereotypes across protected classes; identity characteristics; and sensitive, social, and occupational groups.
217
+ We have taken care across various biases by trying to remove them from training data. However since the model is a generative model, it would tend to produce hallucinations.
218
+ Any disturbing or harmful sterotype produced by the model is purely un-intentional and coincidental.
219
 
220
 
221
  ## Recommendations
222
 
223
  <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
224
+ Recommendation is to not use biases and negative connotation for the model
225
 
226
 
227