gsarti commited on
Commit
449ac0a
β€’
1 Parent(s): 08d76ce

Added NLLB preset

Browse files
Files changed (2) hide show
  1. app.py +23 -10
  2. presets.py +10 -1
app.py CHANGED
@@ -24,7 +24,8 @@ from presets import (
24
  set_chatml_preset,
25
  set_cora_preset,
26
  set_default_preset,
27
- set_mmt_preset,
 
28
  set_towerinstruct_preset,
29
  set_zephyr_preset,
30
  set_gemma_preset,
@@ -242,22 +243,24 @@ with gr.Blocks(css=custom_css) as demo:
242
  "Preset for the <a href='https://huggingface.co/gsarti/cora_mgen' target='_blank'>CORA Multilingual QA</a> model.\nUses special templates for inputs."
243
  )
244
  with gr.Column():
245
- zephyr_preset = gr.Button("Zephyr Template", variant="secondary", interactive=False)
246
  gr.Markdown(
247
- "Preset for models using the <a href='https://huggingface.co/stabilityai/stablelm-2-zephyr-1_6b' target='_blank'>StableLM 2 Zephyr conversational template</a>.\nUses <code><|system|></code>, <code><|user|></code> and <code><|assistant|></code> special tokens."
248
  )
249
  with gr.Row(equal_height=True):
250
  with gr.Column(scale=1):
251
- multilingual_mt_template = gr.Button(
252
- "Multilingual MT", variant="secondary"
253
  )
254
  gr.Markdown(
255
- "Preset for multilingual MT models such as <a href='https://huggingface.co/facebook/nllb-200-distilled-600M' target='_blank'>NLLB</a> and <a href='https://huggingface.co/facebook/mbart-large-50-many-to-many-mmt' target='_blank'>mBART</a> using language tags."
256
  )
257
  with gr.Column(scale=1):
258
- chatml_template = gr.Button("Qwen ChatML", variant="secondary")
 
 
259
  gr.Markdown(
260
- "Preset for models using the <a href='https://github.com/MicrosoftDocs/azure-docs/blob/main/articles/ai-services/openai/includes/chat-markup-language.md' target='_blank'>ChatML conversational template</a>.\nUses <code><|im_start|></code>, <code><|im_end|></code> special tokens."
261
  )
262
  with gr.Column(scale=1):
263
  towerinstruct_template = gr.Button(
@@ -267,6 +270,11 @@ with gr.Blocks(css=custom_css) as demo:
267
  "Preset for models using the <a href='https://huggingface.co/Unbabel/TowerInstruct-7B-v0.1' target='_blank'>Unbabel TowerInstruct</a> conversational template.\nUses <code><|im_start|></code>, <code><|im_end|></code> special tokens."
268
  )
269
  with gr.Row(equal_height=True):
 
 
 
 
 
270
  with gr.Column(scale=1):
271
  gemma_template = gr.Button(
272
  "Gemma Chat Template", variant="secondary", interactive=False
@@ -624,8 +632,13 @@ with gr.Blocks(css=custom_css) as demo:
624
  ],
625
  ).success(preload_model, inputs=load_model_args, cancels=load_model_event)
626
 
627
- multilingual_mt_template.click(**reset_kwargs).then(
628
- set_mmt_preset,
 
 
 
 
 
629
  outputs=[model_name_or_path, input_template, output_template, tokenizer_kwargs],
630
  ).success(preload_model, inputs=load_model_args, cancels=load_model_event)
631
 
 
24
  set_chatml_preset,
25
  set_cora_preset,
26
  set_default_preset,
27
+ set_mbart_mmt_preset,
28
+ set_nllb_mmt_preset,
29
  set_towerinstruct_preset,
30
  set_zephyr_preset,
31
  set_gemma_preset,
 
243
  "Preset for the <a href='https://huggingface.co/gsarti/cora_mgen' target='_blank'>CORA Multilingual QA</a> model.\nUses special templates for inputs."
244
  )
245
  with gr.Column():
246
+ chatml_template = gr.Button("Qwen ChatML", variant="secondary")
247
  gr.Markdown(
248
+ "Preset for models using the <a href='https://github.com/MicrosoftDocs/azure-docs/blob/main/articles/ai-services/openai/includes/chat-markup-language.md' target='_blank'>ChatML conversational template</a>.\nUses <code><|im_start|></code>, <code><|im_end|></code> special tokens."
249
  )
250
  with gr.Row(equal_height=True):
251
  with gr.Column(scale=1):
252
+ mbart_mmt_template = gr.Button(
253
+ "mBART Multilingual MT", variant="secondary"
254
  )
255
  gr.Markdown(
256
+ "Preset for the <a href='https://huggingface.co/facebook/mbart-large-50-many-to-many-mmt' target='_blank'>mBART Many-to-Many</a> multilingual MT model using language tags (default: English to French)."
257
  )
258
  with gr.Column(scale=1):
259
+ nllb_mmt_template = gr.Button(
260
+ "NLLB Multilingual MT", variant="secondary"
261
+ )
262
  gr.Markdown(
263
+ "Preset for the <a href='https://huggingface.co/facebook/nllb-200-distilled-600M' target='_blank'>NLLB 600M</a> multilingual MT model using language tags (default: English to French)."
264
  )
265
  with gr.Column(scale=1):
266
  towerinstruct_template = gr.Button(
 
270
  "Preset for models using the <a href='https://huggingface.co/Unbabel/TowerInstruct-7B-v0.1' target='_blank'>Unbabel TowerInstruct</a> conversational template.\nUses <code><|im_start|></code>, <code><|im_end|></code> special tokens."
271
  )
272
  with gr.Row(equal_height=True):
273
+ with gr.Column():
274
+ zephyr_preset = gr.Button("Zephyr Template", variant="secondary", interactive=False)
275
+ gr.Markdown(
276
+ "Preset for models using the <a href='https://huggingface.co/stabilityai/stablelm-2-zephyr-1_6b' target='_blank'>StableLM 2 Zephyr conversational template</a>.\nUses <code><|system|></code>, <code><|user|></code> and <code><|assistant|></code> special tokens."
277
+ )
278
  with gr.Column(scale=1):
279
  gemma_template = gr.Button(
280
  "Gemma Chat Template", variant="secondary", interactive=False
 
632
  ],
633
  ).success(preload_model, inputs=load_model_args, cancels=load_model_event)
634
 
635
+ mbart_mmt_template.click(**reset_kwargs).then(
636
+ set_mbart_mmt_preset,
637
+ outputs=[model_name_or_path, input_template, output_template, tokenizer_kwargs],
638
+ ).success(preload_model, inputs=load_model_args, cancels=load_model_event)
639
+
640
+ nllb_mmt_template.click(**reset_kwargs).then(
641
+ set_nllb_mmt_preset,
642
  outputs=[model_name_or_path, input_template, output_template, tokenizer_kwargs],
643
  ).success(preload_model, inputs=load_model_args, cancels=load_model_event)
644
 
presets.py CHANGED
@@ -46,7 +46,7 @@ def set_chatml_preset():
46
  )
47
 
48
 
49
- def set_mmt_preset():
50
  return (
51
  "facebook/mbart-large-50-one-to-many-mmt", # model_name_or_path
52
  "{context} {current}", # input_template
@@ -55,6 +55,15 @@ def set_mmt_preset():
55
  )
56
 
57
 
 
 
 
 
 
 
 
 
 
58
  def set_towerinstruct_preset():
59
  return (
60
  "Unbabel/TowerInstruct-7B-v0.1", # model_name_or_path
 
46
  )
47
 
48
 
49
+ def set_mbart_mmt_preset():
50
  return (
51
  "facebook/mbart-large-50-one-to-many-mmt", # model_name_or_path
52
  "{context} {current}", # input_template
 
55
  )
56
 
57
 
58
+ def set_nllb_mmt_preset():
59
+ return (
60
+ "facebook/nllb-200-distilled-600M", # model_name_or_path
61
+ "{context} {current}", # input_template
62
+ "{context} {current}", # output_template
63
+ '{\n\t"src_lang": "eng_Latn",\n\t"tgt_lang": "fra_Latn"\n}', # tokenizer_kwargs
64
+ )
65
+
66
+
67
  def set_towerinstruct_preset():
68
  return (
69
  "Unbabel/TowerInstruct-7B-v0.1", # model_name_or_path