anas-awadalla commited on
Commit
71de382
1 Parent(s): 96f6488
Files changed (1) hide show
  1. app.py +7 -5
app.py CHANGED
@@ -70,7 +70,7 @@ def generate(
70
  example_one_text=None,
71
  example_two_image=None,
72
  example_two_text=None,
73
- tc=True
74
  ):
75
  if not tc:
76
  raise gr.Error("Please read the terms and conditions.")
@@ -175,18 +175,20 @@ with gr.Blocks() as demo:
175
  Blog posts: #1 [An open-source framework for training vision-language models with in-context learning](https://laion.ai/blog/open-flamingo/) // #2 [OpenFlamingo v2: New Models and Enhanced Training Setup]()\n
176
  GitHub: [open_flamingo](https://github.com/mlfoundations/open_flamingo)
177
 
178
- In this demo we implement an interactive interface that showcases the in-context learning capabilities of the OpenFlamingo-9B model, a large multimodal model trained on top of MPT-7B.
 
179
  The model is trained on an interleaved mixture of text and images and is able to generate text conditioned on sequences of images/text. To safeguard against harmful generations, we detect toxic text in the model output and reject it. However, we understand that this is not a perfect solution and we encourage you to use this demo responsibly. If you find that the model is generating harmful text, please report it using this [form](https://forms.gle/StbcPvyyW2p3Pc7z6).
180
  """
181
  )
182
 
183
  with gr.Accordion("See terms and conditions"):
184
  gr.Markdown("""**Please read the following information carefully before proceeding.**
185
- [OpenFlamingo-9B](https://huggingface.co/openflamingo/OpenFlamingo-9B-vitl-mpt7b) is a **research prototype** that aims to enable users to interact with AI through both language and images. AI agents equipped with both language and visual understanding can be useful on a larger variety of tasks compared to models that communicate solely via language. By releasing an open-source research prototype, we hope to help the research community better understand the risks and limitations of modern visual-language AI models and accelerate the development of safer and more reliable methods.
186
- **Limitations.** OpenFlamingo-9B is built on top of the [MPT-7B](https://huggingface.co/mosaicml/mpt-7b) large language model developed by MosaicML. Large language models are trained on mostly unfiltered internet data, and have been shown to be able to produce toxic, unethical, inaccurate, and harmful content. On top of this, OpenFlamingo’s ability to support visual inputs creates additional risks, since it can be used in a wider variety of applications; image+text models may carry additional risks specific to multimodality. Please use discretion when assessing the accuracy or appropriateness of the model’s outputs, and be mindful before sharing its results.
 
187
  **Privacy and data collection.** This demo does NOT store any personal information on its users, and it does NOT store user queries.""")
188
  read_tc = gr.Checkbox(
189
- label="I have read and agree to the terms and conditions")
190
 
191
  with gr.Tab("📷 Image Captioning"):
192
  with gr.Row():
 
70
  example_one_text=None,
71
  example_two_image=None,
72
  example_two_text=None,
73
+ tc=False
74
  ):
75
  if not tc:
76
  raise gr.Error("Please read the terms and conditions.")
 
175
  Blog posts: #1 [An open-source framework for training vision-language models with in-context learning](https://laion.ai/blog/open-flamingo/) // #2 [OpenFlamingo v2: New Models and Enhanced Training Setup]()\n
176
  GitHub: [open_flamingo](https://github.com/mlfoundations/open_flamingo)
177
 
178
+ In this demo we implement an interactive interface that showcases the in-context learning capabilities of the OpenFlamingo-4B model, a large multimodal model trained on top of
179
+ RedPajama-INCITE-Instruct-3B-v1.
180
  The model is trained on an interleaved mixture of text and images and is able to generate text conditioned on sequences of images/text. To safeguard against harmful generations, we detect toxic text in the model output and reject it. However, we understand that this is not a perfect solution and we encourage you to use this demo responsibly. If you find that the model is generating harmful text, please report it using this [form](https://forms.gle/StbcPvyyW2p3Pc7z6).
181
  """
182
  )
183
 
184
  with gr.Accordion("See terms and conditions"):
185
  gr.Markdown("""**Please read the following information carefully before proceeding.**
186
+ [OpenFlamingo-4B](https://huggingface.co/openflamingo/OpenFlamingo-4B-vitl-rpj3b-langinstruct) is a **research prototype** that aims to enable users to interact with AI through both language and images. AI agents equipped with both language and visual understanding can be useful on a larger variety of tasks compared to models that communicate solely via language. By releasing an open-source research prototype, we hope to help the research community better understand the risks and limitations of modern visual-language AI models and accelerate the development of safer and more reliable methods.
187
+ **Limitations.** OpenFlamingo-4B is built on top of the [
188
+ RedPajama-INCITE-Instruct-3B-v1](https://huggingface.co/togethercomputer/RedPajama-INCITE-Instruct-3B-v1) large language model developed by MosaicML. Large language models are trained on mostly unfiltered internet data, and have been shown to be able to produce toxic, unethical, inaccurate, and harmful content. On top of this, OpenFlamingo’s ability to support visual inputs creates additional risks, since it can be used in a wider variety of applications; image+text models may carry additional risks specific to multimodality. Please use discretion when assessing the accuracy or appropriateness of the model’s outputs, and be mindful before sharing its results.
189
  **Privacy and data collection.** This demo does NOT store any personal information on its users, and it does NOT store user queries.""")
190
  read_tc = gr.Checkbox(
191
+ label="I have read and agree to the terms and conditions", value=False)
192
 
193
  with gr.Tab("📷 Image Captioning"):
194
  with gr.Row():