bhavitvyamalik commited on
Commit
ff13355
1 Parent(s): 6088947

future scope

Browse files
Files changed (2) hide show
  1. app.py +4 -3
  2. sections/future_scope.md +4 -3
app.py CHANGED
@@ -87,12 +87,13 @@ with st.beta_expander("Article"):
87
  st.write(read_markdown("abstract.md"))
88
  st.write(read_markdown("caveats.md"))
89
  # st.write("# Methodology")
90
- # st.image(
91
- # "./misc/Multilingual-IC.png", caption="Seq2Seq model for Image-text Captioning."
92
- # )
93
  st.markdown(read_markdown("pretraining.md"))
94
  st.write(read_markdown("challenges.md"))
95
  st.write(read_markdown("social_impact.md"))
 
96
  st.write(read_markdown("references.md"))
97
  # st.write(read_markdown("checkpoints.md"))
98
  st.write(read_markdown("acknowledgements.md"))
 
87
  st.write(read_markdown("abstract.md"))
88
  st.write(read_markdown("caveats.md"))
89
  # st.write("# Methodology")
90
+ st.image(
91
+ "./misc/Multilingual-IC.png", caption="Seq2Seq model for Image-text Captioning."
92
+ )
93
  st.markdown(read_markdown("pretraining.md"))
94
  st.write(read_markdown("challenges.md"))
95
  st.write(read_markdown("social_impact.md"))
96
+ st.write(read_markdown("future_scope.md"))
97
  st.write(read_markdown("references.md"))
98
  # st.write(read_markdown("checkpoints.md"))
99
  st.write(read_markdown("acknowledgements.md"))
sections/future_scope.md CHANGED
@@ -1,4 +1,5 @@
1
  # Future scope of work
2
- We hope to improve this in the future by using:
3
- - Better translating options. Better translators (for e.g. Google Translate API, Large pre-trained seq2seq models for translation) to get more multilingual data, especially in low-resource languages.
4
- - More training time: We found that training image captioning model for a single model takes a lot of compute time and if we want
 
 
1
  # Future scope of work
2
+ We hope to improve this project in the future by using:
3
+ - Better translating options: Translation has a very huge impact on how the end model would perform. Better translators (for e.g. Google Translate API) and language specific seq2seq models for translation are able to generate better data, especially in low-resource languages.
4
+ - More training time: We found that training image captioning model for a single model takes a lot of compute time and if we want to replicate the same then the training time goes up manifold for the same number of samples.
5
+ - Accessibility: Make model deployable on hand-held devices to make it more accessible. Currently, our model is too large because of which not many will be able to access it. However, our final goal is ensure everyone can access it without any computation barriers. We got to know that JAX has an experimental converter `jax2tf`to convert JAX functions to TF. I hope we'll be able to support TFLite support for our model as well in future.