yuangongfdu commited on
Commit
0674a08
1 Parent(s): bfe831e

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +5 -6
app.py CHANGED
@@ -45,7 +45,7 @@ def predict(audio_path, question):
45
  if __name__ == '__main__':
46
  link = "https://github.com/YuanGongND/ltu"
47
  text = "[Github]"
48
- paper_link = "https://arxiv.org/pdf/2305.10790.pdf"
49
  paper_text = "[Paper]"
50
  sample_audio_link = "https://drive.google.com/drive/folders/17yeBevX0LIS1ugt0DZDOoJolwxvncMja?usp=sharing"
51
  sample_audio_text = "[sample audios from AudioSet evaluation set]"
@@ -53,12 +53,11 @@ if __name__ == '__main__':
53
  inputs=[gr.Audio(type="filepath"),
54
  gr.Textbox(value='What can be inferred from the spoken text and sounds? Why?',
55
  label='Edit the textbox to ask your own questions!')],
56
- outputs=[gr.Textbox(label="LTU Output")],
57
  cache_examples=True,
58
- title="Demo of LTU-2 Beta",
59
- description="LTU-2 an improved version of LTU. LTU-2 is stronger in spoken text understanding and music understanding. <br>" +
60
- "LTU is authored by Yuan Gong, Alexander H. Liu, Hongyin Luo, Leonid Karlinsky, and James Glass (MIT & MIT-IBM Watson AI Lab). <br>" +
61
- "**Please note that the model is under construction and may be buggy. It is trained with some new techniques that are not described in LTU paper. I.e., using method described in LTU paper cannot reproduce this model.**<br>" +
62
  "Input should be wav file sampled at 16kHz. This demo trim input audio to 10 seconds. <br>"
63
  "**Research Demo, No Commercial Use (Due to license of LLaMA).**")
64
  demo.launch(debug=False, share=False)
 
45
  if __name__ == '__main__':
46
  link = "https://github.com/YuanGongND/ltu"
47
  text = "[Github]"
48
+ paper_link = "https://www.researchgate.net/publication/374153208_Joint_Audio_and_Speech_Understanding"
49
  paper_text = "[Paper]"
50
  sample_audio_link = "https://drive.google.com/drive/folders/17yeBevX0LIS1ugt0DZDOoJolwxvncMja?usp=sharing"
51
  sample_audio_text = "[sample audios from AudioSet evaluation set]"
 
53
  inputs=[gr.Audio(type="filepath"),
54
  gr.Textbox(value='What can be inferred from the spoken text and sounds? Why?',
55
  label='Edit the textbox to ask your own questions!')],
56
+ outputs=[gr.Textbox(label="LTU-AS Output")],
57
  cache_examples=True,
58
+ title="Demo of LTU-AS",
59
+ description="LTU-AS an improved version of LTU. LTU-AS is stronger in spoken text understanding and music understanding. " + f"<a href='{paper_link}'>{paper_text}</a> <br>" +
60
+ "LTU-AS is authored by Yuan Gong, Alexander H. Liu, Hongyin Luo, Leonid Karlinsky, and James Glass (MIT & MIT-IBM Watson AI Lab). <br>" +
 
61
  "Input should be wav file sampled at 16kHz. This demo trim input audio to 10 seconds. <br>"
62
  "**Research Demo, No Commercial Use (Due to license of LLaMA).**")
63
  demo.launch(debug=False, share=False)