ClownRat commited on
Commit
12e9783
•
1 Parent(s): 41491b0

improve duration.

Browse files
Files changed (1) hide show
  1. app.py +2 -2
app.py CHANGED
@@ -80,7 +80,6 @@ class Chat:
80
  return state
81
 
82
  @torch.inference_mode()
83
- @spaces.GPU
84
  def generate(self, tensor: list, modals: list, prompt: str, first_run: bool, state):
85
  # TODO: support multiple turns of conversation.
86
  assert len(tensor) == len(modals)
@@ -131,6 +130,7 @@ def save_video_to_local(video_path):
131
  return filename
132
 
133
 
 
134
  def generate(image, video, first_run, state, state_, textbox_in, tensor, modals, dtype=torch.float16):
135
  flag = 1
136
  if not textbox_in:
@@ -220,7 +220,7 @@ if __name__ == '__main__':
220
  conv_mode = "llama_2"
221
  model_path = 'DAMO-NLP-SG/VideoLLaMA2-7B'
222
 
223
- handler = Chat(model_path, conv_mode=conv_mode, load_8bit=False, load_4bit=False, device='cuda')
224
  handler.model.to(dtype=torch.float16)
225
 
226
  if not os.path.exists("temp"):
 
80
  return state
81
 
82
  @torch.inference_mode()
 
83
  def generate(self, tensor: list, modals: list, prompt: str, first_run: bool, state):
84
  # TODO: support multiple turns of conversation.
85
  assert len(tensor) == len(modals)
 
130
  return filename
131
 
132
 
133
+ @spaces.GPU(duration=120)
134
  def generate(image, video, first_run, state, state_, textbox_in, tensor, modals, dtype=torch.float16):
135
  flag = 1
136
  if not textbox_in:
 
220
  conv_mode = "llama_2"
221
  model_path = 'DAMO-NLP-SG/VideoLLaMA2-7B'
222
 
223
+ handler = Chat(model_path, conv_mode=conv_mode, load_8bit=True, load_4bit=False, device='cuda')
224
  handler.model.to(dtype=torch.float16)
225
 
226
  if not os.path.exists("temp"):