imthanhlv commited on
Commit
4962857
1 Parent(s): f9e87ac

added prefix for IC

Browse files
Files changed (1) hide show
  1. app.py +4 -6
app.py CHANGED
@@ -243,6 +243,8 @@ def inference(img, text, is_translate):
243
  text = clip.tokenize([text]).to(device)
244
  with torch.no_grad():
245
  prefix = clip_model.encode_text(text).to(device, dtype=torch.float32)
 
 
246
 
247
  else:
248
  if img is None:
@@ -253,13 +255,9 @@ def inference(img, text, is_translate):
253
 
254
  with torch.no_grad():
255
  prefix = clip_model.encode_image(image).to(device, dtype=torch.float32)
 
 
256
 
257
- prefix_embed = model.clip_project(prefix).reshape(1, prefix_length, -1)
258
-
259
- if use_beam_search:
260
- generated_text_prefix = generate_beam(model, tokenizer, embed=prefix_embed)[0]
261
- else:
262
- generated_text_prefix = generate2(model, tokenizer, embed=prefix_embed)
263
  return generated_text_prefix
264
 
265
  title = "CLIP Dual encoder"
 
243
  text = clip.tokenize([text]).to(device)
244
  with torch.no_grad():
245
  prefix = clip_model.encode_text(text).to(device, dtype=torch.float32)
246
+ prefix_embed = model.clip_project(prefix).reshape(1, prefix_length, -1)
247
+ generated_text_prefix = generate_beam(model, tokenizer, embed=prefix_embed)[0]
248
 
249
  else:
250
  if img is None:
 
255
 
256
  with torch.no_grad():
257
  prefix = clip_model.encode_image(image).to(device, dtype=torch.float32)
258
+ prefix_embed = model.clip_project(prefix).reshape(1, prefix_length, -1)
259
+ generated_text_prefix = generate_beam(model, tokenizer, embed=prefix_embed, prompt="Một bức ảnh về")[0]
260
 
 
 
 
 
 
 
261
  return generated_text_prefix
262
 
263
  title = "CLIP Dual encoder"