Upload folder using huggingface_hub
Browse files
modeling_internvl_chat.py
CHANGED
@@ -375,7 +375,7 @@ class InternVLChatModel(PreTrainedModel):
|
|
375 |
vit_embeds = self.mlp1(vit_embeds)
|
376 |
return vit_embeds
|
377 |
|
378 |
-
def chat(self, tokenizer, pixel_values, question, generation_config,
|
379 |
IMG_START_TOKEN='<img>', IMG_END_TOKEN='</img>', IMG_CONTEXT_TOKEN='<IMG_CONTEXT>'):
|
380 |
|
381 |
img_context_token_id = tokenizer.convert_tokens_to_ids(IMG_CONTEXT_TOKEN)
|
@@ -384,8 +384,15 @@ class InternVLChatModel(PreTrainedModel):
|
|
384 |
from .conversation import get_conv_template
|
385 |
|
386 |
template = get_conv_template(self.template)
|
387 |
-
|
388 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
389 |
template.append_message(template.roles[1], None)
|
390 |
query = template.get_prompt()
|
391 |
model_inputs = tokenizer(query, return_tensors='pt')
|
@@ -399,9 +406,11 @@ class InternVLChatModel(PreTrainedModel):
|
|
399 |
**generation_config
|
400 |
)
|
401 |
response = tokenizer.batch_decode(generation_output, skip_special_tokens=True)[0]
|
402 |
-
|
403 |
-
|
404 |
-
|
|
|
|
|
405 |
|
406 |
@torch.no_grad()
|
407 |
def generate(
|
|
|
375 |
vit_embeds = self.mlp1(vit_embeds)
|
376 |
return vit_embeds
|
377 |
|
378 |
+
def chat(self, tokenizer, pixel_values, question, generation_config, history=None, return_history=False,
|
379 |
IMG_START_TOKEN='<img>', IMG_END_TOKEN='</img>', IMG_CONTEXT_TOKEN='<IMG_CONTEXT>'):
|
380 |
|
381 |
img_context_token_id = tokenizer.convert_tokens_to_ids(IMG_CONTEXT_TOKEN)
|
|
|
384 |
from .conversation import get_conv_template
|
385 |
|
386 |
template = get_conv_template(self.template)
|
387 |
+
if history is None:
|
388 |
+
history = []
|
389 |
+
image_tokens = IMG_START_TOKEN + IMG_CONTEXT_TOKEN * self.num_image_token + IMG_END_TOKEN
|
390 |
+
question = image_tokens + '\n' + question
|
391 |
+
else:
|
392 |
+
for (old_question, old_answer) in history:
|
393 |
+
template.append_message(template.roles[0], old_question)
|
394 |
+
template.append_message(template.roles[1], old_answer)
|
395 |
+
template.append_message(template.roles[0], question)
|
396 |
template.append_message(template.roles[1], None)
|
397 |
query = template.get_prompt()
|
398 |
model_inputs = tokenizer(query, return_tensors='pt')
|
|
|
406 |
**generation_config
|
407 |
)
|
408 |
response = tokenizer.batch_decode(generation_output, skip_special_tokens=True)[0]
|
409 |
+
history.append((question, response))
|
410 |
+
if return_history:
|
411 |
+
return response, history
|
412 |
+
else:
|
413 |
+
return response
|
414 |
|
415 |
@torch.no_grad()
|
416 |
def generate(
|
runs/Feb18_13-14-12_SH-IDC1-10-140-37-8/events.out.tfevents.1708234550.SH-IDC1-10-140-37-8.124298.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f0813499a41ad34fcc8faf77fee66dc139237be21fbd16e9abfaab6d39aebb8a
|
3 |
+
size 8764
|
runs/Feb18_14-05-57_SH-IDC1-10-140-37-8/events.out.tfevents.1708237619.SH-IDC1-10-140-37-8.185381.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ecf844c1015f2b40fa8e152da5dce2540ea0f210cbcc59c30e0bf70be62d85f7
|
3 |
+
size 507325
|