minhdang commited on
Commit
74780e2
1 Parent(s): a3c6452

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +8 -8
app.py CHANGED
@@ -62,7 +62,7 @@ MODELS = sorted(list(models.keys()))
62
 
63
 
64
  def generate_prompt_with_history(
65
- text, image, history, vl_chat_processor, tokenizer, max_length=2048
66
  ):
67
  """
68
  Generate a prompt with history for the deepseek application.
@@ -85,8 +85,8 @@ def generate_prompt_with_history(
85
  # Initialize conversation
86
  conversation = vl_chat_processor.new_chat_template()
87
 
88
- if history:
89
- conversation.messages = history
90
 
91
  if image is not None:
92
  if "<image_placeholder>" not in text:
@@ -203,7 +203,7 @@ def predict(
203
  text,
204
  image,
205
  chatbot,
206
- history,
207
  top_p,
208
  temperature,
209
  repetition_penalty,
@@ -244,7 +244,7 @@ def predict(
244
  conversation = generate_prompt_with_history(
245
  text,
246
  image,
247
- history,
248
  vl_chat_processor,
249
  tokenizer,
250
  max_length=max_context_length_tokens,
@@ -291,7 +291,7 @@ def retry(
291
  text,
292
  image,
293
  chatbot,
294
- history,
295
  top_p,
296
  temperature,
297
  repetition_penalty,
@@ -313,7 +313,7 @@ def retry(
313
  text,
314
  image,
315
  chatbot,
316
- history,
317
  top_p,
318
  temperature,
319
  repetition_penalty,
@@ -452,7 +452,7 @@ def build_demo(MODELS):
452
  input_text,
453
  input_image,
454
  chatbot,
455
- history,
456
  top_p,
457
  temperature,
458
  repetition_penalty,
 
62
 
63
 
64
  def generate_prompt_with_history(
65
+ text, image, vl_chat_processor, tokenizer, max_length=2048
66
  ):
67
  """
68
  Generate a prompt with history for the deepseek application.
 
85
  # Initialize conversation
86
  conversation = vl_chat_processor.new_chat_template()
87
 
88
+ # if history:
89
+ # conversation.messages = history
90
 
91
  if image is not None:
92
  if "<image_placeholder>" not in text:
 
203
  text,
204
  image,
205
  chatbot,
206
+ # history,
207
  top_p,
208
  temperature,
209
  repetition_penalty,
 
244
  conversation = generate_prompt_with_history(
245
  text,
246
  image,
247
+ # history,
248
  vl_chat_processor,
249
  tokenizer,
250
  max_length=max_context_length_tokens,
 
291
  text,
292
  image,
293
  chatbot,
294
+ # history,
295
  top_p,
296
  temperature,
297
  repetition_penalty,
 
313
  text,
314
  image,
315
  chatbot,
316
+ # history,
317
  top_p,
318
  temperature,
319
  repetition_penalty,
 
452
  input_text,
453
  input_image,
454
  chatbot,
455
+ # history,
456
  top_p,
457
  temperature,
458
  repetition_penalty,