Elgene commited on
Commit
fffe8df
1 Parent(s): adbde78

remove jpeg conversion

Browse files
Files changed (1) hide show
  1. main.py +25 -9
main.py CHANGED
@@ -31,16 +31,33 @@ model.to(device)
31
  task_prompt = "<s_cord-v2>"
32
  decoder_input_ids = processor.tokenizer(task_prompt, add_special_tokens=False, return_tensors="pt").input_ids
33
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
34
  def generateOutput(fileData):
35
  pil_image = Image.open(BytesIO(fileData))
36
- resized_image = pil_image.resize((800, 600)).convert('RGB')
37
- rgb_image = Image.new('RGB', resized_image.size)
38
- rgb_image.paste(resized_image)
39
- output_buffer = BytesIO()
40
- rgb_image.save(output_buffer, format="JPEG", quality = 100)
41
- jpeg_image = Image.open(BytesIO(output_buffer.getvalue()))
42
-
43
- pixel_values = processor(jpeg_image, return_tensors="pt").pixel_values
44
  outputs = model.generate(
45
  pixel_values.to(device),
46
  decoder_input_ids=decoder_input_ids.to(device),
@@ -61,4 +78,3 @@ async def analyze_image(file: UploadFile = File(...)):
61
  sequence = sequence.replace(processor.tokenizer.eos_token, "").replace(processor.tokenizer.pad_token, "")
62
  sequence = re.sub(r"<.*?>", "", sequence, count=1).strip() # remove first task start token
63
  return processor.token2json(sequence)
64
-
 
31
  task_prompt = "<s_cord-v2>"
32
  decoder_input_ids = processor.tokenizer(task_prompt, add_special_tokens=False, return_tensors="pt").input_ids
33
 
34
+ # def generateOutput(fileData):
35
+ # pil_image = Image.open(BytesIO(fileData))
36
+ # resized_image = pil_image.resize((800, 600)).convert('RGB')
37
+ # rgb_image = Image.new('RGB', resized_image.size)
38
+ # rgb_image.paste(resized_image)
39
+ # output_buffer = BytesIO()
40
+ # rgb_image.save(output_buffer, format="JPEG", quality = 100)
41
+ # jpeg_image = Image.open(BytesIO(output_buffer.getvalue()))
42
+
43
+ # pixel_values = processor(jpeg_image, return_tensors="pt").pixel_values
44
+ # outputs = model.generate(
45
+ # pixel_values.to(device),
46
+ # decoder_input_ids=decoder_input_ids.to(device),
47
+ # max_length=model.decoder.config.max_position_embeddings,
48
+ # pad_token_id=processor.tokenizer.pad_token_id,
49
+ # eos_token_id=processor.tokenizer.eos_token_id,
50
+ # use_cache=True,
51
+ # bad_words_ids=[[processor.tokenizer.unk_token_id]],
52
+ # return_dict_in_generate=True,
53
+ # )
54
+ # return outputs
55
+
56
  def generateOutput(fileData):
57
  pil_image = Image.open(BytesIO(fileData))
58
+ pil_image.resize((800, 600))
59
+ pixel_values = processor(pil_image, return_tensors="pt").pixel_values
60
+
 
 
 
 
 
61
  outputs = model.generate(
62
  pixel_values.to(device),
63
  decoder_input_ids=decoder_input_ids.to(device),
 
78
  sequence = sequence.replace(processor.tokenizer.eos_token, "").replace(processor.tokenizer.pad_token, "")
79
  sequence = re.sub(r"<.*?>", "", sequence, count=1).strip() # remove first task start token
80
  return processor.token2json(sequence)