Spaces:
Sleeping
Sleeping
remove jpeg conversion
Browse files
main.py
CHANGED
@@ -31,16 +31,33 @@ model.to(device)
|
|
31 |
task_prompt = "<s_cord-v2>"
|
32 |
decoder_input_ids = processor.tokenizer(task_prompt, add_special_tokens=False, return_tensors="pt").input_ids
|
33 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
34 |
def generateOutput(fileData):
|
35 |
pil_image = Image.open(BytesIO(fileData))
|
36 |
-
|
37 |
-
|
38 |
-
|
39 |
-
output_buffer = BytesIO()
|
40 |
-
rgb_image.save(output_buffer, format="JPEG", quality = 100)
|
41 |
-
jpeg_image = Image.open(BytesIO(output_buffer.getvalue()))
|
42 |
-
|
43 |
-
pixel_values = processor(jpeg_image, return_tensors="pt").pixel_values
|
44 |
outputs = model.generate(
|
45 |
pixel_values.to(device),
|
46 |
decoder_input_ids=decoder_input_ids.to(device),
|
@@ -61,4 +78,3 @@ async def analyze_image(file: UploadFile = File(...)):
|
|
61 |
sequence = sequence.replace(processor.tokenizer.eos_token, "").replace(processor.tokenizer.pad_token, "")
|
62 |
sequence = re.sub(r"<.*?>", "", sequence, count=1).strip() # remove first task start token
|
63 |
return processor.token2json(sequence)
|
64 |
-
|
|
|
31 |
task_prompt = "<s_cord-v2>"
|
32 |
decoder_input_ids = processor.tokenizer(task_prompt, add_special_tokens=False, return_tensors="pt").input_ids
|
33 |
|
34 |
+
# def generateOutput(fileData):
|
35 |
+
# pil_image = Image.open(BytesIO(fileData))
|
36 |
+
# resized_image = pil_image.resize((800, 600)).convert('RGB')
|
37 |
+
# rgb_image = Image.new('RGB', resized_image.size)
|
38 |
+
# rgb_image.paste(resized_image)
|
39 |
+
# output_buffer = BytesIO()
|
40 |
+
# rgb_image.save(output_buffer, format="JPEG", quality = 100)
|
41 |
+
# jpeg_image = Image.open(BytesIO(output_buffer.getvalue()))
|
42 |
+
|
43 |
+
# pixel_values = processor(jpeg_image, return_tensors="pt").pixel_values
|
44 |
+
# outputs = model.generate(
|
45 |
+
# pixel_values.to(device),
|
46 |
+
# decoder_input_ids=decoder_input_ids.to(device),
|
47 |
+
# max_length=model.decoder.config.max_position_embeddings,
|
48 |
+
# pad_token_id=processor.tokenizer.pad_token_id,
|
49 |
+
# eos_token_id=processor.tokenizer.eos_token_id,
|
50 |
+
# use_cache=True,
|
51 |
+
# bad_words_ids=[[processor.tokenizer.unk_token_id]],
|
52 |
+
# return_dict_in_generate=True,
|
53 |
+
# )
|
54 |
+
# return outputs
|
55 |
+
|
56 |
def generateOutput(fileData):
|
57 |
pil_image = Image.open(BytesIO(fileData))
|
58 |
+
pil_image.resize((800, 600))
|
59 |
+
pixel_values = processor(pil_image, return_tensors="pt").pixel_values
|
60 |
+
|
|
|
|
|
|
|
|
|
|
|
61 |
outputs = model.generate(
|
62 |
pixel_values.to(device),
|
63 |
decoder_input_ids=decoder_input_ids.to(device),
|
|
|
78 |
sequence = sequence.replace(processor.tokenizer.eos_token, "").replace(processor.tokenizer.pad_token, "")
|
79 |
sequence = re.sub(r"<.*?>", "", sequence, count=1).strip() # remove first task start token
|
80 |
return processor.token2json(sequence)
|
|