# 注释掉本地模型相关的导入语句，因为我们只需要API调用方式
# import torch
# from transformers import Qwen2_5_VLForConditionalGeneration, AutoProcessor
# checkpoint = "Qwen/Qwen2.5-VL-7B-Instruct"
# model = Qwen2_5_VLForConditionalGeneration.from_pretrained(checkpoint, torch_dtype=torch.bfloat16, attn_implementation="flash_attention_2",device_map="auto")
# processor = AutoProcessor.from_pretrained(checkpoint)