image processor not included in Transformers?

#1
by JMog - opened

I am using Transformers transformers-4.48.3, however when running the example from https://huggingface.co/docs/transformers/main/model_doc/superglue, i get the following error:

ValueError: Unrecognized image processor in magic-leap-community/superglue_outdoor. Should have a image_processor_type key in its preprocessor_config.json of config.json, or one of the following model_type keys in its config.json: align, aria, beit, bit, blip, blip-2, bridgetower, chameleon, chinese_clip, clip, clipseg, conditional_detr, convnext, convnextv2, cvt, data2vec-vision, deformable_detr, deit, depth_anything, deta, detr, dinat, dinov2, donut-swin, dpt, efficientformer, efficientnet, flava, focalnet, fuyu, git, glpn, grounding-dino, groupvit, hiera, idefics, idefics2, idefics3, ijepa, imagegpt, instructblip, instructblipvideo, kosmos-2, layoutlmv2, layoutlmv3, levit, llava, llava_next, llava_next_video, llava_onevision, mask2former, maskformer, mgp-str, mllama, mobilenet_v1, mobilenet_v2, mobilevit, mobilevitv2, nat, nougat, oneformer, owlv2, owlvit, paligemma, perceiver, pix2struct, pixtral, poolformer, pvt, pvt_v2, qwen2_vl, regnet, resnet, rt_detr, sam, segformer, seggpt, siglip, swiftformer, swin, swin2sr, swinv2, table-transformer, timesformer, timm_wrapper, tvlt, tvp, udop, upernet, van, videomae, vilt, vipllava, vit, vit_hybrid, vit_mae, vit_msn, vitmatte, xclip, yolos, zoedepth

This leads me to believe that it is not actually implemented properly in transformers? anyone who has managed to work around this?

Magic Leap Community org

Hi, SuperGlue's PR has been merged on main but there is no release including it yet, you can use it by checking out this specific commit and install transformers from source to use SuperGlue. Sorry for the inconvenience

Sign up or log in to comment