sgugger commited on
Commit
02de429
1 Parent(s): 8fa5e71

Update with commit f7c618e3b0652a6a0f6e435c404288fc38d18dda

Browse files

See: https://github.com/huggingface/transformers/commit/f7c618e3b0652a6a0f6e435c404288fc38d18dda

Files changed (2) hide show
  1. frameworks.json +1 -1
  2. pipeline_tags.json +1 -0
frameworks.json CHANGED
@@ -146,7 +146,7 @@
146
  {"model_type":"videomae","pytorch":true,"tensorflow":false,"flax":false,"processor":"AutoFeatureExtractor"}
147
  {"model_type":"vilt","pytorch":true,"tensorflow":false,"flax":false,"processor":"AutoProcessor"}
148
  {"model_type":"vision-encoder-decoder","pytorch":true,"tensorflow":true,"flax":true,"processor":"AutoTokenizer"}
149
- {"model_type":"vision-text-dual-encoder","pytorch":true,"tensorflow":false,"flax":true,"processor":"AutoProcessor"}
150
  {"model_type":"visual_bert","pytorch":true,"tensorflow":false,"flax":false,"processor":"AutoTokenizer"}
151
  {"model_type":"vit","pytorch":true,"tensorflow":true,"flax":true,"processor":"AutoFeatureExtractor"}
152
  {"model_type":"vit_hybrid","pytorch":true,"tensorflow":false,"flax":false,"processor":"AutoTokenizer"}
 
146
  {"model_type":"videomae","pytorch":true,"tensorflow":false,"flax":false,"processor":"AutoFeatureExtractor"}
147
  {"model_type":"vilt","pytorch":true,"tensorflow":false,"flax":false,"processor":"AutoProcessor"}
148
  {"model_type":"vision-encoder-decoder","pytorch":true,"tensorflow":true,"flax":true,"processor":"AutoTokenizer"}
149
+ {"model_type":"vision-text-dual-encoder","pytorch":true,"tensorflow":true,"flax":true,"processor":"AutoProcessor"}
150
  {"model_type":"visual_bert","pytorch":true,"tensorflow":false,"flax":false,"processor":"AutoTokenizer"}
151
  {"model_type":"vit","pytorch":true,"tensorflow":true,"flax":true,"processor":"AutoFeatureExtractor"}
152
  {"model_type":"vit_hybrid","pytorch":true,"tensorflow":false,"flax":false,"processor":"AutoTokenizer"}
pipeline_tags.json CHANGED
@@ -770,6 +770,7 @@
770
  {"model_class":"TFViTMAEForPreTraining","pipeline_tag":"pretraining","auto_class":"TF_AutoModelForPreTraining"}
771
  {"model_class":"TFViTMAEModel","pipeline_tag":"feature-extraction","auto_class":"TF_AutoModel"}
772
  {"model_class":"TFViTModel","pipeline_tag":"feature-extraction","auto_class":"TF_AutoModel"}
 
773
  {"model_class":"TFWav2Vec2Model","pipeline_tag":"feature-extraction","auto_class":"TF_AutoModel"}
774
  {"model_class":"TFWhisperForConditionalGeneration","pipeline_tag":"automatic-speech-recognition","auto_class":"TF_AutoModelForSpeechSeq2Seq"}
775
  {"model_class":"TFWhisperModel","pipeline_tag":"feature-extraction","auto_class":"TF_AutoModel"}
 
770
  {"model_class":"TFViTMAEForPreTraining","pipeline_tag":"pretraining","auto_class":"TF_AutoModelForPreTraining"}
771
  {"model_class":"TFViTMAEModel","pipeline_tag":"feature-extraction","auto_class":"TF_AutoModel"}
772
  {"model_class":"TFViTModel","pipeline_tag":"feature-extraction","auto_class":"TF_AutoModel"}
773
+ {"model_class":"TFVisionTextDualEncoderModel","pipeline_tag":"feature-extraction","auto_class":"TF_AutoModel"}
774
  {"model_class":"TFWav2Vec2Model","pipeline_tag":"feature-extraction","auto_class":"TF_AutoModel"}
775
  {"model_class":"TFWhisperForConditionalGeneration","pipeline_tag":"automatic-speech-recognition","auto_class":"TF_AutoModelForSpeechSeq2Seq"}
776
  {"model_class":"TFWhisperModel","pipeline_tag":"feature-extraction","auto_class":"TF_AutoModel"}