sgugger commited on
Commit
1c11bfb
1 Parent(s): 48e4ebf

Update with commit 5b40a37bc4da9dc6cd33876ce9bb3f7f48450a03

Browse files

See: https://github.com/huggingface/transformers/commit/5b40a37bc4da9dc6cd33876ce9bb3f7f48450a03

Files changed (2) hide show
  1. frameworks.json +1 -1
  2. pipeline_tags.json +2 -0
frameworks.json CHANGED
@@ -86,7 +86,7 @@
86
  {"model_type":"vision-text-dual-encoder","pytorch":true,"tensorflow":false,"flax":true,"processor":"AutoProcessor"}
87
  {"model_type":"visual_bert","pytorch":true,"tensorflow":false,"flax":false,"processor":"AutoTokenizer"}
88
  {"model_type":"vit","pytorch":true,"tensorflow":true,"flax":true,"processor":"AutoFeatureExtractor"}
89
- {"model_type":"vit_mae","pytorch":true,"tensorflow":false,"flax":false,"processor":"AutoFeatureExtractor"}
90
  {"model_type":"wav2vec2","pytorch":true,"tensorflow":true,"flax":true,"processor":"AutoProcessor"}
91
  {"model_type":"wavlm","pytorch":true,"tensorflow":false,"flax":false,"processor":"AutoTokenizer"}
92
  {"model_type":"xglm","pytorch":true,"tensorflow":false,"flax":true,"processor":"AutoTokenizer"}
 
86
  {"model_type":"vision-text-dual-encoder","pytorch":true,"tensorflow":false,"flax":true,"processor":"AutoProcessor"}
87
  {"model_type":"visual_bert","pytorch":true,"tensorflow":false,"flax":false,"processor":"AutoTokenizer"}
88
  {"model_type":"vit","pytorch":true,"tensorflow":true,"flax":true,"processor":"AutoFeatureExtractor"}
89
+ {"model_type":"vit_mae","pytorch":true,"tensorflow":true,"flax":false,"processor":"AutoFeatureExtractor"}
90
  {"model_type":"wav2vec2","pytorch":true,"tensorflow":true,"flax":true,"processor":"AutoProcessor"}
91
  {"model_type":"wavlm","pytorch":true,"tensorflow":false,"flax":false,"processor":"AutoTokenizer"}
92
  {"model_type":"xglm","pytorch":true,"tensorflow":false,"flax":true,"processor":"AutoTokenizer"}
pipeline_tags.json CHANGED
@@ -559,6 +559,8 @@
559
  {"model_class":"TFTransfoXLLMHeadModel","pipeline_tag":"text-generation","auto_class":"TF_AutoModelForCausalLM"}
560
  {"model_class":"TFTransfoXLModel","pipeline_tag":"feature-extraction","auto_class":"TF_AutoModel"}
561
  {"model_class":"TFViTForImageClassification","pipeline_tag":"image-classification","auto_class":"TF_AutoModelForImageClassification"}
 
 
562
  {"model_class":"TFViTModel","pipeline_tag":"feature-extraction","auto_class":"TF_AutoModel"}
563
  {"model_class":"TFWav2Vec2Model","pipeline_tag":"feature-extraction","auto_class":"TF_AutoModel"}
564
  {"model_class":"TFXLMForMultipleChoice","pipeline_tag":"multiple-choice","auto_class":"TF_AutoModelForMultipleChoice"}
 
559
  {"model_class":"TFTransfoXLLMHeadModel","pipeline_tag":"text-generation","auto_class":"TF_AutoModelForCausalLM"}
560
  {"model_class":"TFTransfoXLModel","pipeline_tag":"feature-extraction","auto_class":"TF_AutoModel"}
561
  {"model_class":"TFViTForImageClassification","pipeline_tag":"image-classification","auto_class":"TF_AutoModelForImageClassification"}
562
+ {"model_class":"TFViTMAEForPreTraining","pipeline_tag":"pretraining","auto_class":"TF_AutoModelForPreTraining"}
563
+ {"model_class":"TFViTMAEModel","pipeline_tag":"feature-extraction","auto_class":"TF_AutoModel"}
564
  {"model_class":"TFViTModel","pipeline_tag":"feature-extraction","auto_class":"TF_AutoModel"}
565
  {"model_class":"TFWav2Vec2Model","pipeline_tag":"feature-extraction","auto_class":"TF_AutoModel"}
566
  {"model_class":"TFXLMForMultipleChoice","pipeline_tag":"multiple-choice","auto_class":"TF_AutoModelForMultipleChoice"}