sgugger commited on
Commit
646d68f
1 Parent(s): 16747b4

Update with commit adc0ff25028d29af30386f2d7d3f85e290fbef57

Browse files

See: https://github.com/huggingface/transformers/commit/adc0ff25028d29af30386f2d7d3f85e290fbef57

Files changed (2) hide show
  1. frameworks.json +1 -0
  2. pipeline_tags.json +2 -0
frameworks.json CHANGED
@@ -13,6 +13,7 @@
13
  {"model_type":"convbert","pytorch":true,"tensorflow":true,"flax":false,"processor":"AutoTokenizer"}
14
  {"model_type":"convnext","pytorch":true,"tensorflow":true,"flax":false,"processor":"AutoFeatureExtractor"}
15
  {"model_type":"ctrl","pytorch":true,"tensorflow":true,"flax":false,"processor":"AutoTokenizer"}
 
16
  {"model_type":"data2vec-audio","pytorch":true,"tensorflow":false,"flax":false,"processor":"AutoFeatureExtractor"}
17
  {"model_type":"data2vec-text","pytorch":true,"tensorflow":false,"flax":false,"processor":"AutoTokenizer"}
18
  {"model_type":"data2vec-vision","pytorch":true,"tensorflow":true,"flax":false,"processor":"AutoFeatureExtractor"}
 
13
  {"model_type":"convbert","pytorch":true,"tensorflow":true,"flax":false,"processor":"AutoTokenizer"}
14
  {"model_type":"convnext","pytorch":true,"tensorflow":true,"flax":false,"processor":"AutoFeatureExtractor"}
15
  {"model_type":"ctrl","pytorch":true,"tensorflow":true,"flax":false,"processor":"AutoTokenizer"}
16
+ {"model_type":"cvt","pytorch":true,"tensorflow":false,"flax":false,"processor":"AutoFeatureExtractor"}
17
  {"model_type":"data2vec-audio","pytorch":true,"tensorflow":false,"flax":false,"processor":"AutoFeatureExtractor"}
18
  {"model_type":"data2vec-text","pytorch":true,"tensorflow":false,"flax":false,"processor":"AutoTokenizer"}
19
  {"model_type":"data2vec-vision","pytorch":true,"tensorflow":true,"flax":false,"processor":"AutoFeatureExtractor"}
pipeline_tags.json CHANGED
@@ -66,6 +66,8 @@
66
  {"model_class":"ConvBertModel","pipeline_tag":"feature-extraction","auto_class":"AutoModel"}
67
  {"model_class":"ConvNextForImageClassification","pipeline_tag":"image-classification","auto_class":"AutoModelForImageClassification"}
68
  {"model_class":"ConvNextModel","pipeline_tag":"feature-extraction","auto_class":"AutoModel"}
 
 
69
  {"model_class":"DPRQuestionEncoder","pipeline_tag":"feature-extraction","auto_class":"AutoModel"}
70
  {"model_class":"DPTModel","pipeline_tag":"feature-extraction","auto_class":"AutoModel"}
71
  {"model_class":"Data2VecAudioForAudioFrameClassification","pipeline_tag":"audio-frame-classification","auto_class":"AutoModelForAudioFrameClassification"}
 
66
  {"model_class":"ConvBertModel","pipeline_tag":"feature-extraction","auto_class":"AutoModel"}
67
  {"model_class":"ConvNextForImageClassification","pipeline_tag":"image-classification","auto_class":"AutoModelForImageClassification"}
68
  {"model_class":"ConvNextModel","pipeline_tag":"feature-extraction","auto_class":"AutoModel"}
69
+ {"model_class":"CvtForImageClassification","pipeline_tag":"image-classification","auto_class":"AutoModelForImageClassification"}
70
+ {"model_class":"CvtModel","pipeline_tag":"feature-extraction","auto_class":"AutoModel"}
71
  {"model_class":"DPRQuestionEncoder","pipeline_tag":"feature-extraction","auto_class":"AutoModel"}
72
  {"model_class":"DPTModel","pipeline_tag":"feature-extraction","auto_class":"AutoModel"}
73
  {"model_class":"Data2VecAudioForAudioFrameClassification","pipeline_tag":"audio-frame-classification","auto_class":"AutoModelForAudioFrameClassification"}