Update with commit d25e25ee2b63ebfcd099deb689a5a7272574a10f
Browse filesSee: https://github.com/huggingface/transformers/commit/d25e25ee2b63ebfcd099deb689a5a7272574a10f
- frameworks.json +1 -0
- pipeline_tags.json +4 -0
frameworks.json
CHANGED
@@ -78,6 +78,7 @@
|
|
78 |
{"model_type":"vit_mae","pytorch":true,"tensorflow":false,"flax":false,"processor":"AutoFeatureExtractor"}
|
79 |
{"model_type":"wav2vec2","pytorch":true,"tensorflow":true,"flax":true,"processor":"AutoProcessor"}
|
80 |
{"model_type":"wavlm","pytorch":true,"tensorflow":false,"flax":false,"processor":"AutoTokenizer"}
|
|
|
81 |
{"model_type":"xlm","pytorch":true,"tensorflow":true,"flax":false,"processor":"AutoTokenizer"}
|
82 |
{"model_type":"xlm-prophetnet","pytorch":true,"tensorflow":false,"flax":false,"processor":"AutoTokenizer"}
|
83 |
{"model_type":"xlm-roberta","pytorch":true,"tensorflow":true,"flax":false,"processor":"AutoTokenizer"}
|
|
|
78 |
{"model_type":"vit_mae","pytorch":true,"tensorflow":false,"flax":false,"processor":"AutoFeatureExtractor"}
|
79 |
{"model_type":"wav2vec2","pytorch":true,"tensorflow":true,"flax":true,"processor":"AutoProcessor"}
|
80 |
{"model_type":"wavlm","pytorch":true,"tensorflow":false,"flax":false,"processor":"AutoTokenizer"}
|
81 |
+
{"model_type":"xglm","pytorch":true,"tensorflow":false,"flax":true,"processor":"AutoTokenizer"}
|
82 |
{"model_type":"xlm","pytorch":true,"tensorflow":true,"flax":false,"processor":"AutoTokenizer"}
|
83 |
{"model_type":"xlm-prophetnet","pytorch":true,"tensorflow":false,"flax":false,"processor":"AutoTokenizer"}
|
84 |
{"model_type":"xlm-roberta","pytorch":true,"tensorflow":true,"flax":false,"processor":"AutoTokenizer"}
|
pipeline_tags.json
CHANGED
@@ -194,6 +194,8 @@
|
|
194 |
{"model_class":"FlaxVisionTextDualEncoderModel","pipeline_tag":"feature-extraction","auto_class":"Flax_AutoModel"}
|
195 |
{"model_class":"FlaxWav2Vec2ForPreTraining","pipeline_tag":"pretraining","auto_class":"Flax_AutoModelForPreTraining"}
|
196 |
{"model_class":"FlaxWav2Vec2Model","pipeline_tag":"feature-extraction","auto_class":"Flax_AutoModel"}
|
|
|
|
|
197 |
{"model_class":"FunnelBaseModel","pipeline_tag":"feature-extraction","auto_class":"AutoModel"}
|
198 |
{"model_class":"FunnelForMaskedLM","pipeline_tag":"fill-mask","auto_class":"AutoModelForMaskedLM"}
|
199 |
{"model_class":"FunnelForMultipleChoice","pipeline_tag":"multiple-choice","auto_class":"AutoModelForMultipleChoice"}
|
@@ -571,6 +573,8 @@
|
|
571 |
{"model_class":"WavLMForSequenceClassification","pipeline_tag":"audio-classification","auto_class":"AutoModelForAudioClassification"}
|
572 |
{"model_class":"WavLMForXVector","pipeline_tag":"audio-xvector","auto_class":"AutoModelForAudioXVector"}
|
573 |
{"model_class":"WavLMModel","pipeline_tag":"feature-extraction","auto_class":"AutoModel"}
|
|
|
|
|
574 |
{"model_class":"XLMForMultipleChoice","pipeline_tag":"multiple-choice","auto_class":"AutoModelForMultipleChoice"}
|
575 |
{"model_class":"XLMForQuestionAnsweringSimple","pipeline_tag":"question-answering","auto_class":"AutoModelForQuestionAnswering"}
|
576 |
{"model_class":"XLMForSequenceClassification","pipeline_tag":"text-classification","auto_class":"AutoModelForSequenceClassification"}
|
|
|
194 |
{"model_class":"FlaxVisionTextDualEncoderModel","pipeline_tag":"feature-extraction","auto_class":"Flax_AutoModel"}
|
195 |
{"model_class":"FlaxWav2Vec2ForPreTraining","pipeline_tag":"pretraining","auto_class":"Flax_AutoModelForPreTraining"}
|
196 |
{"model_class":"FlaxWav2Vec2Model","pipeline_tag":"feature-extraction","auto_class":"Flax_AutoModel"}
|
197 |
+
{"model_class":"FlaxXGLMForCausalLM","pipeline_tag":"text-generation","auto_class":"Flax_AutoModelForCausalLM"}
|
198 |
+
{"model_class":"FlaxXGLMModel","pipeline_tag":"feature-extraction","auto_class":"Flax_AutoModel"}
|
199 |
{"model_class":"FunnelBaseModel","pipeline_tag":"feature-extraction","auto_class":"AutoModel"}
|
200 |
{"model_class":"FunnelForMaskedLM","pipeline_tag":"fill-mask","auto_class":"AutoModelForMaskedLM"}
|
201 |
{"model_class":"FunnelForMultipleChoice","pipeline_tag":"multiple-choice","auto_class":"AutoModelForMultipleChoice"}
|
|
|
573 |
{"model_class":"WavLMForSequenceClassification","pipeline_tag":"audio-classification","auto_class":"AutoModelForAudioClassification"}
|
574 |
{"model_class":"WavLMForXVector","pipeline_tag":"audio-xvector","auto_class":"AutoModelForAudioXVector"}
|
575 |
{"model_class":"WavLMModel","pipeline_tag":"feature-extraction","auto_class":"AutoModel"}
|
576 |
+
{"model_class":"XGLMForCausalLM","pipeline_tag":"text-generation","auto_class":"AutoModelForCausalLM"}
|
577 |
+
{"model_class":"XGLMModel","pipeline_tag":"feature-extraction","auto_class":"AutoModel"}
|
578 |
{"model_class":"XLMForMultipleChoice","pipeline_tag":"multiple-choice","auto_class":"AutoModelForMultipleChoice"}
|
579 |
{"model_class":"XLMForQuestionAnsweringSimple","pipeline_tag":"question-answering","auto_class":"AutoModelForQuestionAnswering"}
|
580 |
{"model_class":"XLMForSequenceClassification","pipeline_tag":"text-classification","auto_class":"AutoModelForSequenceClassification"}
|