Hub Python Library documentation

추론 타입

Hugging Face's logo
Join the Hugging Face community

and get access to the augmented documentation experience

to get started

추론 타입

이 페이지에는 Hugging Face Hub에서 지원하는 타입(예: 데이터 클래스)이 나열되어 있습니다. 각 작업은 JSON 스키마를 사용하여 지정되며, 이러한 스키마에 의해서 타입이 생성됩니다. 이때 Python 요구 사항으로 인해 일부 사용자 정의가 있을 수 있습니다.

각 작업의 JSON 스키마를 확인하려면 @huggingface.js/tasks를 확인하세요.

라이브러리에서 이 부분은 아직 개발 중이며, 향후 릴리즈에서 개선될 예정입니다.

audio_classification

class huggingface_hub.AudioClassificationInput

< >

( inputs: Any parameters: Optional = None )

Inputs for Audio Classification inference

class huggingface_hub.AudioClassificationOutputElement

< >

( label: str score: float )

Outputs for Audio Classification inference

class huggingface_hub.AudioClassificationParameters

< >

( function_to_apply: Optional = None top_k: Optional = None )

Additional inference parameters Additional inference parameters for Audio Classification

audio_to_audio

class huggingface_hub.AudioToAudioInput

< >

( inputs: Any )

Inputs for Audio to Audio inference

class huggingface_hub.AudioToAudioOutputElement

< >

( blob: Any content_type: str label: str )

Outputs of inference for the Audio To Audio task A generated audio file with its label.

automatic_speech_recognition

class huggingface_hub.AutomaticSpeechRecognitionGenerationParameters

< >

( do_sample: Optional = None early_stopping: Union = None epsilon_cutoff: Optional = None eta_cutoff: Optional = None max_length: Optional = None max_new_tokens: Optional = None min_length: Optional = None min_new_tokens: Optional = None num_beam_groups: Optional = None num_beams: Optional = None penalty_alpha: Optional = None temperature: Optional = None top_k: Optional = None top_p: Optional = None typical_p: Optional = None use_cache: Optional = None )

Parametrization of the text generation process Ad-hoc parametrization of the text generation process

class huggingface_hub.AutomaticSpeechRecognitionInput

< >

( inputs: Any parameters: Optional = None )

Inputs for Automatic Speech Recognition inference

class huggingface_hub.AutomaticSpeechRecognitionOutput

< >

( text: str chunks: Optional = None )

Outputs of inference for the Automatic Speech Recognition task

class huggingface_hub.AutomaticSpeechRecognitionOutputChunk

< >

( text: str timestamps: List )

class huggingface_hub.AutomaticSpeechRecognitionParameters

< >

( generate: Optional = None return_timestamps: Optional = None )

Additional inference parameters Additional inference parameters for Automatic Speech Recognition

chat_completion

class huggingface_hub.ChatCompletionInput

< >

( messages: List model: str frequency_penalty: Optional = None logit_bias: Optional = None logprobs: Optional = None max_tokens: Optional = None n: Optional = None presence_penalty: Optional = None seed: Optional = None stop: Optional = None stream: Optional = None temperature: Optional = None tool_choice: Union = None tool_prompt: Optional = None tools: Optional = None top_logprobs: Optional = None top_p: Optional = None )

Chat Completion Input. Auto-generated from TGI specs. For more details, check out https://github.com/huggingface/huggingface.js/blob/main/packages/tasks/scripts/inference-tgi-import.ts.

class huggingface_hub.ChatCompletionInputFunctionDefinition

< >

( arguments: Any name: str description: Optional = None )

class huggingface_hub.ChatCompletionInputMessage

< >

( role: str content: Optional = None name: Optional = None tool_calls: Optional = None )

class huggingface_hub.ChatCompletionInputTool

< >

( function: ChatCompletionInputFunctionDefinition type: str )

class huggingface_hub.ChatCompletionInputToolCall

< >

( function: ChatCompletionInputFunctionDefinition id: int type: str )

class huggingface_hub.ChatCompletionInputToolTypeClass

< >

( function_name: str )

class huggingface_hub.ChatCompletionOutput

< >

( choices: List created: int id: str model: str object: str system_fingerprint: str usage: ChatCompletionOutputUsage )

Chat Completion Output. Auto-generated from TGI specs. For more details, check out https://github.com/huggingface/huggingface.js/blob/main/packages/tasks/scripts/inference-tgi-import.ts.

class huggingface_hub.ChatCompletionOutputComplete

< >

( finish_reason: str index: int message: ChatCompletionOutputMessage logprobs: Optional = None )

class huggingface_hub.ChatCompletionOutputFunctionDefinition

< >

( arguments: Any name: str description: Optional = None )

class huggingface_hub.ChatCompletionOutputLogprob

< >

( logprob: float token: str top_logprobs: List )

class huggingface_hub.ChatCompletionOutputLogprobs

< >

( content: List )

class huggingface_hub.ChatCompletionOutputMessage

< >

( role: str content: Optional = None name: Optional = None tool_calls: Optional = None )

class huggingface_hub.ChatCompletionOutputToolCall

< >

( function: ChatCompletionOutputFunctionDefinition id: int type: str )

class huggingface_hub.ChatCompletionOutputTopLogprob

< >

( logprob: float token: str )

class huggingface_hub.ChatCompletionOutputUsage

< >

( completion_tokens: int prompt_tokens: int total_tokens: int )

class huggingface_hub.ChatCompletionStreamOutput

< >

( choices: List created: int id: str model: str object: str system_fingerprint: str )

Chat Completion Stream Output. Auto-generated from TGI specs. For more details, check out https://github.com/huggingface/huggingface.js/blob/main/packages/tasks/scripts/inference-tgi-import.ts.

class huggingface_hub.ChatCompletionStreamOutputChoice

< >

( delta: ChatCompletionStreamOutputDelta index: int finish_reason: Optional = None logprobs: Optional = None )

class huggingface_hub.ChatCompletionStreamOutputDelta

< >

( role: str content: Optional = None tool_calls: Optional = None )

class huggingface_hub.ChatCompletionStreamOutputDeltaToolCall

< >

( function: ChatCompletionStreamOutputFunction id: str index: int type: str )

class huggingface_hub.ChatCompletionStreamOutputFunction

< >

( arguments: str name: Optional = None )

class huggingface_hub.ChatCompletionStreamOutputLogprob

< >

( logprob: float token: str top_logprobs: List )

class huggingface_hub.ChatCompletionStreamOutputLogprobs

< >

( content: List )

class huggingface_hub.ChatCompletionStreamOutputTopLogprob

< >

( logprob: float token: str )

depth_estimation

class huggingface_hub.DepthEstimationInput

< >

( inputs: Any parameters: Optional = None )

Inputs for Depth Estimation inference

class huggingface_hub.DepthEstimationOutput

< >

( depth: Any predicted_depth: Any )

Outputs of inference for the Depth Estimation task

document_question_answering

class huggingface_hub.DocumentQuestionAnsweringInput

< >

( inputs: DocumentQuestionAnsweringInputData parameters: Optional = None )

Inputs for Document Question Answering inference

class huggingface_hub.DocumentQuestionAnsweringInputData

< >

( image: Any question: str )

One (document, question) pair to answer

class huggingface_hub.DocumentQuestionAnsweringOutputElement

< >

( answer: str end: int score: float start: int words: List )

Outputs of inference for the Document Question Answering task

class huggingface_hub.DocumentQuestionAnsweringParameters

< >

( doc_stride: Optional = None handle_impossible_answer: Optional = None lang: Optional = None max_answer_len: Optional = None max_question_len: Optional = None max_seq_len: Optional = None top_k: Optional = None word_boxes: Optional = None )

Additional inference parameters Additional inference parameters for Document Question Answering

feature_extraction

class huggingface_hub.FeatureExtractionInput

< >

( inputs: str parameters: Optional = None )

Inputs for Text Embedding inference

fill_mask

class huggingface_hub.FillMaskInput

< >

( inputs: str parameters: Optional = None )

Inputs for Fill Mask inference

class huggingface_hub.FillMaskOutputElement

< >

( score: float sequence: str token: int token_str: Any fill_mask_output_token_str: Optional = None )

Outputs of inference for the Fill Mask task

class huggingface_hub.FillMaskParameters

< >

( targets: Optional = None top_k: Optional = None )

Additional inference parameters Additional inference parameters for Fill Mask

image_classification

class huggingface_hub.ImageClassificationInput

< >

( inputs: Any parameters: Optional = None )

Inputs for Image Classification inference

class huggingface_hub.ImageClassificationOutputElement

< >

( label: str score: float )

Outputs of inference for the Image Classification task

class huggingface_hub.ImageClassificationParameters

< >

( function_to_apply: Optional = None top_k: Optional = None )

Additional inference parameters Additional inference parameters for Image Classification

image_segmentation

class huggingface_hub.ImageSegmentationInput

< >

( inputs: Any parameters: Optional = None )

Inputs for Image Segmentation inference

class huggingface_hub.ImageSegmentationOutputElement

< >

( label: str mask: Any score: Optional = None )

Outputs of inference for the Image Segmentation task A predicted mask / segment

class huggingface_hub.ImageSegmentationParameters

< >

( mask_threshold: Optional = None overlap_mask_area_threshold: Optional = None subtask: Optional = None threshold: Optional = None )

Additional inference parameters Additional inference parameters for Image Segmentation

image_to_image

class huggingface_hub.ImageToImageInput

< >

( inputs: Any parameters: Optional = None )

Inputs for Image To Image inference

class huggingface_hub.ImageToImageOutput

< >

( image: Any )

Outputs of inference for the Image To Image task

class huggingface_hub.ImageToImageParameters

< >

( guidance_scale: Optional = None negative_prompt: Optional = None num_inference_steps: Optional = None target_size: Optional = None )

Additional inference parameters Additional inference parameters for Image To Image

class huggingface_hub.ImageToImageTargetSize

< >

( height: int width: int )

The size in pixel of the output image

image_to_text

class huggingface_hub.ImageToTextGenerationParameters

< >

( do_sample: Optional = None early_stopping: Union = None epsilon_cutoff: Optional = None eta_cutoff: Optional = None max_length: Optional = None max_new_tokens: Optional = None min_length: Optional = None min_new_tokens: Optional = None num_beam_groups: Optional = None num_beams: Optional = None penalty_alpha: Optional = None temperature: Optional = None top_k: Optional = None top_p: Optional = None typical_p: Optional = None use_cache: Optional = None )

Parametrization of the text generation process Ad-hoc parametrization of the text generation process

class huggingface_hub.ImageToTextInput

< >

( inputs: Any parameters: Optional = None )

Inputs for Image To Text inference

class huggingface_hub.ImageToTextOutput

< >

( generated_text: Any image_to_text_output_generated_text: Optional = None )

Outputs of inference for the Image To Text task

class huggingface_hub.ImageToTextParameters

< >

( generate: Optional = None max_new_tokens: Optional = None )

Additional inference parameters Additional inference parameters for Image To Text

object_detection

class huggingface_hub.ObjectDetectionBoundingBox

< >

( xmax: int xmin: int ymax: int ymin: int )

The predicted bounding box. Coordinates are relative to the top left corner of the input image.

class huggingface_hub.ObjectDetectionInput

< >

( inputs: Any parameters: Optional = None )

Inputs for Object Detection inference

class huggingface_hub.ObjectDetectionOutputElement

< >

( box: ObjectDetectionBoundingBox label: str score: float )

Outputs of inference for the Object Detection task

class huggingface_hub.ObjectDetectionParameters

< >

( threshold: Optional = None )

Additional inference parameters Additional inference parameters for Object Detection

question_answering

class huggingface_hub.QuestionAnsweringInput

< >

( inputs: QuestionAnsweringInputData parameters: Optional = None )

Inputs for Question Answering inference

class huggingface_hub.QuestionAnsweringInputData

< >

( context: str question: str )

One (context, question) pair to answer

class huggingface_hub.QuestionAnsweringOutputElement

< >

( answer: str end: int score: float start: int )

Outputs of inference for the Question Answering task

class huggingface_hub.QuestionAnsweringParameters

< >

( align_to_words: Optional = None doc_stride: Optional = None handle_impossible_answer: Optional = None max_answer_len: Optional = None max_question_len: Optional = None max_seq_len: Optional = None top_k: Optional = None )

Additional inference parameters Additional inference parameters for Question Answering

sentence_similarity

class huggingface_hub.SentenceSimilarityInput

< >

( inputs: SentenceSimilarityInputData parameters: Optional = None )

Inputs for Sentence similarity inference

class huggingface_hub.SentenceSimilarityInputData

< >

( sentences: List source_sentence: str )

summarization

class huggingface_hub.SummarizationGenerationParameters

< >

( clean_up_tokenization_spaces: Optional = None generate_parameters: Optional = None truncation: Optional = None )

Additional inference parameters Additional inference parameters for Text2text Generation

class huggingface_hub.SummarizationInput

< >

( inputs: str parameters: Optional = None )

Inputs for Summarization inference Inputs for Text2text Generation inference

class huggingface_hub.SummarizationOutput

< >

( summary_text: str )

Outputs of inference for the Summarization task

table_question_answering

class huggingface_hub.TableQuestionAnsweringInput

< >

( inputs: TableQuestionAnsweringInputData parameters: Optional = None )

Inputs for Table Question Answering inference

class huggingface_hub.TableQuestionAnsweringInputData

< >

( question: str table: Dict )

One (table, question) pair to answer

class huggingface_hub.TableQuestionAnsweringOutputElement

< >

( answer: str cells: List coordinates: List aggregator: Optional = None )

Outputs of inference for the Table Question Answering task

text2text_generation

class huggingface_hub.Text2TextGenerationInput

< >

( inputs: str parameters: Optional = None )

Inputs for Text2text Generation inference

class huggingface_hub.Text2TextGenerationOutput

< >

( generated_text: Any text2_text_generation_output_generated_text: Optional = None )

Outputs of inference for the Text2text Generation task

class huggingface_hub.Text2TextGenerationParameters

< >

( clean_up_tokenization_spaces: Optional = None generate_parameters: Optional = None truncation: Optional = None )

Additional inference parameters Additional inference parameters for Text2text Generation

text_classification

class huggingface_hub.TextClassificationInput

< >

( inputs: str parameters: Optional = None )

Inputs for Text Classification inference

class huggingface_hub.TextClassificationOutputElement

< >

( label: str score: float )

Outputs of inference for the Text Classification task

class huggingface_hub.TextClassificationParameters

< >

( function_to_apply: Optional = None top_k: Optional = None )

Additional inference parameters Additional inference parameters for Text Classification

text_generation

class huggingface_hub.TextGenerationInput

< >

( inputs: str parameters: Optional = None stream: Optional = None )

Text Generation Input. Auto-generated from TGI specs. For more details, check out https://github.com/huggingface/huggingface.js/blob/main/packages/tasks/scripts/inference-tgi-import.ts.

class huggingface_hub.TextGenerationInputGenerateParameters

< >

( best_of: Optional = None decoder_input_details: Optional = None details: Optional = None do_sample: Optional = None frequency_penalty: Optional = None grammar: Optional = None max_new_tokens: Optional = None repetition_penalty: Optional = None return_full_text: Optional = None seed: Optional = None stop: Optional = None temperature: Optional = None top_k: Optional = None top_n_tokens: Optional = None top_p: Optional = None truncate: Optional = None typical_p: Optional = None watermark: Optional = None )

class huggingface_hub.TextGenerationInputGrammarType

< >

( type: TypeEnum value: Any )

class huggingface_hub.TextGenerationOutput

< >

( generated_text: str details: Optional = None )

Text Generation Output. Auto-generated from TGI specs. For more details, check out https://github.com/huggingface/huggingface.js/blob/main/packages/tasks/scripts/inference-tgi-import.ts.

class huggingface_hub.TextGenerationOutputBestOfSequence

< >

( finish_reason: TextGenerationOutputFinishReason generated_text: str generated_tokens: int prefill: List tokens: List seed: Optional = None top_tokens: Optional = None )

class huggingface_hub.TextGenerationOutputDetails

< >

( finish_reason: TextGenerationOutputFinishReason generated_tokens: int prefill: List tokens: List best_of_sequences: Optional = None seed: Optional = None top_tokens: Optional = None )

class huggingface_hub.TextGenerationOutputPrefillToken

< >

( id: int logprob: float text: str )

class huggingface_hub.TextGenerationOutputToken

< >

( id: int logprob: float special: bool text: str )

class huggingface_hub.TextGenerationStreamOutput

< >

( index: int token: TextGenerationStreamOutputToken details: Optional = None generated_text: Optional = None top_tokens: Optional = None )

Text Generation Stream Output. Auto-generated from TGI specs. For more details, check out https://github.com/huggingface/huggingface.js/blob/main/packages/tasks/scripts/inference-tgi-import.ts.

class huggingface_hub.TextGenerationStreamOutputStreamDetails

< >

( finish_reason: TextGenerationOutputFinishReason generated_tokens: int seed: Optional = None )

class huggingface_hub.TextGenerationStreamOutputToken

< >

( id: int logprob: float special: bool text: str )

text_to_audio

class huggingface_hub.TextToAudioGenerationParameters

< >

( do_sample: Optional = None early_stopping: Union = None epsilon_cutoff: Optional = None eta_cutoff: Optional = None max_length: Optional = None max_new_tokens: Optional = None min_length: Optional = None min_new_tokens: Optional = None num_beam_groups: Optional = None num_beams: Optional = None penalty_alpha: Optional = None temperature: Optional = None top_k: Optional = None top_p: Optional = None typical_p: Optional = None use_cache: Optional = None )

Parametrization of the text generation process Ad-hoc parametrization of the text generation process

class huggingface_hub.TextToAudioInput

< >

( inputs: str parameters: Optional = None )

Inputs for Text To Audio inference

class huggingface_hub.TextToAudioOutput

< >

( audio: Any sampling_rate: Any text_to_audio_output_sampling_rate: Optional = None )

Outputs of inference for the Text To Audio task

class huggingface_hub.TextToAudioParameters

< >

( generate: Optional = None )

Additional inference parameters Additional inference parameters for Text To Audio

text_to_image

class huggingface_hub.TextToImageInput

< >

( inputs: str parameters: Optional = None )

Inputs for Text To Image inference

class huggingface_hub.TextToImageOutput

< >

( image: Any )

Outputs of inference for the Text To Image task

class huggingface_hub.TextToImageParameters

< >

( guidance_scale: Optional = None negative_prompt: Optional = None num_inference_steps: Optional = None scheduler: Optional = None target_size: Optional = None )

Additional inference parameters Additional inference parameters for Text To Image

class huggingface_hub.TextToImageTargetSize

< >

( height: int width: int )

The size in pixel of the output image

token_classification

class huggingface_hub.TokenClassificationInput

< >

( inputs: str parameters: Optional = None )

Inputs for Token Classification inference

class huggingface_hub.TokenClassificationOutputElement

< >

( label: Any score: float end: Optional = None entity_group: Optional = None start: Optional = None word: Optional = None )

Outputs of inference for the Token Classification task

class huggingface_hub.TokenClassificationParameters

< >

( aggregation_strategy: Optional = None ignore_labels: Optional = None stride: Optional = None )

Additional inference parameters Additional inference parameters for Token Classification

translation

class huggingface_hub.TranslationGenerationParameters

< >

( clean_up_tokenization_spaces: Optional = None generate_parameters: Optional = None truncation: Optional = None )

Additional inference parameters Additional inference parameters for Text2text Generation

class huggingface_hub.TranslationInput

< >

( inputs: str parameters: Optional = None )

Inputs for Translation inference Inputs for Text2text Generation inference

class huggingface_hub.TranslationOutput

< >

( translation_text: str )

Outputs of inference for the Translation task

video_classification

class huggingface_hub.VideoClassificationInput

< >

( inputs: Any parameters: Optional = None )

Inputs for Video Classification inference

class huggingface_hub.VideoClassificationOutputElement

< >

( label: str score: float )

Outputs of inference for the Video Classification task

class huggingface_hub.VideoClassificationParameters

< >

( frame_sampling_rate: Optional = None function_to_apply: Optional = None num_frames: Optional = None top_k: Optional = None )

Additional inference parameters Additional inference parameters for Video Classification

visual_question_answering

class huggingface_hub.VisualQuestionAnsweringInput

< >

( inputs: VisualQuestionAnsweringInputData parameters: Optional = None )

Inputs for Visual Question Answering inference

class huggingface_hub.VisualQuestionAnsweringInputData

< >

( image: Any question: Any )

One (image, question) pair to answer

class huggingface_hub.VisualQuestionAnsweringOutputElement

< >

( label: Any score: float answer: Optional = None )

Outputs of inference for the Visual Question Answering task

class huggingface_hub.VisualQuestionAnsweringParameters

< >

( top_k: Optional = None )

Additional inference parameters Additional inference parameters for Visual Question Answering

zero_shot_classification

class huggingface_hub.ZeroShotClassificationInput

< >

( inputs: ZeroShotClassificationInputData parameters: Optional = None )

Inputs for Zero Shot Classification inference

class huggingface_hub.ZeroShotClassificationInputData

< >

( candidate_labels: List text: str )

The input text data, with candidate labels

class huggingface_hub.ZeroShotClassificationOutputElement

< >

( label: str score: float )

Outputs of inference for the Zero Shot Classification task

class huggingface_hub.ZeroShotClassificationParameters

< >

( hypothesis_template: Optional = None multi_label: Optional = None )

Additional inference parameters Additional inference parameters for Zero Shot Classification

zero_shot_image_classification

class huggingface_hub.ZeroShotImageClassificationInput

< >

( inputs: ZeroShotImageClassificationInputData parameters: Optional = None )

Inputs for Zero Shot Image Classification inference

class huggingface_hub.ZeroShotImageClassificationInputData

< >

( candidate_labels: List image: Any )

The input image data, with candidate labels

class huggingface_hub.ZeroShotImageClassificationOutputElement

< >

( label: str score: float )

Outputs of inference for the Zero Shot Image Classification task

class huggingface_hub.ZeroShotImageClassificationParameters

< >

( hypothesis_template: Optional = None )

Additional inference parameters Additional inference parameters for Zero Shot Image Classification

zero_shot_object_detection

class huggingface_hub.ZeroShotObjectDetectionBoundingBox

< >

( xmax: int xmin: int ymax: int ymin: int )

The predicted bounding box. Coordinates are relative to the top left corner of the input image.

class huggingface_hub.ZeroShotObjectDetectionInput

< >

( inputs: ZeroShotObjectDetectionInputData parameters: Optional = None )

Inputs for Zero Shot Object Detection inference

class huggingface_hub.ZeroShotObjectDetectionInputData

< >

( candidate_labels: List image: Any )

The input image data, with candidate labels

class huggingface_hub.ZeroShotObjectDetectionOutputElement

< >

( box: ZeroShotObjectDetectionBoundingBox label: str score: float )

Outputs of inference for the Zero Shot Object Detection task

< > Update on GitHub