{"index": {"('MODELS', 'inverse-text-processing', 'generic-itn')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/audio/itn/generic_inverse_text_processing.py", "imports": ["os", "typing"], "module": "modelscope.models.audio.itn.generic_inverse_text_processing"}, "('MODELS', 'speaker-verification', 'generic-sv')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/audio/sv/generic_speaker_verification.py", "imports": ["os", "typing"], "module": "modelscope.models.audio.sv.generic_speaker_verification"}, "('MODELS', 'speaker-diarization', 'generic-sv')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/audio/sv/generic_speaker_verification.py", "imports": ["os", "typing"], "module": "modelscope.models.audio.sv.generic_speaker_verification"}, "('MODELS', 'speaker-verification', 'ecapa-tdnn-sv')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/audio/sv/ecapa_tdnn.py", "imports": ["math", "typing", "os", "torch", "torchaudio"], "module": "modelscope.models.audio.sv.ecapa_tdnn"}, "('MODELS', 'acoustic-noise-suppression', 'speech_dfsmn_ans')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/audio/ans/denoise_net.py", "imports": ["torch"], "module": "modelscope.models.audio.ans.denoise_net"}, "('MODELS', 'acoustic-noise-suppression', 'speech_frcrn_ans_cirm_16k')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/audio/ans/frcrn.py", "imports": ["os", "torch", "typing"], "module": "modelscope.models.audio.ans.frcrn"}, "('MODELS', 'text-to-speech', 'sambert-hifigan')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/audio/tts/sambert_hifi.py", "imports": ["datetime", "shutil", "matplotlib", "numpy", "json", "yaml", "os", "zipfile", "wave", "__future__"], "module": "modelscope.models.audio.tts.sambert_hifi"}, "('MODELS', 'punctuation', 'generic-punc')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/audio/punc/generic_punctuation.py", "imports": ["os", "typing"], "module": "modelscope.models.audio.punc.generic_punctuation"}, "('MODELS', 'speech-separation', 'speech_mossformer_separation_temporal_8k')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/audio/separation/mossformer.py", "imports": ["os", "torch", "typing", "copy"], "module": "modelscope.models.audio.separation.mossformer"}, "('MODELS', 'keyword-spotting', 'speech_dfsmn_kws_char_farfield')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/audio/kws/farfield/model.py", "imports": ["os", "tempfile", "typing"], "module": "modelscope.models.audio.kws.farfield.model"}, "('MODELS', 'keyword-spotting', 'kws-kwsbp')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/audio/kws/generic_key_word_spotting.py", "imports": ["os", "typing"], "module": "modelscope.models.audio.kws.generic_key_word_spotting"}, "('MODELS', 'keyword-spotting', 'speech_kws_fsmn_char_ctc_nearfield')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/audio/kws/nearfield/model.py", "imports": ["typing", "os", "tempfile", "torch", "sys"], "module": "modelscope.models.audio.kws.nearfield.model"}, "('MODELS', 'auto-speech-recognition', 'wenet-asr')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/audio/asr/wenet_automatic_speech_recognition.py", "imports": ["os", "wenetruntime", "json", "typing"], "module": "modelscope.models.audio.asr.wenet_automatic_speech_recognition"}, "('MODELS', 'auto-speech-recognition', 'generic-asr')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/audio/asr/generic_automatic_speech_recognition.py", "imports": ["os", "typing"], "module": "modelscope.models.audio.asr.generic_automatic_speech_recognition"}, "('MODELS', 'voice-activity-detection', 'generic-asr')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/audio/asr/generic_automatic_speech_recognition.py", "imports": ["os", "typing"], "module": "modelscope.models.audio.asr.generic_automatic_speech_recognition"}, "('MODELS', 'language-score-prediction', 'generic-asr')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/audio/asr/generic_automatic_speech_recognition.py", "imports": ["os", "typing"], "module": "modelscope.models.audio.asr.generic_automatic_speech_recognition"}, "('MODELS', 'speech-timestamp', 'generic-asr')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/audio/asr/generic_automatic_speech_recognition.py", "imports": ["os", "typing"], "module": "modelscope.models.audio.asr.generic_automatic_speech_recognition"}, "('MODELS', 'protein-structure', 'unifold')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/science/unifold/model.py", "imports": ["os", "argparse", "torch", "typing"], "module": "modelscope.models.science.unifold.model"}, "('MODELS', 'backbone', 'xlm-roberta')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/xlm_roberta/backbone.py", "imports": ["transformers", "math", "torch", "packaging"], "module": "modelscope.models.nlp.xlm_roberta.backbone"}, "('MODELS', 'text-error-correction', 'bart')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/bart/text_error_correction.py", "imports": ["os", "torch", "typing"], "module": "modelscope.models.nlp.bart.text_error_correction"}, "('MODELS', 'backbone', 'T5')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/T5/backbone.py", "imports": ["math", "transformers", "typing", "os", "torch", "copy", "warnings"], "module": "modelscope.models.nlp.T5.backbone"}, "('MODELS', 'text2text-generation', 'T5')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/T5/text2text_generation.py", "imports": ["transformers", "typing", "torch", "copy", "warnings"], "module": "modelscope.models.nlp.T5.text2text_generation"}, "('MODELS', 'translation-evaluation', 'unite')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/unite/modeling_unite.py", "imports": ["math", "numpy", "transformers", "dataclasses", "typing", "packaging", "torch", "warnings"], "module": "modelscope.models.nlp.unite.modeling_unite"}, "('BACKBONES', 'backbone', 'gpt-neo')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/gpt_neo/backbone.py", "imports": ["transformers"], "module": "modelscope.models.nlp.gpt_neo.backbone"}, "('BACKBONES', 'backbone', 'gpt2')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/gpt2/backbone.py", "imports": ["transformers"], "module": "modelscope.models.nlp.gpt2.backbone"}, "('HEADS', 'fill-mask', 'roberta-mlm')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/heads/torch_pretrain_head.py", "imports": ["torch", "transformers", "typing"], "module": "modelscope.models.nlp.heads.torch_pretrain_head"}, "('HEADS', 'text-classification', 'text-classification')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/heads/text_classification_head.py", "imports": ["torch", "typing"], "module": "modelscope.models.nlp.heads.text_classification_head"}, "('HEADS', 'sentence-similarity', 'text-classification')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/heads/text_classification_head.py", "imports": ["torch", "typing"], "module": "modelscope.models.nlp.heads.text_classification_head"}, "('HEADS', 'nli', 'text-classification')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/heads/text_classification_head.py", "imports": ["torch", "typing"], "module": "modelscope.models.nlp.heads.text_classification_head"}, "('HEADS', 'sentiment-classification', 'text-classification')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/heads/text_classification_head.py", "imports": ["torch", "typing"], "module": "modelscope.models.nlp.heads.text_classification_head"}, "('HEADS', 'token-classification', 'lstm-crf')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/heads/crf_head.py", "imports": ["torch", "transformers", "typing"], "module": "modelscope.models.nlp.heads.crf_head"}, "('HEADS', 'named-entity-recognition', 'lstm-crf')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/heads/crf_head.py", "imports": ["torch", "transformers", "typing"], "module": "modelscope.models.nlp.heads.crf_head"}, "('HEADS', 'word-segmentation', 'lstm-crf')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/heads/crf_head.py", "imports": ["torch", "transformers", "typing"], "module": "modelscope.models.nlp.heads.crf_head"}, "('HEADS', 'part-of-speech', 'lstm-crf')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/heads/crf_head.py", "imports": ["torch", "transformers", "typing"], "module": "modelscope.models.nlp.heads.crf_head"}, "('HEADS', 'transformer-crf', 'transformer-crf')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/heads/crf_head.py", "imports": ["torch", "transformers", "typing"], "module": "modelscope.models.nlp.heads.crf_head"}, "('HEADS', 'token-classification', 'transformer-crf')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/heads/crf_head.py", "imports": ["torch", "transformers", "typing"], "module": "modelscope.models.nlp.heads.crf_head"}, "('HEADS', 'named-entity-recognition', 'transformer-crf')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/heads/crf_head.py", "imports": ["torch", "transformers", "typing"], "module": "modelscope.models.nlp.heads.crf_head"}, "('HEADS', 'word-segmentation', 'transformer-crf')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/heads/crf_head.py", "imports": ["torch", "transformers", "typing"], "module": "modelscope.models.nlp.heads.crf_head"}, "('HEADS', 'part-of-speech', 'transformer-crf')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/heads/crf_head.py", "imports": ["torch", "transformers", "typing"], "module": "modelscope.models.nlp.heads.crf_head"}, "('HEADS', 'information-extraction', 'information-extraction')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/heads/infromation_extraction_head.py", "imports": ["torch"], "module": "modelscope.models.nlp.heads.infromation_extraction_head"}, "('HEADS', 'relation-extraction', 'information-extraction')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/heads/infromation_extraction_head.py", "imports": ["torch"], "module": "modelscope.models.nlp.heads.infromation_extraction_head"}, "('HEADS', 'text-ranking', 'text-ranking')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/heads/text_ranking_head.py", "imports": ["torch", "typing"], "module": "modelscope.models.nlp.heads.text_ranking_head"}, "('HEADS', 'token-classification', 'token-classification')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/heads/token_classification_head.py", "imports": ["torch", "typing"], "module": "modelscope.models.nlp.heads.token_classification_head"}, "('HEADS', 'named-entity-recognition', 'token-classification')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/heads/token_classification_head.py", "imports": ["torch", "typing"], "module": "modelscope.models.nlp.heads.token_classification_head"}, "('HEADS', 'part-of-speech', 'token-classification')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/heads/token_classification_head.py", "imports": ["torch", "typing"], "module": "modelscope.models.nlp.heads.token_classification_head"}, "('HEADS', 'fill-mask', 'bert-mlm')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/heads/fill_mask_head.py", "imports": ["torch", "transformers", "typing"], "module": "modelscope.models.nlp.heads.fill_mask_head"}, "('HEADS', 'fill-mask', 'fill-mask')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/heads/fill_mask_head.py", "imports": ["torch", "transformers", "typing"], "module": "modelscope.models.nlp.heads.fill_mask_head"}, "('HEADS', 'text-generation', 'text-generation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/heads/text_generation_head.py", "imports": ["torch", "typing"], "module": "modelscope.models.nlp.heads.text_generation_head"}, "('MODELS', 'backbone', 'transformers')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/hf_transformers/backbone.py", "imports": ["transformers"], "module": "modelscope.models.nlp.hf_transformers.backbone"}, "('MODELS', 'code-generation', 'codegeex')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/codegeex/codegeex_for_code_generation.py", "imports": ["torch", "typing", "copy"], "module": "modelscope.models.nlp.codegeex.codegeex_for_code_generation"}, "('MODELS', 'code-translation', 'codegeex')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/codegeex/codegeex_for_code_translation.py", "imports": ["torch", "typing", "copy"], "module": "modelscope.models.nlp.codegeex.codegeex_for_code_translation"}, "('MODELS', 'table-question-answering', 'space-T-en')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/space_T_en/text_to_sql.py", "imports": ["os", "torch", "text2sql_lgesql", "typing"], "module": "modelscope.models.nlp.space_T_en.text_to_sql"}, "('MODELS', 'text-generation', 'palm-v2')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/palm_v2/text_generation.py", "imports": ["math", "subprocess", "numpy", "json", "transformers", "dataclasses", "typing", "os", "copy", "torch", "codecs"], "module": "modelscope.models.nlp.palm_v2.text_generation"}, "('MODELS', 'document-grounded-dialog-retrieval', 'doc2bot')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/dgds/document_grounded_dialog_retrieval.py", "imports": ["os", "torch", "typing"], "module": "modelscope.models.nlp.dgds.document_grounded_dialog_retrieval"}, "('MODELS', 'document-grounded-dialog-rerank', 'doc2bot')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/dgds/document_grounded_dialog_rerank.py", "imports": ["os", "torch", "typing"], "module": "modelscope.models.nlp.dgds.document_grounded_dialog_rerank"}, "('MODELS', 'document-grounded-dialog-generate', 'doc2bot')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/dgds/document_grounded_dialog_generate.py", "imports": ["os", "torch", "typing"], "module": "modelscope.models.nlp.dgds.document_grounded_dialog_generate"}, "('MODELS', 'backbone', 'lstm')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/lstm/backbone.py", "imports": ["torch"], "module": "modelscope.models.nlp.lstm.backbone"}, "('MODELS', 'token-classification', 'lstm-crf')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/lstm/token_classification.py", "imports": [], "module": "modelscope.models.nlp.lstm.token_classification"}, "('MODELS', 'named-entity-recognition', 'lstm-crf')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/lstm/token_classification.py", "imports": [], "module": "modelscope.models.nlp.lstm.token_classification"}, "('MODELS', 'part-of-speech', 'lstm-crf')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/lstm/token_classification.py", "imports": [], "module": "modelscope.models.nlp.lstm.token_classification"}, "('MODELS', 'word-segmentation', 'lstm-crf')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/lstm/token_classification.py", "imports": [], "module": "modelscope.models.nlp.lstm.token_classification"}, "('MODELS', 'word-segmentation', 'lstm-crf-for-word-segmentation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/lstm/token_classification.py", "imports": [], "module": "modelscope.models.nlp.lstm.token_classification"}, "('BACKBONES', 'backbone', 'bloom')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/bloom/backbone.py", "imports": ["transformers"], "module": "modelscope.models.nlp.bloom.backbone"}, "('MODELS', 'text-classification', 'text-classification')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/task_models/text_classification.py", "imports": ["numpy", "typing"], "module": "modelscope.models.nlp.task_models.text_classification"}, "('MODELS', 'feature-extraction', 'feature-extraction')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/task_models/feature_extraction.py", "imports": ["numpy", "typing"], "module": "modelscope.models.nlp.task_models.feature_extraction"}, "('MODELS', 'information-extraction', 'information-extraction')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/task_models/information_extraction.py", "imports": ["numpy", "typing"], "module": "modelscope.models.nlp.task_models.information_extraction"}, "('MODELS', 'relation-extraction', 'information-extraction')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/task_models/information_extraction.py", "imports": ["numpy", "typing"], "module": "modelscope.models.nlp.task_models.information_extraction"}, "('MODELS', 'token-classification', 'token-classification')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/task_models/token_classification.py", "imports": ["torch", "typing"], "module": "modelscope.models.nlp.task_models.token_classification"}, "('MODELS', 'part-of-speech', 'token-classification')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/task_models/token_classification.py", "imports": ["torch", "typing"], "module": "modelscope.models.nlp.task_models.token_classification"}, "('MODELS', 'named-entity-recognition', 'token-classification-for-ner')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/task_models/token_classification.py", "imports": ["torch", "typing"], "module": "modelscope.models.nlp.task_models.token_classification"}, "('MODELS', 'transformer-crf', 'transformer-crf')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/task_models/token_classification.py", "imports": ["torch", "typing"], "module": "modelscope.models.nlp.task_models.token_classification"}, "('MODELS', 'token-classification', 'transformer-crf')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/task_models/token_classification.py", "imports": ["torch", "typing"], "module": "modelscope.models.nlp.task_models.token_classification"}, "('MODELS', 'token-classification', 'transformer-crf-for-word-segmentation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/task_models/token_classification.py", "imports": ["torch", "typing"], "module": "modelscope.models.nlp.task_models.token_classification"}, "('MODELS', 'named-entity-recognition', 'transformer-crf')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/task_models/token_classification.py", "imports": ["torch", "typing"], "module": "modelscope.models.nlp.task_models.token_classification"}, "('MODELS', 'part-of-speech', 'transformer-crf')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/task_models/token_classification.py", "imports": ["torch", "typing"], "module": "modelscope.models.nlp.task_models.token_classification"}, "('MODELS', 'word-segmentation', 'transformer-crf')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/task_models/token_classification.py", "imports": ["torch", "typing"], "module": "modelscope.models.nlp.task_models.token_classification"}, "('MODELS', 'word-segmentation', 'transformer-crf-for-word-segmentation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/task_models/token_classification.py", "imports": ["torch", "typing"], "module": "modelscope.models.nlp.task_models.token_classification"}, "('MODELS', 'fill-mask', 'fill-mask')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/task_models/fill_mask.py", "imports": ["torch", "numpy", "typing"], "module": "modelscope.models.nlp.task_models.fill_mask"}, "('MODELS', 'text-generation', 'text-generation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/task_models/text_generation.py", "imports": ["numpy", "transformers", "typing"], "module": "modelscope.models.nlp.task_models.text_generation"}, "('MODELS', 'text-ranking', 'text-ranking')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/task_models/text_ranking.py", "imports": ["numpy", "typing"], "module": "modelscope.models.nlp.task_models.text_ranking"}, "('MODELS', 'text-classification', 'peer')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/peer/text_classification.py", "imports": ["torch", "copy"], "module": "modelscope.models.nlp.peer.text_classification"}, "('MODELS', 'nli', 'peer')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/peer/text_classification.py", "imports": ["torch", "copy"], "module": "modelscope.models.nlp.peer.text_classification"}, "('MODELS', 'sentiment-classification', 'peer')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/peer/text_classification.py", "imports": ["torch", "copy"], "module": "modelscope.models.nlp.peer.text_classification"}, "('MODELS', 'sentence-similarity', 'peer')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/peer/text_classification.py", "imports": ["torch", "copy"], "module": "modelscope.models.nlp.peer.text_classification"}, "('MODELS', 'zero-shot-classification', 'peer')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/peer/text_classification.py", "imports": ["torch", "copy"], "module": "modelscope.models.nlp.peer.text_classification"}, "('MODELS', 'backbone', 'deberta_v2')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/deberta_v2/backbone.py", "imports": ["torch", "collections", "transformers", "typing"], "module": "modelscope.models.nlp.deberta_v2.backbone"}, "('MODELS', 'fill-mask', 'deberta_v2')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/deberta_v2/fill_mask.py", "imports": ["torch", "transformers", "typing"], "module": "modelscope.models.nlp.deberta_v2.fill_mask"}, "('MODELS', 'backbone', 'megatron-bert')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/megatron_bert/backbone.py", "imports": ["transformers", "math", "torch"], "module": "modelscope.models.nlp.megatron_bert.backbone"}, "('MODELS', 'fill-mask', 'megatron-bert')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/megatron_bert/fill_mask.py", "imports": ["transformers", "torch"], "module": "modelscope.models.nlp.megatron_bert.fill_mask"}, "('MODELS', 'backbone', 'structbert')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/structbert/backbone.py", "imports": ["math", "transformers", "dataclasses", "typing", "packaging", "torch"], "module": "modelscope.models.nlp.structbert.backbone"}, "('MODELS', 'text-classification', 'structbert')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/structbert/text_classification.py", "imports": ["torch"], "module": "modelscope.models.nlp.structbert.text_classification"}, "('MODELS', 'nli', 'structbert')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/structbert/text_classification.py", "imports": ["torch"], "module": "modelscope.models.nlp.structbert.text_classification"}, "('MODELS', 'sentiment-classification', 'structbert')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/structbert/text_classification.py", "imports": ["torch"], "module": "modelscope.models.nlp.structbert.text_classification"}, "('MODELS', 'sentence-similarity', 'structbert')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/structbert/text_classification.py", "imports": ["torch"], "module": "modelscope.models.nlp.structbert.text_classification"}, "('MODELS', 'zero-shot-classification', 'structbert')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/structbert/text_classification.py", "imports": ["torch"], "module": "modelscope.models.nlp.structbert.text_classification"}, "('MODELS', 'faq-question-answering', 'structbert')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/structbert/faq_question_answering.py", "imports": ["math", "typing", "os", "torch", "collections"], "module": "modelscope.models.nlp.structbert.faq_question_answering"}, "('MODELS', 'token-classification', 'structbert')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/structbert/token_classification.py", "imports": ["torch"], "module": "modelscope.models.nlp.structbert.token_classification"}, "('MODELS', 'word-segmentation', 'structbert')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/structbert/token_classification.py", "imports": ["torch"], "module": "modelscope.models.nlp.structbert.token_classification"}, "('MODELS', 'part-of-speech', 'structbert')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/structbert/token_classification.py", "imports": ["torch"], "module": "modelscope.models.nlp.structbert.token_classification"}, "('MODELS', 'fill-mask', 'structbert')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/structbert/fill_mask.py", "imports": ["transformers", "torch"], "module": "modelscope.models.nlp.structbert.fill_mask"}, "('MODELS', 'backbone', 'plug-mental')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/plug_mental/backbone.py", "imports": ["math", "transformers", "dataclasses", "typing", "packaging", "torch"], "module": "modelscope.models.nlp.plug_mental.backbone"}, "('MODELS', 'text-classification', 'plug-mental')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/plug_mental/text_classification.py", "imports": ["torch"], "module": "modelscope.models.nlp.plug_mental.text_classification"}, "('MODELS', 'nli', 'plug-mental')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/plug_mental/text_classification.py", "imports": ["torch"], "module": "modelscope.models.nlp.plug_mental.text_classification"}, "('MODELS', 'sentiment-classification', 'plug-mental')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/plug_mental/text_classification.py", "imports": ["torch"], "module": "modelscope.models.nlp.plug_mental.text_classification"}, "('MODELS', 'sentence-similarity', 'plug-mental')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/plug_mental/text_classification.py", "imports": ["torch"], "module": "modelscope.models.nlp.plug_mental.text_classification"}, "('MODELS', 'zero-shot-classification', 'plug-mental')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/plug_mental/text_classification.py", "imports": ["torch"], "module": "modelscope.models.nlp.plug_mental.text_classification"}, "('MODELS', 'text-generation', 'gpt3')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/gpt3/text_generation.py", "imports": ["torch", "collections", "transformers", "typing"], "module": "modelscope.models.nlp.gpt3.text_generation"}, "('MODELS', 'translation', 'csanmt-translation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/csanmt/translation.py", "imports": ["tensorflow", "collections", "math", "typing"], "module": "modelscope.models.nlp.csanmt.translation"}, "('MODELS', 'table-question-answering', 'space-T-cn')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/space_T_cn/table_question_answering.py", "imports": ["numpy", "transformers", "typing", "os", "torch"], "module": "modelscope.models.nlp.space_T_cn.table_question_answering"}, "('MODELS', 'text-summarization', 'mglm')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/mglm/mglm_for_text_summarization.py", "imports": ["megatron_util", "numpy", "typing", "os", "random", "torch"], "module": "modelscope.models.nlp.mglm.mglm_for_text_summarization"}, "('MODELS', 'backbone', 'veco')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/veco/backbone.py", "imports": ["transformers"], "module": "modelscope.models.nlp.veco.backbone"}, "('MODELS', 'nli', 'veco')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/veco/text_classification.py", "imports": ["transformers"], "module": "modelscope.models.nlp.veco.text_classification"}, "('MODELS', 'sentiment-classification', 'veco')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/veco/text_classification.py", "imports": ["transformers"], "module": "modelscope.models.nlp.veco.text_classification"}, "('MODELS', 'sentence-similarity', 'veco')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/veco/text_classification.py", "imports": ["transformers"], "module": "modelscope.models.nlp.veco.text_classification"}, "('MODELS', 'text-classification', 'veco')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/veco/text_classification.py", "imports": ["transformers"], "module": "modelscope.models.nlp.veco.text_classification"}, "('MODELS', 'token-classification', 'veco')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/veco/token_classification.py", "imports": ["transformers", "torch"], "module": "modelscope.models.nlp.veco.token_classification"}, "('MODELS', 'fill-mask', 'veco')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/veco/fill_mask.py", "imports": ["transformers"], "module": "modelscope.models.nlp.veco.fill_mask"}, "('MODELS', 'backbone', 'ponet')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/ponet/backbone.py", "imports": ["math", "distutils", "transformers", "packaging", "torch"], "module": "modelscope.models.nlp.ponet.backbone"}, "('MODELS', 'document-segmentation', 'ponet-for-document-segmentation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/ponet/document_segmentation.py", "imports": ["torch", "typing"], "module": "modelscope.models.nlp.ponet.document_segmentation"}, "('MODELS', 'extractive-summarization', 'ponet-for-document-segmentation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/ponet/document_segmentation.py", "imports": ["torch", "typing"], "module": "modelscope.models.nlp.ponet.document_segmentation"}, "('MODELS', 'fill-mask', 'ponet')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/ponet/fill_mask.py", "imports": ["transformers", "torch"], "module": "modelscope.models.nlp.ponet.fill_mask"}, "('MODELS', 'fid-dialogue', 'fid-plug')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/fid_plug/text_generation.py", "imports": ["os", "io", "transformers", "torch"], "module": "modelscope.models.nlp.fid_plug.text_generation"}, "('MODELS', 'task-oriented-conversation', 'space-modeling')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/space/dialog_modeling.py", "imports": ["os", "typing"], "module": "modelscope.models.nlp.space.dialog_modeling"}, "('MODELS', 'task-oriented-conversation', 'space-dst')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/space/dialog_state_tracking.py", "imports": ["torch", "transformers", "typing"], "module": "modelscope.models.nlp.space.dialog_state_tracking"}, "('MODELS', 'task-oriented-conversation', 'space-intent')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/space/dialog_intent_prediction.py", "imports": ["os", "typing"], "module": "modelscope.models.nlp.space.dialog_intent_prediction"}, "('MODELS', 'backbone', 'bert')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/bert/backbone.py", "imports": ["transformers", "math", "torch", "packaging"], "module": "modelscope.models.nlp.bert.backbone"}, "('MODELS', 'text-classification', 'bert')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/bert/text_classification.py", "imports": [], "module": "modelscope.models.nlp.bert.text_classification"}, "('MODELS', 'nli', 'bert')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/bert/text_classification.py", "imports": [], "module": "modelscope.models.nlp.bert.text_classification"}, "('MODELS', 'sentiment-classification', 'bert')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/bert/text_classification.py", "imports": [], "module": "modelscope.models.nlp.bert.text_classification"}, "('MODELS', 'sentence-similarity', 'bert')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/bert/text_classification.py", "imports": [], "module": "modelscope.models.nlp.bert.text_classification"}, "('MODELS', 'zero-shot-classification', 'bert')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/bert/text_classification.py", "imports": [], "module": "modelscope.models.nlp.bert.text_classification"}, "('MODELS', 'document-segmentation', 'bert-for-document-segmentation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/bert/document_segmentation.py", "imports": ["torch", "typing"], "module": "modelscope.models.nlp.bert.document_segmentation"}, "('MODELS', 'sentence-embedding', 'bert')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/bert/sentence_embedding.py", "imports": ["torch"], "module": "modelscope.models.nlp.bert.sentence_embedding"}, "('MODELS', 'siamese-uie', 'bert')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/bert/siamese_uie.py", "imports": ["torch", "copy"], "module": "modelscope.models.nlp.bert.siamese_uie"}, "('MODELS', 'word-alignment', 'bert')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/bert/word_alignment.py", "imports": ["torch"], "module": "modelscope.models.nlp.bert.word_alignment"}, "('MODELS', 'token-classification', 'bert')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/bert/token_classification.py", "imports": [], "module": "modelscope.models.nlp.bert.token_classification"}, "('MODELS', 'part-of-speech', 'bert')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/bert/token_classification.py", "imports": [], "module": "modelscope.models.nlp.bert.token_classification"}, "('MODELS', 'word-segmentation', 'bert')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/bert/token_classification.py", "imports": [], "module": "modelscope.models.nlp.bert.token_classification"}, "('MODELS', 'fill-mask', 'bert')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/bert/fill_mask.py", "imports": [], "module": "modelscope.models.nlp.bert.fill_mask"}, "('MODELS', 'text-ranking', 'bert')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/bert/text_ranking.py", "imports": [], "module": "modelscope.models.nlp.bert.text_ranking"}, "('MODELS', 'text-classification', 'user-satisfaction-estimation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/use/user_satisfaction_estimation.py", "imports": ["numpy", "transformers", "typing", "os", "torch"], "module": "modelscope.models.nlp.use.user_satisfaction_estimation"}, "('MODELS', 'text-generation', 'gpt-moe')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/gpt_moe/text_generation.py", "imports": ["transformers", "typing"], "module": "modelscope.models.nlp.gpt_moe.text_generation"}, "('MODELS', 'face-reconstruction', 'face_reconstruction')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_reconstruction/models/facerecon_model.py", "imports": ["os", "numpy", "torch", "cv2"], "module": "modelscope.models.cv.face_reconstruction.models.facerecon_model"}, "('MODELS', 'panorama-depth-estimation', 'unifuse-depth-estimation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/panorama_depth_estimation/unifuse_model.py", "imports": ["os", "numpy", "torchvision", "torch"], "module": "modelscope.models.cv.panorama_depth_estimation.unifuse_model"}, "('NECKS', 'default', 'SemanticFPNWrapper')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_panoptic_segmentation/head/semantic_fpn_wrapper.py", "imports": ["mmdet", "mmcv", "torch"], "module": "modelscope.models.cv.video_panoptic_segmentation.head.semantic_fpn_wrapper"}, "('TRANSFORMER_LAYER', 'default', 'KernelUpdator')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_instance_segmentation/head/kernel_updator.py", "imports": ["mmcv", "torch"], "module": "modelscope.models.cv.video_instance_segmentation.head.kernel_updator"}, "('HEADS', 'default', 'VideoKernelUpdateHead')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_panoptic_segmentation/head/kernel_update_head.py", "imports": ["numpy", "mmdet", "mmcv", "torch"], "module": "modelscope.models.cv.video_panoptic_segmentation.head.kernel_update_head"}, "('HEADS', 'default', 'VideoKernelIterHead')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_panoptic_segmentation/head/kernel_iter_head.py", "imports": ["mmdet", "torch"], "module": "modelscope.models.cv.video_panoptic_segmentation.head.kernel_iter_head"}, "('MODELS', 'video-panoptic-segmentation', 'swinb-video-panoptic-segmentation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_panoptic_segmentation/video_k_net.py", "imports": ["numpy", "mmdet", "mmcv", "torch"], "module": "modelscope.models.cv.video_panoptic_segmentation.video_k_net"}, "('TRACKERS', 'default', 'QuasiDenseEmbedTracker')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_panoptic_segmentation/track/quasi_dense_embed_tracker.py", "imports": ["mmdet", "mmcv", "torch"], "module": "modelscope.models.cv.video_panoptic_segmentation.track.quasi_dense_embed_tracker"}, "('MODELS', 'image-matching', 'quadtree-attention-image-matching')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_matching/quadtree_attention_model.py", "imports": ["pathlib", "numpy", "cv2", "os", "torch"], "module": "modelscope.models.cv.image_matching.quadtree_attention_model"}, "('MODELS', 'image-denoising', 'nafnet')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_denoise/nafnet_for_image_denoise.py", "imports": ["os", "torch", "typing"], "module": "modelscope.models.cv.image_denoise.nafnet_for_image_denoise"}, "('MODELS', 'face-emotion', 'face-emotion')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_emotion/emotion_model.py", "imports": ["os", "sys", "torch"], "module": "modelscope.models.cv.face_emotion.emotion_model"}, "('MODELS', 'image-inpainting', 'FFTInpainting')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_inpainting/model.py", "imports": ["os", "torch", "typing"], "module": "modelscope.models.cv.image_inpainting.model"}, "('MODELS', 'hand-static', 'hand-static')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/hand_static/hand_model.py", "imports": ["numpy", "torchvision", "cv2", "os", "torch", "sys", "PIL"], "module": "modelscope.models.cv.hand_static.hand_model"}, "('MODELS', 'image-depth-estimation', 'bts-depth-estimation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_depth_estimation_bts/depth_estimation_bts_model.py", "imports": ["os", "torch"], "module": "modelscope.models.cv.image_depth_estimation_bts.depth_estimation_bts_model"}, "('MODELS', 'language-guided-video-summarization', 'clip-it-language-guided-video-summarization')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/language_guided_video_summarization/summarizer.py", "imports": ["videofeatures_clipit", "numpy", "bmt_clipit", "typing", "os", "torch", "argparse"], "module": "modelscope.models.cv.language_guided_video_summarization.summarizer"}, "('MODELS', 'image-segmentation', 'm2fp')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_human_parsing/m2fp_net.py", "imports": ["os", "torch", "typing"], "module": "modelscope.models.cv.image_human_parsing.m2fp_net"}, "('MODELS', 'movie-scene-segmentation', 'resnet50-bert')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/movie_scene_segmentation/model.py", "imports": ["einops", "math", "numpy", "torchvision", "typing", "os", "shotdetect_scenedetect_lgss", "torch", "PIL"], "module": "modelscope.models.cv.movie_scene_segmentation.model"}, "('MODELS', 'image-multi-view-depth-estimation', 'image-casmvs-depth-estimation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_mvs_depth_estimation/casmvs_model.py", "imports": ["easydict", "numpy", "cv2", "os", "torch"], "module": "modelscope.models.cv.image_mvs_depth_estimation.casmvs_model"}, "('MODELS', 'image-segmentation', 'swinL-panoptic-segmentation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_panoptic_segmentation/panseg_model.py", "imports": ["os", "torch"], "module": "modelscope.models.cv.image_panoptic_segmentation.panseg_model"}, "('MODELS', 'image-segmentation', 'r50-panoptic-segmentation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_panoptic_segmentation/r50_panseg_model.py", "imports": ["easycv"], "module": "modelscope.models.cv.image_panoptic_segmentation.r50_panseg_model"}, "('MODELS', 'image-portrait-enhancement', 'gpen')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_portrait_enhancement/image_portrait_enhancement.py", "imports": ["os", "torch", "math", "typing"], "module": "modelscope.models.cv.image_portrait_enhancement.image_portrait_enhancement"}, "('MODELS', 'facial-expression-recognition', 'fer')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/facial_expression_recognition/fer/facial_expression_recognition.py", "imports": ["numpy", "cv2", "os", "torch", "PIL"], "module": "modelscope.models.cv.facial_expression_recognition.fer.facial_expression_recognition"}, "('MODELS', 'image-classification', 'ClassificationModel')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_classification/mmcls_model.py", "imports": ["os"], "module": "modelscope.models.cv.image_classification.mmcls_model"}, "('MODELS', 'image-classification', 'content-check')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_classification/resnet50_cc.py", "imports": ["math", "torchvision", "os", "torch", "collections"], "module": "modelscope.models.cv.image_classification.resnet50_cc"}, "('BACKBONES', 'default', 'BEiTv2')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_classification/backbones/beit_v2.py", "imports": ["einops", "math", "itertools", "typing", "os", "torch", "functools", "warnings", "collections", "mmcv", "mmcls"], "module": "modelscope.models.cv.image_classification.backbones.beit_v2"}, "('BACKBONES', 'default', 'NextViT')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_classification/backbones/nextvit.py", "imports": ["einops", "math", "itertools", "typing", "os", "torch", "functools", "warnings", "collections", "mmcv", "mmcls"], "module": "modelscope.models.cv.image_classification.backbones.nextvit"}, "('MODELS', 'video-depth-estimation', 'dro-resnet18-depth-estimation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_depth_estimation/dro_model.py", "imports": ["numpy", "cv2", "os", "torch", "glob", "tqdm"], "module": "modelscope.models.cv.video_depth_estimation.dro_model"}, "('MODELS', 'face-recognition', 'rts-backbone')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_recognition/torchkit/rts_backbone.py", "imports": ["os", "collections", "math", "torch"], "module": "modelscope.models.cv.face_recognition.torchkit.rts_backbone"}, "('MODELS', 'video-summarization', 'pgl-video-summarization')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_summarization/summarizer.py", "imports": ["os", "numpy", "torch", "typing"], "module": "modelscope.models.cv.video_summarization.summarizer"}, "('PREPROCESSORS', 'cv', 'image-driving-perception-preprocessor')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_driving_perception/preprocessor.py", "imports": ["torch", "cv2", "numpy", "typing"], "module": "modelscope.models.cv.image_driving_perception.preprocessor"}, "('MODELS', 'image-driving-perception', 'yolopv2')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_driving_perception/image_driving_percetion_model.py", "imports": ["numpy", "cv2", "typing", "os", "torch"], "module": "modelscope.models.cv.image_driving_perception.image_driving_percetion_model"}, "('MODELS', 'pointcloud-sceneflow-estimation', 'rcp-sceneflow-estimation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/pointcloud_sceneflow_estimation/rcp_model.py", "imports": ["os", "numpy", "torch"], "module": "modelscope.models.cv.pointcloud_sceneflow_estimation.rcp_model"}, "('MODELS', 'human-reconstruction', 'human-reconstruction')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/human_reconstruction/Reconstruction.py", "imports": ["numpy", "torchvision", "cv2", "typing", "os", "torch", "skimage", "PIL"], "module": "modelscope.models.cv.human_reconstruction.Reconstruction"}, "('MODELS', 'lineless-table-recognition', 'LoreModel')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/table_recognition/model_lore.py", "imports": ["math", "numpy", "typing", "os", "torch", "copy"], "module": "modelscope.models.cv.table_recognition.model_lore"}, "('MODELS', 'referring-video-object-segmentation', 'swinT-referring-video-object-segmentation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/referring_video_object_segmentation/model.py", "imports": ["os", "torch", "typing"], "module": "modelscope.models.cv.referring_video_object_segmentation.model"}, "('MODELS', 'video-super-resolution', 'msrresnet-lite')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_super_resolution/msrresnet_lite_model.py", "imports": ["os", "torch", "functools", "typing"], "module": "modelscope.models.cv.video_super_resolution.msrresnet_lite_model"}, "('MODELS', 'video-super-resolution', 'real-basicvsr')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_super_resolution/real_basicvsr_for_video_super_resolution.py", "imports": ["os", "torch", "typing"], "module": "modelscope.models.cv.video_super_resolution.real_basicvsr_for_video_super_resolution"}, "('MODELS', 'image-depth-estimation', 'newcrfs-depth-estimation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_depth_estimation/newcrfs_model.py", "imports": ["os", "numpy", "torch"], "module": "modelscope.models.cv.image_depth_estimation.newcrfs_model"}, "('MODELS', 'video-human-matting', 'video-human-matting')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_human_matting/model.py", "imports": ["numpy", "torchvision", "typing", "os", "torch"], "module": "modelscope.models.cv.video_human_matting.model"}, "('HEADS', 'default', 'FCNMaskNHead')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/object_detection/mmdet_ms/roi_heads/mask_heads/fcn_mask_head.py", "imports": ["numpy", "mmdet", "torch", "warnings", "mmcv"], "module": "modelscope.models.cv.object_detection.mmdet_ms.roi_heads.mask_heads.fcn_mask_head"}, "('HEADS', 'default', 'ConvFCBBoxNHead')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/object_detection/mmdet_ms/roi_heads/bbox_heads/convfc_bbox_head.py", "imports": ["mmdet", "torch"], "module": "modelscope.models.cv.object_detection.mmdet_ms.roi_heads.bbox_heads.convfc_bbox_head"}, "('HEADS', 'default', 'Shared2FCBBoxNHead')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/object_detection/mmdet_ms/roi_heads/bbox_heads/convfc_bbox_head.py", "imports": ["mmdet", "torch"], "module": "modelscope.models.cv.object_detection.mmdet_ms.roi_heads.bbox_heads.convfc_bbox_head"}, "('HEADS', 'default', 'Shared4Conv1FCBBoxNHead')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/object_detection/mmdet_ms/roi_heads/bbox_heads/convfc_bbox_head.py", "imports": ["mmdet", "torch"], "module": "modelscope.models.cv.object_detection.mmdet_ms.roi_heads.bbox_heads.convfc_bbox_head"}, "('BACKBONES', 'default', 'ViT')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/object_detection/mmdet_ms/backbones/vit.py", "imports": ["math", "mmdet", "torch", "functools", "timm"], "module": "modelscope.models.cv.object_detection.mmdet_ms.backbones.vit"}, "('NECKS', 'default', 'FPNF')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/object_detection/mmdet_ms/necks/fpn.py", "imports": ["mmdet", "mmcv", "torch"], "module": "modelscope.models.cv.object_detection.mmdet_ms.necks.fpn"}, "('HEADS', 'default', 'AnchorNHead')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/object_detection/mmdet_ms/dense_heads/anchor_head.py", "imports": ["mmdet"], "module": "modelscope.models.cv.object_detection.mmdet_ms.dense_heads.anchor_head"}, "('HEADS', 'default', 'RPNNHead')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/object_detection/mmdet_ms/dense_heads/rpn_head.py", "imports": ["mmdet", "mmcv", "torch", "copy"], "module": "modelscope.models.cv.object_detection.mmdet_ms.dense_heads.rpn_head"}, "('MODELS', 'image-object-detection', 'YOLOX')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/object_detection/yolox_pai.py", "imports": ["easycv"], "module": "modelscope.models.cv.object_detection.yolox_pai"}, "('MODELS', 'image-object-detection', 'image-object-detection-auto')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/object_detection/yolox_pai.py", "imports": ["easycv"], "module": "modelscope.models.cv.object_detection.yolox_pai"}, "('MODELS', 'domain-specific-object-detection', 'YOLOX')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/object_detection/yolox_pai.py", "imports": ["easycv"], "module": "modelscope.models.cv.object_detection.yolox_pai"}, "('MODELS', 'image-object-detection', 'DINO')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/object_detection/dino.py", "imports": ["easycv"], "module": "modelscope.models.cv.object_detection.dino"}, "('MODELS', 'human-detection', 'detection')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/object_detection/mmdet_model.py", "imports": ["os", "numpy", "torch"], "module": "modelscope.models.cv.object_detection.mmdet_model"}, "('MODELS', 'image-object-detection', 'detection')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/object_detection/mmdet_model.py", "imports": ["os", "numpy", "torch"], "module": "modelscope.models.cv.object_detection.mmdet_model"}, "('MODELS', 'video-object-detection', 'realtime-video-object-detection')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/stream_yolo/realtime_video_detector.py", "imports": ["time", "numpy", "json", "cv2", "os", "torch", "logging", "argparse", "tqdm"], "module": "modelscope.models.cv.stream_yolo.realtime_video_detector"}, "('MODELS', 'image-reid-person', 'passvitb')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_reid_person/pass_model.py", "imports": ["os", "enum", "torch"], "module": "modelscope.models.cv.image_reid_person.pass_model"}, "('MODELS', 'ocr-recognition', 'OCRRecognition')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/ocr_recognition/model.py", "imports": ["os", "torch"], "module": "modelscope.models.cv.ocr_recognition.model"}, "('PREPROCESSORS', 'cv', 'ocr-recognition')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/ocr_recognition/preprocessor.py", "imports": ["numpy", "cv2", "os", "torch", "PIL"], "module": "modelscope.models.cv.ocr_recognition.preprocessor"}, "('MODELS', 'image-segmentation', 'vision-middleware')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/vision_middleware/model.py", "imports": ["os", "torch", "json", "typing"], "module": "modelscope.models.cv.vision_middleware.model"}, "('MODELS', 'image-colorization', 'ddcolor')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_colorization/ddcolor/ddcolor_for_image_colorization.py", "imports": ["numpy", "typing", "os", "torch", "copy"], "module": "modelscope.models.cv.image_colorization.ddcolor.ddcolor_for_image_colorization"}, "('MODELS', 'image-fewshot-detection', 'defrcn')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_defrcn_fewshot/defrcn_for_fewshot.py", "imports": ["os", "torch", "typing"], "module": "modelscope.models.cv.image_defrcn_fewshot.defrcn_for_fewshot"}, "('DETECTORS', 'default', 'Petr3D')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/object_detection_3d/depe/mmdet3d_plugin/models/detectors/petr3d.py", "imports": ["numpy", "mmdet", "torch", "mmcv", "mmdet3d"], "module": "modelscope.models.cv.object_detection_3d.depe.mmdet3d_plugin.models.detectors.petr3d"}, "('BACKBONES', 'default', 'VoVNet')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/object_detection_3d/depe/mmdet3d_plugin/models/backbones/vovnet.py", "imports": ["mmdet", "collections", "mmcv", "torch"], "module": "modelscope.models.cv.object_detection_3d.depe.mmdet3d_plugin.models.backbones.vovnet"}, "('NECKS', 'default', 'CPFPN')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/object_detection_3d/depe/mmdet3d_plugin/models/necks/cp_fpn.py", "imports": ["mmdet", "mmcv", "torch"], "module": "modelscope.models.cv.object_detection_3d.depe.mmdet3d_plugin.models.necks.cp_fpn"}, "('TRANSFORMER', 'default', 'PETRDNTransformer')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/object_detection_3d/depe/mmdet3d_plugin/models/utils/petr_transformer.py", "imports": ["math", "torch", "copy", "warnings", "mmdet", "mmcv", "typing"], "module": "modelscope.models.cv.object_detection_3d.depe.mmdet3d_plugin.models.utils.petr_transformer"}, "('TRANSFORMER_LAYER', 'default', 'PETRTransformerDecoderLayer')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/object_detection_3d/depe/mmdet3d_plugin/models/utils/petr_transformer.py", "imports": ["math", "torch", "copy", "warnings", "mmdet", "mmcv", "typing"], "module": "modelscope.models.cv.object_detection_3d.depe.mmdet3d_plugin.models.utils.petr_transformer"}, "('ATTENTION', 'default', 'PETRMultiheadAttention')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/object_detection_3d/depe/mmdet3d_plugin/models/utils/petr_transformer.py", "imports": ["math", "torch", "copy", "warnings", "mmdet", "mmcv", "typing"], "module": "modelscope.models.cv.object_detection_3d.depe.mmdet3d_plugin.models.utils.petr_transformer"}, "('TRANSFORMER_LAYER_SEQUENCE', 'default', 'PETRTransformerEncoder')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/object_detection_3d/depe/mmdet3d_plugin/models/utils/petr_transformer.py", "imports": ["math", "torch", "copy", "warnings", "mmdet", "mmcv", "typing"], "module": "modelscope.models.cv.object_detection_3d.depe.mmdet3d_plugin.models.utils.petr_transformer"}, "('TRANSFORMER_LAYER_SEQUENCE', 'default', 'PETRTransformerDecoder')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/object_detection_3d/depe/mmdet3d_plugin/models/utils/petr_transformer.py", "imports": ["math", "torch", "copy", "warnings", "mmdet", "mmcv", "typing"], "module": "modelscope.models.cv.object_detection_3d.depe.mmdet3d_plugin.models.utils.petr_transformer"}, "('POSITIONAL_ENCODING', 'default', 'SinePositionalEncoding3D')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/object_detection_3d/depe/mmdet3d_plugin/models/utils/positional_encoding.py", "imports": ["mmcv", "math", "torch"], "module": "modelscope.models.cv.object_detection_3d.depe.mmdet3d_plugin.models.utils.positional_encoding"}, "('HEADS', 'default', 'PETRv2DEDNHead')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/object_detection_3d/depe/mmdet3d_plugin/models/dense_heads/petrv2_dednhead.py", "imports": ["math", "numpy", "mmdet", "torch", "copy", "mmcv", "mmdet3d"], "module": "modelscope.models.cv.object_detection_3d.depe.mmdet3d_plugin.models.dense_heads.petrv2_dednhead"}, "('BBOX_ASSIGNERS', 'default', 'HungarianAssigner3D')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/object_detection_3d/depe/mmdet3d_plugin/core/bbox/assigners/hungarian_assigner_3d.py", "imports": ["scipy", "mmdet", "torch"], "module": "modelscope.models.cv.object_detection_3d.depe.mmdet3d_plugin.core.bbox.assigners.hungarian_assigner_3d"}, "('BBOX_CODERS', 'default', 'NMSFreeCoder')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/object_detection_3d/depe/mmdet3d_plugin/core/bbox/coders/nms_free_coder.py", "imports": ["mmdet", "torch"], "module": "modelscope.models.cv.object_detection_3d.depe.mmdet3d_plugin.core.bbox.coders.nms_free_coder"}, "('MATCH_COST', 'default', 'BBox3DL1Cost')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/object_detection_3d/depe/mmdet3d_plugin/core/bbox/match_costs/match_cost.py", "imports": ["mmdet", "torch"], "module": "modelscope.models.cv.object_detection_3d.depe.mmdet3d_plugin.core.bbox.match_costs.match_cost"}, "('PIPELINES', 'default', 'PadMultiViewImage')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/object_detection_3d/depe/mmdet3d_plugin/datasets/pipelines/transform_3d.py", "imports": ["torch", "copy", "numpy", "mmdet", "mmcv", "mmdet3d", "PIL"], "module": "modelscope.models.cv.object_detection_3d.depe.mmdet3d_plugin.datasets.pipelines.transform_3d"}, "('PIPELINES', 'default', 'NormalizeMultiviewImage')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/object_detection_3d/depe/mmdet3d_plugin/datasets/pipelines/transform_3d.py", "imports": ["torch", "copy", "numpy", "mmdet", "mmcv", "mmdet3d", "PIL"], "module": "modelscope.models.cv.object_detection_3d.depe.mmdet3d_plugin.datasets.pipelines.transform_3d"}, "('PIPELINES', 'default', 'ResizeCropFlipImage')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/object_detection_3d/depe/mmdet3d_plugin/datasets/pipelines/transform_3d.py", "imports": ["torch", "copy", "numpy", "mmdet", "mmcv", "mmdet3d", "PIL"], "module": "modelscope.models.cv.object_detection_3d.depe.mmdet3d_plugin.datasets.pipelines.transform_3d"}, "('PIPELINES', 'default', 'LoadMultiViewImageFromMultiSweepsFiles')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/object_detection_3d/depe/mmdet3d_plugin/datasets/pipelines/loading.py", "imports": ["numpy", "mmdet", "mmcv"], "module": "modelscope.models.cv.object_detection_3d.depe.mmdet3d_plugin.datasets.pipelines.loading"}, "('DATASETS', 'default', 'CustomNuScenesDataset')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/object_detection_3d/depe/mmdet3d_plugin/datasets/nuscenes_dataset.py", "imports": ["numpy", "mmdet", "mmdet3d"], "module": "modelscope.models.cv.object_detection_3d.depe.mmdet3d_plugin.datasets.nuscenes_dataset"}, "('MODELS', 'object-detection-3d', 'depe')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/object_detection_3d/depe/depe_detect.py", "imports": ["os", "numpy", "torch", "typing"], "module": "modelscope.models.cv.object_detection_3d.depe.depe_detect"}, "('MODELS', 'image-debanding', 'rrdb')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_debanding/rrdb/rrdb_image_debanding.py", "imports": ["os", "torch", "typing"], "module": "modelscope.models.cv.image_debanding.rrdb.rrdb_image_debanding"}, "('MODELS', 'text-driven-segmentation', 'text-driven-segmentation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/text_driven_segmentation/lseg_model.py", "imports": ["numpy", "json", "typing", "os", "torch", "PIL"], "module": "modelscope.models.cv.text_driven_segmentation.lseg_model"}, "('MODELS', 'video-deinterlace', 'video-deinterlace')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_deinterlace/UNet_for_video_deinterlace.py", "imports": ["os", "torch", "typing", "copy"], "module": "modelscope.models.cv.video_deinterlace.UNet_for_video_deinterlace"}, "('MODELS', 'image-quality-assessment-degradation', 'image-quality-assessment-degradation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_quality_assessment_degradation/image_quality_assessment_degradation.py", "imports": ["os", "torch", "typing"], "module": "modelscope.models.cv.image_quality_assessment_degradation.image_quality_assessment_degradation"}, "('MODELS', 'image-segmentation', 'cascade_mask_rcnn_swin')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_instance_segmentation/model.py", "imports": ["os", "torch", "typing"], "module": "modelscope.models.cv.image_instance_segmentation.model"}, "('MODELS', 'image-segmentation', 'maskdino_swin')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_instance_segmentation/maskdino_model.py", "imports": ["os", "torch", "typing"], "module": "modelscope.models.cv.image_instance_segmentation.maskdino_model"}, "('MODELS', 'image-quality-assessment-mos', 'image-quality-assessment-man')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_quality_assessment_man/image_quality_assessment_man.py", "imports": ["os", "torch", "typing"], "module": "modelscope.models.cv.image_quality_assessment_man.image_quality_assessment_man"}, "('MODELS', 'controllable-image-generation', 'controllable-image-generation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/controllable_image_generation/controlnet.py", "imports": ["einops", "control_ldm", "math", "numpy", "cv2", "typing", "os", "random", "tempfile", "torch", "sys", "PIL"], "module": "modelscope.models.cv.controllable_image_generation.controlnet"}, "('MODELS', 'crowd-counting', 'HRNetCrowdCounting')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/crowd_counting/cc_model.py", "imports": ["os", "torch", "typing"], "module": "modelscope.models.cv.crowd_counting.cc_model"}, "('MODELS', 'semantic-segmentation', 'detection')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/salient_detection/salient_model.py", "imports": ["torchvision", "cv2", "os", "torch", "PIL"], "module": "modelscope.models.cv.salient_detection.salient_model"}, "('MODELS', 'video-object-detection', 'longshortnet')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_streaming_perception/longshortnet/longshortnet.py", "imports": ["time", "numpy", "json", "cv2", "os", "torch", "logging", "argparse", "tqdm"], "module": "modelscope.models.cv.video_streaming_perception.longshortnet.longshortnet"}, "('MODELS', 'video-object-segmentation', 'video-object-segmentation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_object_segmentation/model.py", "imports": ["os", "torch", "typing"], "module": "modelscope.models.cv.video_object_segmentation.model"}, "('MODELS', 'domain-specific-object-detection', 'tinynas-damoyolo')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/tinynas_detection/tinynas_damoyolo.py", "imports": [], "module": "modelscope.models.cv.tinynas_detection.tinynas_damoyolo"}, "('MODELS', 'image-object-detection', 'tinynas-damoyolo')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/tinynas_detection/tinynas_damoyolo.py", "imports": [], "module": "modelscope.models.cv.tinynas_detection.tinynas_damoyolo"}, "('MODELS', 'image-object-detection', 'tinynas-detection')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/tinynas_detection/tinynas_detector.py", "imports": [], "module": "modelscope.models.cv.tinynas_detection.tinynas_detector"}, "('MODELS', 'face-2d-keypoints', 'flc')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/facial_landmark_confidence/flc/facial_landmark_confidence.py", "imports": ["numpy", "cv2", "os", "torch", "PIL"], "module": "modelscope.models.cv.facial_landmark_confidence.flc.facial_landmark_confidence"}, "('MODELS', 'image-face-fusion', 'image-face-fusion')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_face_fusion/image_face_fusion.py", "imports": ["numpy", "torchvision", "typing", "cv2", "os", "torch", "collections", "PIL"], "module": "modelscope.models.cv.image_face_fusion.image_face_fusion"}, "('MODELS', 'image-paintbyexample', 'Stablediffusion-Paintbyexample')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_paintbyexample/model.py", "imports": ["omegaconf", "typing", "os", "paint_ldm", "torch"], "module": "modelscope.models.cv.image_paintbyexample.model"}, "('MODELS', 'image-deblurring', 'nafnet')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_deblur/nafnet_for_image_deblur.py", "imports": ["os", "torch", "typing"], "module": "modelscope.models.cv.image_deblur.nafnet_for_image_deblur"}, "('MODELS', 'image-demoireing', 'image-restoration')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_restoration/image_restoration_model.py", "imports": ["os", "numpy", "torch", "cv2"], "module": "modelscope.models.cv.image_restoration.image_restoration_model"}, "('MODELS', 'ocr-detection', 'OCRDetection')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/ocr_detection/model.py", "imports": ["os", "numpy", "torch", "typing"], "module": "modelscope.models.cv.ocr_detection.model"}, "('PREPROCESSORS', 'cv', 'ocr-detection')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/ocr_detection/preprocessor.py", "imports": ["math", "numpy", "cv2", "typing", "os", "torch", "PIL"], "module": "modelscope.models.cv.ocr_detection.preprocessor"}, "('MODELS', 'video-text-retrieval', 'vop-retrieval-model')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/vop_retrieval/model.py", "imports": ["os", "torch"], "module": "modelscope.models.cv.vop_retrieval.model"}, "('MODELS', 'video-text-retrieval', 'vop-retrieval-model-se')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/vop_retrieval/model_se.py", "imports": ["os", "torch"], "module": "modelscope.models.cv.vop_retrieval.model_se"}, "('MODELS', 'face-attribute-recognition', 'fairface')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_attribute_recognition/fair_face/face_attribute_recognition.py", "imports": ["numpy", "torchvision", "cv2", "os", "torch", "PIL"], "module": "modelscope.models.cv.face_attribute_recognition.fair_face.face_attribute_recognition"}, "('MODELS', 'nerf-recon-acc', 'nerf-recon-acc')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/nerf_recon_acc/nerf_recon_acc.py", "imports": ["time", "numpy", "cv2", "os", "torch", "glob", "tqdm"], "module": "modelscope.models.cv.nerf_recon_acc.nerf_recon_acc"}, "('PREPROCESSORS', 'cv', 'nerf-recon-acc-preprocessor')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/nerf_recon_acc/nerf_preprocess.py", "imports": ["subprocess", "numpy", "cv2", "typing", "os", "tensorflow", "glob"], "module": "modelscope.models.cv.nerf_recon_acc.nerf_preprocess"}, "('MODELS', 'image-body-reshaping', 'image-body-reshaping')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_body_reshaping/image_body_reshaping.py", "imports": ["numpy", "cv2", "typing", "os", "torch"], "module": "modelscope.models.cv.image_body_reshaping.image_body_reshaping"}, "('MODELS', 'body-2d-keypoints', 'body-2d-keypoints')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/body_2d_keypoints/hrnet_v2.py", "imports": ["os", "numpy", "torch"], "module": "modelscope.models.cv.body_2d_keypoints.hrnet_v2"}, "('MODELS', 'image-classification', 'bnext')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_binary_quant_classification/binary_quant_model.py", "imports": ["os", "collections", "torch"], "module": "modelscope.models.cv.image_binary_quant_classification.binary_quant_model"}, "('MODELS', 'video-inpainting', 'video-inpainting')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_inpainting/inpainting_model.py", "imports": ["torch", "numpy", "math", "torchvision"], "module": "modelscope.models.cv.video_inpainting.inpainting_model"}, "('MODELS', 'image-classification', 'EasyRobustModel')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/robust_image_classification/easyrobust_model.py", "imports": ["os", "torch"], "module": "modelscope.models.cv.robust_image_classification.easyrobust_model"}, "('MODELS', 'face-human-hand-detection', 'face-human-hand-detection')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_human_hand_detection/det_infer.py", "imports": ["torch", "numpy", "cv2"], "module": "modelscope.models.cv.face_human_hand_detection.det_infer"}, "('MODELS', 'bad-image-detecting', 'bad-image-detecting')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/bad_image_detecting/bad_image_detecting.py", "imports": ["numpy", "torchvision", "typing", "os", "torch"], "module": "modelscope.models.cv.bad_image_detecting.bad_image_detecting"}, "('MODELS', 'face-detection', 'mtcnn')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_detection/mtcnn/models/detector.py", "imports": ["os", "numpy", "torch", "PIL"], "module": "modelscope.models.cv.face_detection.mtcnn.models.detector"}, "('MODELS', 'face-detection', 'mogface')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_detection/mogface/models/detectors.py", "imports": ["os", "numpy", "torch", "cv2"], "module": "modelscope.models.cv.face_detection.mogface.models.detectors"}, "('DETECTORS', 'default', 'CustomSingleStageDetector')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_detection/scrfd/mmdet_patch/models/detectors/single_stage.py", "imports": ["mmdet", "torch"], "module": "modelscope.models.cv.face_detection.scrfd.mmdet_patch.models.detectors.single_stage"}, "('DETECTORS', 'default', 'SCRFD')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_detection/scrfd/mmdet_patch/models/detectors/scrfd.py", "imports": ["mmdet", "torch"], "module": "modelscope.models.cv.face_detection.scrfd.mmdet_patch.models.detectors.scrfd"}, "('DETECTORS', 'default', 'TinyMog')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_detection/scrfd/mmdet_patch/models/detectors/tinymog.py", "imports": ["mmdet", "torch"], "module": "modelscope.models.cv.face_detection.scrfd.mmdet_patch.models.detectors.tinymog"}, "('BACKBONES', 'default', 'MobileNetV1')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_detection/scrfd/mmdet_patch/models/backbones/mobilenet.py", "imports": ["mmdet", "mmcv", "torch"], "module": "modelscope.models.cv.face_detection.scrfd.mmdet_patch.models.backbones.mobilenet"}, "('BACKBONES', 'default', 'ResNetV1e')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_detection/scrfd/mmdet_patch/models/backbones/resnet.py", "imports": ["mmdet", "mmcv", "torch"], "module": "modelscope.models.cv.face_detection.scrfd.mmdet_patch.models.backbones.resnet"}, "('HEADS', 'default', 'SCRFDHead')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_detection/scrfd/mmdet_patch/models/dense_heads/scrfd_head.py", "imports": ["numpy", "mmdet", "mmcv", "torch"], "module": "modelscope.models.cv.face_detection.scrfd.mmdet_patch.models.dense_heads.scrfd_head"}, "('DATASETS', 'default', 'RetinaFaceDataset')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_detection/scrfd/mmdet_patch/datasets/retinaface.py", "imports": ["numpy", "mmdet"], "module": "modelscope.models.cv.face_detection.scrfd.mmdet_patch.datasets.retinaface"}, "('PIPELINES', 'default', 'DefaultFormatBundleV2')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_detection/scrfd/mmdet_patch/datasets/pipelines/formating.py", "imports": ["numpy", "mmdet", "mmcv", "torch"], "module": "modelscope.models.cv.face_detection.scrfd.mmdet_patch.datasets.pipelines.formating"}, "('PIPELINES', 'default', 'LoadAnnotationsV2')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_detection/scrfd/mmdet_patch/datasets/pipelines/loading.py", "imports": ["os", "mmdet", "numpy", "pycocotools"], "module": "modelscope.models.cv.face_detection.scrfd.mmdet_patch.datasets.pipelines.loading"}, "('PIPELINES', 'default', 'RotateV2')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_detection/scrfd/mmdet_patch/datasets/pipelines/auto_augment.py", "imports": ["copy", "numpy", "mmdet", "mmcv", "cv2"], "module": "modelscope.models.cv.face_detection.scrfd.mmdet_patch.datasets.pipelines.auto_augment"}, "('PIPELINES', 'default', 'ResizeV2')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_detection/scrfd/mmdet_patch/datasets/pipelines/transforms.py", "imports": ["numpy", "mmdet", "mmcv"], "module": "modelscope.models.cv.face_detection.scrfd.mmdet_patch.datasets.pipelines.transforms"}, "('PIPELINES', 'default', 'RandomFlipV2')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_detection/scrfd/mmdet_patch/datasets/pipelines/transforms.py", "imports": ["numpy", "mmdet", "mmcv"], "module": "modelscope.models.cv.face_detection.scrfd.mmdet_patch.datasets.pipelines.transforms"}, "('PIPELINES', 'default', 'RandomSquareCrop')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_detection/scrfd/mmdet_patch/datasets/pipelines/transforms.py", "imports": ["numpy", "mmdet", "mmcv"], "module": "modelscope.models.cv.face_detection.scrfd.mmdet_patch.datasets.pipelines.transforms"}, "('MODELS', 'face-detection', 'scrfd')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_detection/scrfd/scrfd_detect.py", "imports": ["numpy", "typing", "os", "torch", "copy"], "module": "modelscope.models.cv.face_detection.scrfd.scrfd_detect"}, "('MODELS', 'card-detection', 'scrfd')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_detection/scrfd/scrfd_detect.py", "imports": ["numpy", "typing", "os", "torch", "copy"], "module": "modelscope.models.cv.face_detection.scrfd.scrfd_detect"}, "('PREPROCESSORS', 'cv', 'object-detection-scrfd')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_detection/scrfd/preprocessor.py", "imports": ["numpy", "typing", "PIL"], "module": "modelscope.models.cv.face_detection.scrfd.preprocessor"}, "('MODELS', 'face-detection', 'tinymog')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_detection/scrfd/tinymog_detect.py", "imports": ["os", "torch", "typing", "copy"], "module": "modelscope.models.cv.face_detection.scrfd.tinymog_detect"}, "('MODELS', 'face-detection', 'ulfd')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_detection/ulfd_slim/detection.py", "imports": ["os", "numpy", "torch", "cv2"], "module": "modelscope.models.cv.face_detection.ulfd_slim.detection"}, "('MODELS', 'face-detection', 'retinaface')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_detection/retinaface/detection.py", "imports": ["torch", "numpy", "cv2"], "module": "modelscope.models.cv.face_detection.retinaface.detection"}, "('MODELS', 'video-frame-interpolation', 'video-frame-interpolation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_frame_interpolation/VFINet_for_video_frame_interpolation.py", "imports": ["os", "torch", "typing", "copy"], "module": "modelscope.models.cv.video_frame_interpolation.VFINet_for_video_frame_interpolation"}, "('MODELS', 'image-object-detection', 'vidt')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/vidt/model.py", "imports": ["os", "torch"], "module": "modelscope.models.cv.vidt.model"}, "('MODELS', 'body-3d-keypoints', 'hdformer')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/body_3d_keypoints/hdformer/hdformer_detector.py", "imports": ["os", "numpy", "torch", "typing"], "module": "modelscope.models.cv.body_3d_keypoints.hdformer.hdformer_detector"}, "('MODELS', 'body-3d-keypoints', 'body-3d-keypoints')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/body_3d_keypoints/cannonical_pose/body_3d_pose.py", "imports": ["numpy", "typing", "os", "torch", "logging"], "module": "modelscope.models.cv.body_3d_keypoints.cannonical_pose.body_3d_pose"}, "('MODELS', 'vision-efficient-tuning', 'vision-efficient-tuning')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/vision_efficient_tuning/model.py", "imports": ["torch", "typing"], "module": "modelscope.models.cv.vision_efficient_tuning.model"}, "('MODELS', 'image-quality-assessment-mos', 'image-quality-assessment-mos')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_quality_assessment_mos/image_quality_assessment_mos.py", "imports": ["os", "torch", "typing"], "module": "modelscope.models.cv.image_quality_assessment_mos.image_quality_assessment_mos"}, "('MODELS', 'open-vocabulary-detection', 'open-vocabulary-detection-vild')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/open_vocabulary_detection_vild/vild.py", "imports": ["scipy", "numpy", "typing", "os", "torch", "tensorflow", "clip"], "module": "modelscope.models.cv.open_vocabulary_detection_vild.vild"}, "('MODELS', 'human-wholebody-keypoint', 'human-wholebody-keypoint')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/human_wholebody_keypoint/human_wholebody_keypoint.py", "imports": ["easycv"], "module": "modelscope.models.cv.human_wholebody_keypoint.human_wholebody_keypoint"}, "('MODELS', 'semantic-segmentation', 'ddpm')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_semantic_segmentation/ddpm_segmentation_model.py", "imports": ["os", "torch", "ddpm_guided_diffusion", "typing"], "module": "modelscope.models.cv.image_semantic_segmentation.ddpm_segmentation_model"}, "('HEADS', 'default', 'Mask2FormerHeadFromMMSeg')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_semantic_segmentation/vit_adapter/models/decode_heads/mask2former_head_from_mmseg.py", "imports": ["mmdet", "mmcv", "torch", "copy"], "module": "modelscope.models.cv.image_semantic_segmentation.vit_adapter.models.decode_heads.mask2former_head_from_mmseg"}, "('DETECTORS', 'default', 'EncoderDecoderMask2Former')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_semantic_segmentation/vit_adapter/models/segmentors/encoder_decoder_mask2former.py", "imports": ["mmdet", "torch"], "module": "modelscope.models.cv.image_semantic_segmentation.vit_adapter.models.segmentors.encoder_decoder_mask2former"}, "('BACKBONES', 'default', 'BEiTAdapter')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_semantic_segmentation/vit_adapter/models/backbone/beit_adapter.py", "imports": ["math", "mmdet", "torch", "timm", "logging"], "module": "modelscope.models.cv.image_semantic_segmentation.vit_adapter.models.backbone.beit_adapter"}, "('BACKBONES', 'default', 'BASEBEiT')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_semantic_segmentation/vit_adapter/models/backbone/base/beit.py", "imports": ["math", "functools", "torch", "timm", "mmdet", "mmcv"], "module": "modelscope.models.cv.image_semantic_segmentation.vit_adapter.models.backbone.base.beit"}, "('PIPELINES', 'default', 'ResizeToMultiple')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_semantic_segmentation/vit_adapter/utils/data_process_func.py", "imports": ["mmdet", "mmcv"], "module": "modelscope.models.cv.image_semantic_segmentation.vit_adapter.utils.data_process_func"}, "('HEADS', 'default', 'MaskFormerSemanticHead')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_semantic_segmentation/pan_merge/maskformer_semantic_head.py", "imports": ["mmdet", "torch"], "module": "modelscope.models.cv.image_semantic_segmentation.pan_merge.maskformer_semantic_head"}, "('MODELS', 'image-segmentation', 'swinL-semantic-segmentation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_semantic_segmentation/semantic_seg_model.py", "imports": ["os", "numpy", "torch"], "module": "modelscope.models.cv.image_semantic_segmentation.semantic_seg_model"}, "('MODELS', 'image-segmentation', 'vitadapter-semantic-segmentation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_semantic_segmentation/semantic_seg_model.py", "imports": ["os", "numpy", "torch"], "module": "modelscope.models.cv.image_semantic_segmentation.semantic_seg_model"}, "('MODELS', 'image-segmentation', 'Segformer')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_semantic_segmentation/segformer.py", "imports": ["easycv"], "module": "modelscope.models.cv.image_semantic_segmentation.segformer"}, "('MODELS', 'shop-segmentation', 'shop-segmentation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/shop_segmentation/shop_seg_model.py", "imports": ["numpy", "json", "typing", "os", "torch", "PIL"], "module": "modelscope.models.cv.shop_segmentation.shop_seg_model"}, "('MODELS', 'image-super-resolution', 'ecbsr')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/super_resolution/ecbsr_model.py", "imports": ["os", "torch", "typing"], "module": "modelscope.models.cv.super_resolution.ecbsr_model"}, "('MODELS', 'indoor-layout-estimation', 'panovit-layout-estimation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/indoor_layout_estimation/panovit.py", "imports": ["os", "numpy", "yacs", "torch"], "module": "modelscope.models.cv.indoor_layout_estimation.panovit"}, "('MODELS', 'hand-2d-keypoints', 'HRNet-Hand2D-Keypoints')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/hand_2d_keypoints/hand_2d_keypoints.py", "imports": ["easycv"], "module": "modelscope.models.cv.hand_2d_keypoints.hand_2d_keypoints"}, "('MODELS', 'image-skychange', 'image-skychange')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_skychange/skychange_model.py", "imports": ["time", "math", "pdb", "json", "cv2", "typing", "os", "torch", "collections"], "module": "modelscope.models.cv.image_skychange.skychange_model"}, "('PREPROCESSORS', 'cv', 'image-sky-change-preprocessor')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_skychange/preprocessor.py", "imports": ["pdb", "numpy", "json", "numbers", "torchvision", "cv2", "typing", "torch"], "module": "modelscope.models.cv.image_skychange.preprocessor"}, "('HEADS', 'default', 'MaskScoringNRoIHead')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/abnormal_object_detection/mmdet_ms/roi_head/mask_scoring_roi_head.py", "imports": ["mmdet", "torch"], "module": "modelscope.models.cv.abnormal_object_detection.mmdet_ms.roi_head.mask_scoring_roi_head"}, "('ROI_EXTRACTORS', 'default', 'SingleRoINExtractor')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/abnormal_object_detection/mmdet_ms/roi_head/roi_extractors/single_level_roi_extractor.py", "imports": ["mmdet", "mmcv", "torch"], "module": "modelscope.models.cv.abnormal_object_detection.mmdet_ms.roi_head.roi_extractors.single_level_roi_extractor"}, "('MODELS', 'image-object-detection', 'MaskScoring')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/abnormal_object_detection/mmdet_model.py", "imports": ["os", "numpy", "torch"], "module": "modelscope.models.cv.abnormal_object_detection.mmdet_model"}, "('MODELS', 'image-color-enhancement', 'adaint')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_color_enhance/adaint/adaint.py", "imports": ["numbers", "torchvision", "typing", "os", "torch"], "module": "modelscope.models.cv.image_color_enhance.adaint.adaint"}, "('MODELS', 'image-color-enhancement', 'deeplpfnet')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_color_enhance/deeplpf/deeplpf_image_color_enhance.py", "imports": ["os", "torch", "typing"], "module": "modelscope.models.cv.image_color_enhance.deeplpf.deeplpf_image_color_enhance"}, "('MODELS', 'image-color-enhancement', 'csrnet')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_color_enhance/image_color_enhance.py", "imports": ["os", "torch", "typing"], "module": "modelscope.models.cv.image_color_enhance.image_color_enhance"}, "('MODELS', 'face-2d-keypoints', 'face-2d-keypoints')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_2d_keypoints/face_2d_keypoints_align.py", "imports": ["easycv"], "module": "modelscope.models.cv.face_2d_keypoints.face_2d_keypoints_align"}, "('MODELS', 'video-stabilization', 'video-stabilization')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_stabilization/DUTRAFTStabilizer.py", "imports": ["math", "numpy", "cv2", "typing", "os", "tempfile", "torch", "sys"], "module": "modelscope.models.cv.video_stabilization.DUTRAFTStabilizer"}, "('MODELS', 'video-instance-segmentation', 'swinb-video-instance-segmentation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_instance_segmentation/video_knet.py", "imports": ["mmdet", "torch"], "module": "modelscope.models.cv.video_instance_segmentation.video_knet"}, "('HEADS', 'default', 'KernelFrameIterHeadVideo')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_instance_segmentation/head/kernel_frame_iter_head.py", "imports": ["mmdet", "mmcv", "torch"], "module": "modelscope.models.cv.video_instance_segmentation.head.kernel_frame_iter_head"}, "('HEADS', 'default', 'ConvKernelHeadVideo')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_instance_segmentation/head/kernel_head.py", "imports": ["mmdet", "mmcv", "torch"], "module": "modelscope.models.cv.video_instance_segmentation.head.kernel_head"}, "('HEADS', 'default', 'KernelUpdateHead')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_instance_segmentation/head/kernel_update_head.py", "imports": ["numpy", "mmdet", "mmcv", "torch"], "module": "modelscope.models.cv.video_instance_segmentation.head.kernel_update_head"}, "('HEADS', 'default', 'KernelIterHeadVideo')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_instance_segmentation/head/kernel_iter_head.py", "imports": ["mmdet", "torch"], "module": "modelscope.models.cv.video_instance_segmentation.head.kernel_iter_head"}, "('MATCH_COST', 'default', 'MaskCost')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_instance_segmentation/track/mask_hungarian_assigner.py", "imports": ["scipy", "numpy", "mmdet", "torch"], "module": "modelscope.models.cv.video_instance_segmentation.track.mask_hungarian_assigner"}, "('BBOX_ASSIGNERS', 'default', 'MaskHungarianAssignerVideo')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_instance_segmentation/track/mask_hungarian_assigner.py", "imports": ["scipy", "numpy", "mmdet", "torch"], "module": "modelscope.models.cv.video_instance_segmentation.track.mask_hungarian_assigner"}, "('HEADS', 'default', 'KernelUpdateHeadVideo')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_instance_segmentation/track/kernel_update_head.py", "imports": ["numpy", "mmdet", "mmcv", "torch"], "module": "modelscope.models.cv.video_instance_segmentation.track.kernel_update_head"}, "('NECKS', 'default', 'MSDeformAttnPixelDecoder')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_instance_segmentation/neck/msdeformattn_decoder.py", "imports": ["mmdet", "mmcv", "torch"], "module": "modelscope.models.cv.video_instance_segmentation.neck.msdeformattn_decoder"}, "('MODELS', 'product-segmentation', 'product-segmentation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/product_segmentation/seg_infer.py", "imports": ["torch", "numpy", "cv2", "PIL"], "module": "modelscope.models.cv.product_segmentation.seg_infer"}, "('MODELS', 'product-retrieval-embedding', 'product-retrieval-embedding')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/product_retrieval_embedding/item_model.py", "imports": ["os", "numpy", "torch", "typing"], "module": "modelscope.models.cv.product_retrieval_embedding.item_model"}, "('MODELS', 'image-classification', 'image-probing-model')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_probing_model/model.py", "imports": ["os", "torch", "json", "typing"], "module": "modelscope.models.cv.image_probing_model.model"}, "('MODELS', 'text-to-video-synthesis', 'latent-text-to-video-synthesis')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/video_synthesis/text_to_video_synthesis_model.py", "imports": ["einops", "open_clip", "typing", "os", "torch"], "module": "modelscope.models.multi_modal.video_synthesis.text_to_video_synthesis_model"}, "('MODELS', 'visual-question-answering', 'mplug')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/mplug_for_all_tasks.py", "imports": ["os", "typing"], "module": "modelscope.models.multi_modal.mplug_for_all_tasks"}, "('MODELS', 'image-captioning', 'mplug')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/mplug_for_all_tasks.py", "imports": ["os", "typing"], "module": "modelscope.models.multi_modal.mplug_for_all_tasks"}, "('MODELS', 'image-text-retrieval', 'mplug')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/mplug_for_all_tasks.py", "imports": ["os", "typing"], "module": "modelscope.models.multi_modal.mplug_for_all_tasks"}, "('MODELS', 'video-question-answering', 'hitea')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/mplug_for_all_tasks.py", "imports": ["os", "typing"], "module": "modelscope.models.multi_modal.mplug_for_all_tasks"}, "('MODELS', 'video-captioning', 'hitea')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/mplug_for_all_tasks.py", "imports": ["os", "typing"], "module": "modelscope.models.multi_modal.mplug_for_all_tasks"}, "('MODELS', 'document-vl-embedding', 'vldoc')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/vldoc/model.py", "imports": ["math", "json", "re", "torchvision", "os", "copy", "torch", "sys", "logging"], "module": "modelscope.models.multi_modal.vldoc.model"}, "('MODELS', 'image-captioning', 'ofa')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/ofa_for_all_tasks.py", "imports": ["math", "json", "re", "typing", "string", "os", "torch", "functools"], "module": "modelscope.models.multi_modal.ofa_for_all_tasks"}, "('MODELS', 'ocr-recognition', 'ofa')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/ofa_for_all_tasks.py", "imports": ["math", "json", "re", "typing", "string", "os", "torch", "functools"], "module": "modelscope.models.multi_modal.ofa_for_all_tasks"}, "('MODELS', 'visual-grounding', 'ofa')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/ofa_for_all_tasks.py", "imports": ["math", "json", "re", "typing", "string", "os", "torch", "functools"], "module": "modelscope.models.multi_modal.ofa_for_all_tasks"}, "('MODELS', 'visual-question-answering', 'ofa')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/ofa_for_all_tasks.py", "imports": ["math", "json", "re", "typing", "string", "os", "torch", "functools"], "module": "modelscope.models.multi_modal.ofa_for_all_tasks"}, "('MODELS', 'visual-entailment', 'ofa')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/ofa_for_all_tasks.py", "imports": ["math", "json", "re", "typing", "string", "os", "torch", "functools"], "module": "modelscope.models.multi_modal.ofa_for_all_tasks"}, "('MODELS', 'image-classification', 'ofa')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/ofa_for_all_tasks.py", "imports": ["math", "json", "re", "typing", "string", "os", "torch", "functools"], "module": "modelscope.models.multi_modal.ofa_for_all_tasks"}, "('MODELS', 'text-summarization', 'ofa')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/ofa_for_all_tasks.py", "imports": ["math", "json", "re", "typing", "string", "os", "torch", "functools"], "module": "modelscope.models.multi_modal.ofa_for_all_tasks"}, "('MODELS', 'text-classification', 'ofa')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/ofa_for_all_tasks.py", "imports": ["math", "json", "re", "typing", "string", "os", "torch", "functools"], "module": "modelscope.models.multi_modal.ofa_for_all_tasks"}, "('MODELS', 'auto-speech-recognition', 'ofa')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/ofa_for_all_tasks.py", "imports": ["math", "json", "re", "typing", "string", "os", "torch", "functools"], "module": "modelscope.models.multi_modal.ofa_for_all_tasks"}, "('MODELS', 'sudoku', 'ofa')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/ofa_for_all_tasks.py", "imports": ["math", "json", "re", "typing", "string", "os", "torch", "functools"], "module": "modelscope.models.multi_modal.ofa_for_all_tasks"}, "('MODELS', 'text2sql', 'ofa')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/ofa_for_all_tasks.py", "imports": ["math", "json", "re", "typing", "string", "os", "torch", "functools"], "module": "modelscope.models.multi_modal.ofa_for_all_tasks"}, "('MODELS', 'generative-multi-modal-embedding', 'gemm-generative-multi-modal')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/gemm/gemm_model.py", "imports": ["numpy", "json", "torchvision", "typing", "os", "torch", "PIL"], "module": "modelscope.models.multi_modal.gemm.gemm_model"}, "('MODELS', 'text-to-image-synthesis', 'ofa')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/ofa_for_text_to_image_synthesis_model.py", "imports": ["numpy", "json", "pkg_resources", "torchvision", "typing", "os", "torch", "taming", "PIL"], "module": "modelscope.models.multi_modal.ofa_for_text_to_image_synthesis_model"}, "('MODELS', 'video-temporal-grounding', 'soonet')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/soonet/model.py", "imports": ["os", "torch"], "module": "modelscope.models.multi_modal.soonet.model"}, "('MODELS', 'multi-modal-similarity', 'team-multi-modal-similarity')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/team/team_model.py", "imports": ["numpy", "torchvision", "cv2", "typing", "tokenizers", "torch", "PIL"], "module": "modelscope.models.multi_modal.team.team_model"}, "('MODELS', 'video-multi-modal-embedding', 'video-clip-multi-modal-embedding')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/mmr/models/clip_for_mm_video_embedding.py", "imports": ["uuid", "numpy", "json", "typing", "os", "random", "urllib", "tempfile", "decord", "torch", "PIL"], "module": "modelscope.models.multi_modal.mmr.models.clip_for_mm_video_embedding"}, "('MODELS', 'text-to-image-synthesis', 'diffusion-text-to-image-synthesis')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/diffusion/model.py", "imports": ["numpy", "json", "typing", "os", "torch"], "module": "modelscope.models.multi_modal.diffusion.model"}, "('MODELS', 'backbone', 'mgeo')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/mgeo/backbone.py", "imports": ["math", "transformers", "dataclasses", "typing", "os", "random", "torch", "warnings"], "module": "modelscope.models.multi_modal.mgeo.backbone"}, "('MODELS', 'text-classification', 'mgeo')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/mgeo/text_classification.py", "imports": ["torch"], "module": "modelscope.models.multi_modal.mgeo.text_classification"}, "('MODELS', 'nli', 'mgeo')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/mgeo/text_classification.py", "imports": ["torch"], "module": "modelscope.models.multi_modal.mgeo.text_classification"}, "('MODELS', 'sentiment-classification', 'mgeo')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/mgeo/text_classification.py", "imports": ["torch"], "module": "modelscope.models.multi_modal.mgeo.text_classification"}, "('MODELS', 'sentence-similarity', 'mgeo')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/mgeo/text_classification.py", "imports": ["torch"], "module": "modelscope.models.multi_modal.mgeo.text_classification"}, "('MODELS', 'zero-shot-classification', 'mgeo')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/mgeo/text_classification.py", "imports": ["torch"], "module": "modelscope.models.multi_modal.mgeo.text_classification"}, "('MODELS', 'token-classification', 'mgeo')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/mgeo/token_classification.py", "imports": ["torch"], "module": "modelscope.models.multi_modal.mgeo.token_classification"}, "('MODELS', 'part-of-speech', 'mgeo')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/mgeo/token_classification.py", "imports": ["torch"], "module": "modelscope.models.multi_modal.mgeo.token_classification"}, "('MODELS', 'word-segmentation', 'mgeo')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/mgeo/token_classification.py", "imports": ["torch"], "module": "modelscope.models.multi_modal.mgeo.token_classification"}, "('MODELS', 'text-ranking', 'mgeo')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/mgeo/text_ranking.py", "imports": ["torch"], "module": "modelscope.models.multi_modal.mgeo.text_ranking"}, "('MODELS', 'multi-modal-embedding', 'clip-multi-modal-embedding')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/clip/model.py", "imports": ["numpy", "json", "typing", "os", "torch", "collections"], "module": "modelscope.models.multi_modal.clip.model"}, "('MODELS', 'text-to-image-synthesis', 'multi-stage-diffusion-text-to-image-synthesis')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/multi_stage_diffusion/model.py", "imports": ["math", "numpy", "json", "typing", "os", "torch", "PIL"], "module": "modelscope.models.multi_modal.multi_stage_diffusion.model"}, "('MODELS', 'generative-multi-modal-embedding', 'rleg-generative-multi-modal')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/rleg/rleg.py", "imports": ["torch", "torchvision", "typing"], "module": "modelscope.models.multi_modal.rleg.rleg"}, "('METRICS', 'default', 'accuracy')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/metrics/accuracy_metric.py", "imports": ["numpy", "typing"], "module": "modelscope.metrics.accuracy_metric"}, "('METRICS', 'default', 'image-colorization-metric')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/metrics/image_colorization_metric.py", "imports": ["torch", "scipy", "numpy", "typing"], "module": "modelscope.metrics.image_colorization_metric"}, "('METRICS', 'default', 'text-ranking-metric')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/metrics/text_ranking_metric.py", "imports": ["numpy", "typing"], "module": "modelscope.metrics.text_ranking_metric"}, "('METRICS', 'default', 'mAP')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/metrics/map_metric.py", "imports": ["numpy", "typing"], "module": "modelscope.metrics.map_metric"}, "('METRICS', 'default', 'video-super-resolution-metric')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/metrics/video_super_resolution_metric/video_super_resolution_metric.py", "imports": ["numpy", "typing"], "module": "modelscope.metrics.video_super_resolution_metric.video_super_resolution_metric"}, "('METRICS', 'default', 'referring-video-object-segmentation-metric')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/metrics/referring_video_object_segmentation_metric.py", "imports": ["pycocotools", "numpy", "typing", "torch", "tqdm"], "module": "modelscope.metrics.referring_video_object_segmentation_metric"}, "('METRICS', 'default', 'image-quality-assessment-degradation-metric')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/metrics/image_quality_assessment_degradation_metric.py", "imports": ["scipy", "numpy", "cv2", "typing", "os", "tempfile", "torch", "sys", "collections", "tqdm"], "module": "modelscope.metrics.image_quality_assessment_degradation_metric"}, "('METRICS', 'default', 'inbatch_recall')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/metrics/inbatch_recall_metric.py", "imports": ["torch", "numpy", "typing"], "module": "modelscope.metrics.inbatch_recall_metric"}, "('METRICS', 'default', 'movie-scene-segmentation-metric')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/metrics/movie_scene_segmentation_metric.py", "imports": ["numpy", "typing"], "module": "modelscope.metrics.movie_scene_segmentation_metric"}, "('METRICS', 'default', 'image-denoise-metric')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/metrics/image_denoise_metric.py", "imports": ["torch", "cv2", "numpy", "typing"], "module": "modelscope.metrics.image_denoise_metric"}, "('METRICS', 'default', 'image-quality-assessment-mos-metric')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/metrics/image_quality_assessment_mos_metric.py", "imports": ["scipy", "numpy", "typing", "cv2", "os", "tempfile", "torch", "sys", "tqdm"], "module": "modelscope.metrics.image_quality_assessment_mos_metric"}, "('METRICS', 'default', 'image-portrait-enhancement-metric')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/metrics/image_portrait_enhancement_metric.py", "imports": ["cv2", "numpy", "typing"], "module": "modelscope.metrics.image_portrait_enhancement_metric"}, "('METRICS', 'default', 'ocr-recognition-metric')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/metrics/ocr_recognition_metric.py", "imports": ["edit_distance", "numpy", "torch", "typing"], "module": "modelscope.metrics.ocr_recognition_metric"}, "('METRICS', 'default', 'image-inpainting-metric')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/metrics/image_inpainting_metric.py", "imports": ["torch", "scipy", "numpy", "typing"], "module": "modelscope.metrics.image_inpainting_metric"}, "('METRICS', 'default', 'prediction-saving-wrapper')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/metrics/prediction_saving_wrapper.py", "imports": ["numpy", "sklearn", "typing"], "module": "modelscope.metrics.prediction_saving_wrapper"}, "('METRICS', 'default', 'seq-cls-metric')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/metrics/sequence_classification_metric.py", "imports": ["numpy", "sklearn", "typing"], "module": "modelscope.metrics.sequence_classification_metric"}, "('METRICS', 'default', 'ppl')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/metrics/ppl_metric.py", "imports": ["torch", "numpy", "math", "typing"], "module": "modelscope.metrics.ppl_metric"}, "('METRICS', 'default', 'video-frame-interpolation-metric')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/metrics/video_frame_interpolation_metric.py", "imports": ["math", "lpips", "numpy", "typing", "torch"], "module": "modelscope.metrics.video_frame_interpolation_metric"}, "('METRICS', 'default', 'loss-metric')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/metrics/loss_metric.py", "imports": ["numpy", "sklearn", "typing"], "module": "modelscope.metrics.loss_metric"}, "('METRICS', 'default', 'video-stabilization-metric')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/metrics/video_stabilization_metric.py", "imports": ["numpy", "typing", "cv2", "os", "tempfile", "sys", "tqdm"], "module": "modelscope.metrics.video_stabilization_metric"}, "('METRICS', 'default', 'ned')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/metrics/ned_metric.py", "imports": ["numpy", "typing"], "module": "modelscope.metrics.ned_metric"}, "('METRICS', 'default', 'text-gen-metric')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/metrics/text_generation_metric.py", "imports": ["rouge", "nltk", "typing"], "module": "modelscope.metrics.text_generation_metric"}, "('METRICS', 'default', 'bleu')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/metrics/bleu_metric.py", "imports": ["sacrebleu", "typing", "itertools"], "module": "modelscope.metrics.bleu_metric"}, "('METRICS', 'default', 'token-cls-metric')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/metrics/token_classification_metric.py", "imports": ["numpy", "importlib", "typing"], "module": "modelscope.metrics.token_classification_metric"}, "('METRICS', 'default', 'image-ins-seg-coco-metric')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/metrics/image_instance_segmentation_metric.py", "imports": ["pycocotools", "numpy", "typing", "os", "tempfile", "collections"], "module": "modelscope.metrics.image_instance_segmentation_metric"}, "('METRICS', 'default', 'image-color-enhance-metric')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/metrics/image_color_enhance_metric.py", "imports": ["cv2", "numpy", "typing"], "module": "modelscope.metrics.image_color_enhance_metric"}, "('METRICS', 'default', 'audio-noise-metric')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/metrics/audio_noise_metric.py", "imports": ["typing"], "module": "modelscope.metrics.audio_noise_metric"}, "('METRICS', 'default', 'video-summarization-metric')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/metrics/video_summarization_metric.py", "imports": ["numpy", "typing"], "module": "modelscope.metrics.video_summarization_metric"}, "('PIPELINES', 'speech-separation', 'speech-separation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/audio/separation_pipeline.py", "imports": ["soundfile", "io", "numpy", "typing", "torch"], "module": "modelscope.pipelines.audio.separation_pipeline"}, "('PIPELINES', 'text-to-speech', 'sambert-hifigan-tts')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/audio/text_to_speech_pipeline.py", "imports": ["numpy", "typing"], "module": "modelscope.pipelines.audio.text_to_speech_pipeline"}, "('PIPELINES', 'keyword-spotting', 'speech_dfsmn_kws_char_farfield')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/audio/kws_farfield_pipeline.py", "imports": ["soundfile", "io", "numpy", "typing", "wave"], "module": "modelscope.pipelines.audio.kws_farfield_pipeline"}, "('PIPELINES', 'punctuation', 'punc-inference')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/audio/punctuation_processing_pipeline.py", "imports": ["os", "shutil", "yaml", "typing"], "module": "modelscope.pipelines.audio.punctuation_processing_pipeline"}, "('PIPELINES', 'inverse-text-processing', 'itn-inference')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/audio/inverse_text_processing_pipeline.py", "imports": ["os", "shutil", "yaml", "typing"], "module": "modelscope.pipelines.audio.inverse_text_processing_pipeline"}, "('PIPELINES', 'speaker-verification', 'sv-inference')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/audio/speaker_verification_pipeline.py", "imports": ["os", "shutil", "yaml", "typing"], "module": "modelscope.pipelines.audio.speaker_verification_pipeline"}, "('PIPELINES', 'speech-timestamp', 'speech-timestamp-inference')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/audio/timestamp_pipeline.py", "imports": ["json", "yaml", "typing", "os", "funasr"], "module": "modelscope.pipelines.audio.timestamp_pipeline"}, "('PIPELINES', 'acoustic-echo-cancellation', 'speech-dfsmn-aec-psm-16k')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/audio/linear_aec_pipeline.py", "imports": ["scipy", "numpy", "yaml", "typing", "os", "torch", "importlib"], "module": "modelscope.pipelines.audio.linear_aec_pipeline"}, "('PIPELINES', 'speaker-verification', 'speaker-verification')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/audio/speaker_verification_light_pipeline.py", "imports": ["soundfile", "torch", "io", "typing"], "module": "modelscope.pipelines.audio.speaker_verification_light_pipeline"}, "('PIPELINES', 'speaker-diarization', 'speaker-diarization-inference')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/audio/speaker_diarization_pipeline.py", "imports": ["shutil", "numpy", "json", "yaml", "typing", "os"], "module": "modelscope.pipelines.audio.speaker_diarization_pipeline"}, "('PIPELINES', 'acoustic-noise-suppression', 'speech_dfsmn_ans_psm_48k_causal')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/audio/ans_dfsmn_pipeline.py", "imports": ["librosa", "soundfile", "io", "numpy", "typing", "os", "torch", "sys", "collections"], "module": "modelscope.pipelines.audio.ans_dfsmn_pipeline"}, "('PIPELINES', 'voice-activity-detection', 'vad-inference')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/audio/voice_activity_detection_pipeline.py", "imports": ["json", "yaml", "typing", "os", "funasr"], "module": "modelscope.pipelines.audio.voice_activity_detection_pipeline"}, "('PIPELINES', 'acoustic-noise-suppression', 'speech_frcrn_ans_cirm_16k')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/audio/ans_pipeline.py", "imports": ["librosa", "soundfile", "io", "numpy", "typing", "torch"], "module": "modelscope.pipelines.audio.ans_pipeline"}, "('PIPELINES', 'auto-speech-recognition', 'asr-inference')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/audio/asr_inference_pipeline.py", "imports": ["os", "json", "yaml", "typing"], "module": "modelscope.pipelines.audio.asr_inference_pipeline"}, "('PIPELINES', 'auto-speech-recognition', 'asr-wenet-inference')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/audio/asr_wenet_inference_pipeline.py", "imports": ["typing"], "module": "modelscope.pipelines.audio.asr_wenet_inference_pipeline"}, "('PIPELINES', 'keyword-spotting', 'kws-kwsbp')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/audio/kws_kwsbp_pipeline.py", "imports": ["os", "json", "typing"], "module": "modelscope.pipelines.audio.kws_kwsbp_pipeline"}, "('PIPELINES', 'language-score-prediction', 'language-score-prediction')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/audio/lm_infer_pipeline.py", "imports": ["os", "typing"], "module": "modelscope.pipelines.audio.lm_infer_pipeline"}, "('PIPELINES', 'protein-structure', 'unifold-protein-structure')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/science/protein_structure_pipeline.py", "imports": ["time", "numpy", "json", "typing", "unicore", "os", "torch"], "module": "modelscope.pipelines.science.protein_structure_pipeline"}, "('PIPELINES', 'text-generation', 'text-generation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/text_generation_pipeline.py", "imports": ["os", "torch", "typing"], "module": "modelscope.pipelines.nlp.text_generation_pipeline"}, "('PIPELINES', 'text2text-generation', 'translation_en_to_de')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/text_generation_pipeline.py", "imports": ["os", "torch", "typing"], "module": "modelscope.pipelines.nlp.text_generation_pipeline"}, "('PIPELINES', 'text2text-generation', 'translation_en_to_ro')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/text_generation_pipeline.py", "imports": ["os", "torch", "typing"], "module": "modelscope.pipelines.nlp.text_generation_pipeline"}, "('PIPELINES', 'text2text-generation', 'translation_en_to_fr')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/text_generation_pipeline.py", "imports": ["os", "torch", "typing"], "module": "modelscope.pipelines.nlp.text_generation_pipeline"}, "('PIPELINES', 'text2text-generation', 'text2text-generation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/text_generation_pipeline.py", "imports": ["os", "torch", "typing"], "module": "modelscope.pipelines.nlp.text_generation_pipeline"}, "('PIPELINES', 'text-generation', 'gpt-moe-generation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/distributed_gpt_moe_pipeline.py", "imports": ["torch", "typing"], "module": "modelscope.pipelines.nlp.distributed_gpt_moe_pipeline"}, "('PIPELINES', 'document-segmentation', 'document-segmentation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/document_segmentation_pipeline.py", "imports": ["numpy", "re", "transformers", "typing", "datasets", "torch"], "module": "modelscope.pipelines.nlp.document_segmentation_pipeline"}, "('PIPELINES', 'text-generation', 'plug-generation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/distributed_plug_pipeline.py", "imports": ["torch", "typing"], "module": "modelscope.pipelines.nlp.distributed_plug_pipeline"}, "('PIPELINES', 'document-grounded-dialog-generate', 'document-grounded-dialog-generate')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/document_grounded_dialog_generate_pipeline.py", "imports": ["typing"], "module": "modelscope.pipelines.nlp.document_grounded_dialog_generate_pipeline"}, "('PIPELINES', 'named-entity-recognition', 'named-entity-recognition')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/named_entity_recognition_pipeline.py", "imports": ["typing"], "module": "modelscope.pipelines.nlp.named_entity_recognition_pipeline"}, "('PIPELINES', 'named-entity-recognition', 'named-entity-recognition-thai')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/named_entity_recognition_pipeline.py", "imports": ["typing"], "module": "modelscope.pipelines.nlp.named_entity_recognition_pipeline"}, "('PIPELINES', 'named-entity-recognition', 'named-entity-recognition-viet')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/named_entity_recognition_pipeline.py", "imports": ["typing"], "module": "modelscope.pipelines.nlp.named_entity_recognition_pipeline"}, "('PIPELINES', 'feature-extraction', 'feature-extraction')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/feature_extraction_pipeline.py", "imports": ["os", "torch", "typing"], "module": "modelscope.pipelines.nlp.feature_extraction_pipeline"}, "('PIPELINES', 'extractive-summarization', 'extractive-summarization')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/extractive_summarization_pipeline.py", "imports": ["numpy", "re", "typing", "datasets", "torch"], "module": "modelscope.pipelines.nlp.extractive_summarization_pipeline"}, "('PIPELINES', 'text-classification', 'domain-classification')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/fasttext_text_classification_pipeline.py", "imports": ["numpy", "fasttext", "typing", "sentencepiece", "os"], "module": "modelscope.pipelines.nlp.fasttext_text_classification_pipeline"}, "('PIPELINES', 'token-classification', 'token-classification')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/token_classification_pipeline.py", "imports": ["torch", "numpy", "typing"], "module": "modelscope.pipelines.nlp.token_classification_pipeline"}, "('PIPELINES', 'token-classification', 'part-of-speech')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/token_classification_pipeline.py", "imports": ["torch", "numpy", "typing"], "module": "modelscope.pipelines.nlp.token_classification_pipeline"}, "('PIPELINES', 'token-classification', 'word-segmentation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/token_classification_pipeline.py", "imports": ["torch", "numpy", "typing"], "module": "modelscope.pipelines.nlp.token_classification_pipeline"}, "('PIPELINES', 'token-classification', 'named-entity-recognition')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/token_classification_pipeline.py", "imports": ["torch", "numpy", "typing"], "module": "modelscope.pipelines.nlp.token_classification_pipeline"}, "('PIPELINES', 'part-of-speech', 'part-of-speech')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/token_classification_pipeline.py", "imports": ["torch", "numpy", "typing"], "module": "modelscope.pipelines.nlp.token_classification_pipeline"}, "('PIPELINES', 'sentence-embedding', 'sentence-embedding')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/sentence_embedding_pipeline.py", "imports": ["torch", "numpy", "typing"], "module": "modelscope.pipelines.nlp.sentence_embedding_pipeline"}, "('PIPELINES', 'sentence-similarity', 'translation-quality-estimation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/translation_quality_estimation_pipeline.py", "imports": ["io", "transformers", "typing", "os", "torch"], "module": "modelscope.pipelines.nlp.translation_quality_estimation_pipeline"}, "('PIPELINES', 'task-oriented-conversation', 'dialog-state-tracking')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/dialog_state_tracking_pipeline.py", "imports": ["typing"], "module": "modelscope.pipelines.nlp.dialog_state_tracking_pipeline"}, "('PIPELINES', 'table-question-answering', 'table-question-answering-pipeline')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/table_question_answering_pipeline.py", "imports": ["json", "transformers", "typing", "os", "torch"], "module": "modelscope.pipelines.nlp.table_question_answering_pipeline"}, "('PIPELINES', 'faq-question-answering', 'faq-question-answering')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/faq_question_answering_pipeline.py", "imports": ["typing"], "module": "modelscope.pipelines.nlp.faq_question_answering_pipeline"}, "('PIPELINES', 'siamese-uie', 'siamese-uie')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/siamese_uie_pipeline.py", "imports": ["pathlib", "time", "math", "scipy", "json", "typing", "os", "torch", "copy", "logging", "tqdm"], "module": "modelscope.pipelines.nlp.siamese_uie_pipeline"}, "('PIPELINES', 'text-classification', 'sentiment-analysis')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/text_classification_pipeline.py", "imports": ["torch", "numpy", "typing"], "module": "modelscope.pipelines.nlp.text_classification_pipeline"}, "('PIPELINES', 'nli', 'nli')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/text_classification_pipeline.py", "imports": ["torch", "numpy", "typing"], "module": "modelscope.pipelines.nlp.text_classification_pipeline"}, "('PIPELINES', 'sentence-similarity', 'sentence-similarity')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/text_classification_pipeline.py", "imports": ["torch", "numpy", "typing"], "module": "modelscope.pipelines.nlp.text_classification_pipeline"}, "('PIPELINES', 'text-classification', 'text-classification')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/text_classification_pipeline.py", "imports": ["torch", "numpy", "typing"], "module": "modelscope.pipelines.nlp.text_classification_pipeline"}, "('PIPELINES', 'text-classification', 'sentiment-classification')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/text_classification_pipeline.py", "imports": ["torch", "numpy", "typing"], "module": "modelscope.pipelines.nlp.text_classification_pipeline"}, "('PIPELINES', 'text-classification', 'sentence-similarity')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/text_classification_pipeline.py", "imports": ["torch", "numpy", "typing"], "module": "modelscope.pipelines.nlp.text_classification_pipeline"}, "('PIPELINES', 'sentiment-classification', 'sentiment-classification')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/text_classification_pipeline.py", "imports": ["torch", "numpy", "typing"], "module": "modelscope.pipelines.nlp.text_classification_pipeline"}, "('PIPELINES', 'fill-mask', 'fill-mask')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/fill_mask_pipeline.py", "imports": ["numpy", "typing"], "module": "modelscope.pipelines.nlp.fill_mask_pipeline"}, "('PIPELINES', 'fill-mask', 'fill-mask-ponet')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/fill_mask_pipeline.py", "imports": ["numpy", "typing"], "module": "modelscope.pipelines.nlp.fill_mask_pipeline"}, "('PIPELINES', 'zero-shot-classification', 'zero-shot-classification')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/zero_shot_classification_pipeline.py", "imports": ["torch", "scipy", "typing"], "module": "modelscope.pipelines.nlp.zero_shot_classification_pipeline"}, "('PIPELINES', 'task-oriented-conversation', 'dialog-intent-prediction')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/dialog_intent_prediction_pipeline.py", "imports": ["typing"], "module": "modelscope.pipelines.nlp.dialog_intent_prediction_pipeline"}, "('PIPELINES', 'word-alignment', 'word-alignment')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/word_alignment_pipeline.py", "imports": ["numpy", "typing"], "module": "modelscope.pipelines.nlp.word_alignment_pipeline"}, "('PIPELINES', 'text-summarization', 'mglm-text-summarization')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/mglm_text_summarization_pipeline.py", "imports": ["os", "typing"], "module": "modelscope.pipelines.nlp.mglm_text_summarization_pipeline"}, "('PIPELINES', 'text-error-correction', 'text-error-correction')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/text_error_correction_pipeline.py", "imports": ["torch", "typing"], "module": "modelscope.pipelines.nlp.text_error_correction_pipeline"}, "('PIPELINES', 'word-segmentation', 'word-segmentation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/word_segmentation_pipeline.py", "imports": ["torch", "typing"], "module": "modelscope.pipelines.nlp.word_segmentation_pipeline"}, "('PIPELINES', 'word-segmentation', 'multilingual-word-segmentation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/word_segmentation_pipeline.py", "imports": ["torch", "typing"], "module": "modelscope.pipelines.nlp.word_segmentation_pipeline"}, "('PIPELINES', 'word-segmentation', 'word-segmentation-thai')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/word_segmentation_pipeline.py", "imports": ["torch", "typing"], "module": "modelscope.pipelines.nlp.word_segmentation_pipeline"}, "('PIPELINES', 'text-classification', 'language_identification')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/language_identification_pipline.py", "imports": ["numpy", "re", "typing", "os", "tensorflow"], "module": "modelscope.pipelines.nlp.language_identification_pipline"}, "('PIPELINES', 'document-grounded-dialog-retrieval', 'document-grounded-dialog-retrieval')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/document_grounded_dialog_retrieval_pipeline.py", "imports": ["numpy", "json", "typing", "os", "faiss"], "module": "modelscope.pipelines.nlp.document_grounded_dialog_retrieval_pipeline"}, "('PIPELINES', 'translation', 'csanmt-translation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/translation_pipeline.py", "imports": ["jieba", "subword_nmt", "numpy", "sacremoses", "typing", "os", "tensorflow"], "module": "modelscope.pipelines.nlp.translation_pipeline"}, "('PIPELINES', 'task-oriented-conversation', 'dialog-modeling')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/dialog_modeling_pipeline.py", "imports": ["typing"], "module": "modelscope.pipelines.nlp.dialog_modeling_pipeline"}, "('PIPELINES', 'translation', 'interactive-translation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/interactive_translation_pipeline.py", "imports": ["jieba", "subword_nmt", "numpy", "sacremoses", "typing", "os", "tensorflow"], "module": "modelscope.pipelines.nlp.interactive_translation_pipeline"}, "('PIPELINES', 'fid-dialogue', 'fid-dialogue')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/fid_dialogue_pipeline.py", "imports": ["torch", "re", "typing"], "module": "modelscope.pipelines.nlp.fid_dialogue_pipeline"}, "('PIPELINES', 'text-classification', 'user-satisfaction-estimation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/user_satisfaction_estimation_pipeline.py", "imports": ["torch", "numpy", "typing"], "module": "modelscope.pipelines.nlp.user_satisfaction_estimation_pipeline"}, "('PIPELINES', 'text-ranking', 'text-ranking')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/text_ranking_pipeline.py", "imports": ["numpy", "typing"], "module": "modelscope.pipelines.nlp.text_ranking_pipeline"}, "('PIPELINES', 'code-translation', 'codegeex-code-translation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/codegeex_code_translation_pipeline.py", "imports": ["typing"], "module": "modelscope.pipelines.nlp.codegeex_code_translation_pipeline"}, "('PIPELINES', 'text-summarization', 'text-generation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/summarization_pipeline.py", "imports": ["torch", "typing"], "module": "modelscope.pipelines.nlp.summarization_pipeline"}, "('PIPELINES', 'translation-evaluation', 'translation-evaluation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/translation_evaluation_pipeline.py", "imports": ["numpy", "enum", "typing", "os", "torch"], "module": "modelscope.pipelines.nlp.translation_evaluation_pipeline"}, "('PIPELINES', 'text-generation', 'gpt3-generation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/distributed_gpt3_pipeline.py", "imports": ["torch", "typing"], "module": "modelscope.pipelines.nlp.distributed_gpt3_pipeline"}, "('PIPELINES', 'document-grounded-dialog-rerank', 'document-grounded-dialog-rerank')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/document_grounded_dialog_rerank_pipeline.py", "imports": ["time", "numpy", "re", "transformers", "typing", "pprint", "os", "random", "torch", "sys", "collections", "ujson"], "module": "modelscope.pipelines.nlp.document_grounded_dialog_rerank_pipeline"}, "('PIPELINES', 'table-question-answering', 'conversational-text-to-sql')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/conversational_text_to_sql_pipeline.py", "imports": ["torch", "text2sql_lgesql", "typing"], "module": "modelscope.pipelines.nlp.conversational_text_to_sql_pipeline"}, "('PIPELINES', 'information-extraction', 'relation-extraction')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/information_extraction_pipeline.py", "imports": ["torch", "typing"], "module": "modelscope.pipelines.nlp.information_extraction_pipeline"}, "('PIPELINES', 'relation-extraction', 'relation-extraction')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/information_extraction_pipeline.py", "imports": ["torch", "typing"], "module": "modelscope.pipelines.nlp.information_extraction_pipeline"}, "('PIPELINES', 'translation', 'automatic-post-editing')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/automatic_post_editing_pipeline.py", "imports": ["jieba", "html", "numpy", "sacremoses", "typing", "sentencepiece", "os", "tensorflow"], "module": "modelscope.pipelines.nlp.automatic_post_editing_pipeline"}, "('PIPELINES', 'code-generation', 'codegeex-code-generation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/codegeex_code_generation_pipeline.py", "imports": ["typing"], "module": "modelscope.pipelines.nlp.codegeex_code_generation_pipeline"}, "('PIPELINES', 'body-3d-keypoints', 'canonical_body-3d-keypoints_video')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/body_3d_keypoints_pipeline.py", "imports": ["datetime", "matplotlib", "mpl_toolkits", "numpy", "cv2", "typing", "os", "tempfile", "torch"], "module": "modelscope.pipelines.cv.body_3d_keypoints_pipeline"}, "('PIPELINES', 'image-color-enhancement', 'adaint-image-color-enhance')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_color_enhance_pipeline.py", "imports": ["torch", "torchvision", "typing"], "module": "modelscope.pipelines.cv.image_color_enhance_pipeline"}, "('PIPELINES', 'image-color-enhancement', 'deeplpf-image-color-enhance')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_color_enhance_pipeline.py", "imports": ["torch", "torchvision", "typing"], "module": "modelscope.pipelines.cv.image_color_enhance_pipeline"}, "('PIPELINES', 'image-color-enhancement', 'csrnet-image-color-enhance')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_color_enhance_pipeline.py", "imports": ["torch", "torchvision", "typing"], "module": "modelscope.pipelines.cv.image_color_enhance_pipeline"}, "('PIPELINES', 'image-paintbyexample', 'stablediffusion-paintbyexample')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_paintbyexample_pipeline.py", "imports": ["einops", "numpy", "torchvision", "cv2", "typing", "torch", "PIL"], "module": "modelscope.pipelines.cv.image_paintbyexample_pipeline"}, "('PIPELINES', 'animal-recognition', 'resnet101-animal-recognition')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/animal_recognition_pipeline.py", "imports": ["numpy", "torchvision", "typing", "cv2", "os", "torch", "PIL"], "module": "modelscope.pipelines.cv.animal_recognition_pipeline"}, "('PIPELINES', 'image-depth-estimation', 'image-bts-depth-estimation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_bts_depth_estimation_pipeline.py", "imports": ["albumentations", "numpy", "cv2", "typing", "torch"], "module": "modelscope.pipelines.cv.image_bts_depth_estimation_pipeline"}, "('PIPELINES', 'image-segmentation', 'image-panoptic-segmentation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_panoptic_segmentation_pipeline.py", "imports": ["numpy", "cv2", "typing", "torch", "PIL"], "module": "modelscope.pipelines.cv.image_panoptic_segmentation_pipeline"}, "('PIPELINES', 'image-segmentation', 'image-panoptic-segmentation-easycv')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_panoptic_segmentation_pipeline.py", "imports": ["numpy", "cv2", "typing", "torch", "PIL"], "module": "modelscope.pipelines.cv.image_panoptic_segmentation_pipeline"}, "('PIPELINES', 'hand-2d-keypoints', 'hrnetv2w18_hand-2d-keypoints_image')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/hand_2d_keypoints_pipeline.py", "imports": ["os"], "module": "modelscope.pipelines.cv.hand_2d_keypoints_pipeline"}, "('PIPELINES', 'image-skychange', 'image-skychange')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_skychange_pipeline.py", "imports": ["time", "pdb", "numpy", "typing", "cv2", "PIL"], "module": "modelscope.pipelines.cv.image_skychange_pipeline"}, "('PIPELINES', 'video-inpainting', 'video-inpainting')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/video_inpainting_pipeline.py", "imports": ["typing"], "module": "modelscope.pipelines.cv.video_inpainting_pipeline"}, "('PIPELINES', 'image-inpainting', 'image-inpainting-sdv2')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_inpainting_sdv2_pipeline.py", "imports": ["math", "diffusers", "numpy", "cv2", "typing", "os", "tempfile", "torch", "sys"], "module": "modelscope.pipelines.cv.image_inpainting_sdv2_pipeline"}, "('PIPELINES', 'action-recognition', 'TAdaConv_action-recognition')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/action_recognition_pipeline.py", "imports": ["os", "torch", "math", "typing"], "module": "modelscope.pipelines.cv.action_recognition_pipeline"}, "('PIPELINES', 'action-recognition', 'patchshift-action-recognition')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/action_recognition_pipeline.py", "imports": ["os", "torch", "math", "typing"], "module": "modelscope.pipelines.cv.action_recognition_pipeline"}, "('PIPELINES', 'video-text-retrieval', 'vop-video-text-retrieval')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/vop_retrieval_pipeline.py", "imports": ["pickle", "math", "numpy", "gzip", "typing", "os", "random", "torch", "collections", "tqdm"], "module": "modelscope.pipelines.cv.vop_retrieval_pipeline"}, "('PIPELINES', 'image-denoising', 'nafnet-image-denoise')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_denoise_pipeline.py", "imports": ["torch", "torchvision", "typing"], "module": "modelscope.pipelines.cv.image_denoise_pipeline"}, "('PIPELINES', 'image-segmentation', 'vision-middleware-multi-task')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/vision_middleware_pipeline.py", "imports": ["math", "numpy", "torchvision", "typing", "os", "torch", "mmcv"], "module": "modelscope.pipelines.cv.vision_middleware_pipeline"}, "('PIPELINES', 'text-driven-segmentation', 'text-driven-segmentation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/text_driven_segmentation_pipleline.py", "imports": ["typing"], "module": "modelscope.pipelines.cv.text_driven_segmentation_pipleline"}, "('PIPELINES', 'video-single-object-tracking', 'procontext-vitb-video-single-object-tracking')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/video_single_object_tracking_pipeline.py", "imports": ["os", "cv2", "typing"], "module": "modelscope.pipelines.cv.video_single_object_tracking_pipeline"}, "('PIPELINES', 'video-single-object-tracking', 'ostrack-vitb-video-single-object-tracking')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/video_single_object_tracking_pipeline.py", "imports": ["os", "cv2", "typing"], "module": "modelscope.pipelines.cv.video_single_object_tracking_pipeline"}, "('PIPELINES', 'image-classification', 'image-structured-model-probing')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_structured_model_probing_pipeline.py", "imports": ["math", "numpy", "torchvision", "typing", "os", "torch", "mmcv"], "module": "modelscope.pipelines.cv.image_structured_model_probing_pipeline"}, "('PIPELINES', 'image-face-fusion', 'image-face-fusion')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_face_fusion_pipeline.py", "imports": ["numpy", "typing"], "module": "modelscope.pipelines.cv.image_face_fusion_pipeline"}, "('PIPELINES', 'controllable-image-generation', 'controllable-image-generation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/controllable_image_generation_pipeline.py", "imports": ["math", "subprocess", "numpy", "cv2", "typing", "os", "tempfile", "torch", "glob"], "module": "modelscope.pipelines.cv.controllable_image_generation_pipeline"}, "('PIPELINES', 'table-recognition', 'dla34-table-recognition')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/table_recognition_pipeline.py", "imports": ["math", "numpy", "cv2", "typing", "os", "torch", "PIL"], "module": "modelscope.pipelines.cv.table_recognition_pipeline"}, "('PIPELINES', 'language-guided-video-summarization', 'clip-it-video-summarization')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/language_guided_video_summarization_pipeline.py", "imports": ["shutil", "numpy", "cv2", "typing", "os", "random", "tempfile", "torch", "clip", "PIL"], "module": "modelscope.pipelines.cv.language_guided_video_summarization_pipeline"}, "('PIPELINES', 'image-classification', 'resnet50-image-classification-cc')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/content_check_pipeline.py", "imports": ["numpy", "torchvision", "cv2", "typing", "os", "torch", "PIL"], "module": "modelscope.pipelines.cv.content_check_pipeline"}, "('PIPELINES', 'image-quality-assessment-mos', 'image-quality-assessment-man')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_quality_assessment_man_pipeline.py", "imports": ["math", "numpy", "torchvision", "cv2", "typing", "tempfile", "torch"], "module": "modelscope.pipelines.cv.image_quality_assessment_man_pipeline"}, "('PIPELINES', 'portrait-matting', 'unet-image-matting')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_matting_pipeline.py", "imports": ["numpy", "cv2", "typing", "os", "tensorflow"], "module": "modelscope.pipelines.cv.image_matting_pipeline"}, "('PIPELINES', 'universal-matting', 'unet-universal-matting')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_matting_pipeline.py", "imports": ["numpy", "cv2", "typing", "os", "tensorflow"], "module": "modelscope.pipelines.cv.image_matting_pipeline"}, "('PIPELINES', 'lineless-table-recognition', 'lore-lineless-table-recognition')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/lineless_table_recognition_pipeline.py", "imports": ["math", "numpy", "cv2", "typing", "os", "torch", "PIL"], "module": "modelscope.pipelines.cv.lineless_table_recognition_pipeline"}, "('PIPELINES', 'body-2d-keypoints', 'hrnetv2w32_body-2d-keypoints_image')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/body_2d_keypoints_pipeline.py", "imports": ["numpy", "json", "torchvision", "cv2", "typing", "os", "torch", "PIL"], "module": "modelscope.pipelines.cv.body_2d_keypoints_pipeline"}, "('PIPELINES', 'image-depth-estimation', 'image-depth-estimation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_depth_estimation_pipeline.py", "imports": ["numpy", "cv2", "typing", "torch", "PIL"], "module": "modelscope.pipelines.cv.image_depth_estimation_pipeline"}, "('PIPELINES', 'image-portrait-stylization', 'unet-person-image-cartoon')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_cartoon_pipeline.py", "imports": ["numpy", "cv2", "typing", "os", "tensorflow"], "module": "modelscope.pipelines.cv.image_cartoon_pipeline"}, "('PIPELINES', 'ocr-recognition', 'convnextTiny-ocr-recognition')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/ocr_recognition_pipeline.py", "imports": [], "module": "modelscope.pipelines.cv.ocr_recognition_pipeline"}, "('PIPELINES', 'image-to-image-translation', 'image-to-image-translation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_to_image_translation_pipeline.py", "imports": ["io", "numpy", "torchvision", "cv2", "typing", "os", "torch", "sys", "PIL"], "module": "modelscope.pipelines.cv.image_to_image_translation_pipeline"}, "('PIPELINES', 'open-vocabulary-detection', 'open-vocabulary-detection-vild')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_open_vocabulary_detection_pipeline.py", "imports": ["numpy", "typing", "cv2", "os", "torch", "PIL"], "module": "modelscope.pipelines.cv.image_open_vocabulary_detection_pipeline"}, "('PIPELINES', 'face-2d-keypoints', 'manual-facial-landmark-confidence-flcm')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/facial_landmark_confidence_pipeline.py", "imports": ["numpy", "typing", "cv2", "os", "torch", "PIL"], "module": "modelscope.pipelines.cv.facial_landmark_confidence_pipeline"}, "('PIPELINES', 'image-super-resolution', 'mobile-image-super-resolution')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/mobile_image_super_resolution_pipeline.py", "imports": ["numpy", "torchvision", "typing", "torch", "skimage"], "module": "modelscope.pipelines.cv.mobile_image_super_resolution_pipeline"}, "('PIPELINES', 'image-segmentation', 'image-semantic-segmentation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_semantic_segmentation_pipeline.py", "imports": ["numpy", "cv2", "typing", "torch", "PIL"], "module": "modelscope.pipelines.cv.image_semantic_segmentation_pipeline"}, "('PIPELINES', 'image-matching', 'image-matching')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_matching_pipeline.py", "imports": ["numpy", "cv2", "typing", "torch", "PIL"], "module": "modelscope.pipelines.cv.image_matching_pipeline"}, "('PIPELINES', 'face-reconstruction', 'resnet50-face-reconstruction')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/face_reconstruction_pipeline.py", "imports": ["shutil", "scipy", "numpy", "cv2", "typing", "os", "torch", "tensorflow", "face_alignment", "PIL"], "module": "modelscope.pipelines.cv.face_reconstruction_pipeline"}, "('PIPELINES', 'video-frame-interpolation', 'video-frame-interpolation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/video_frame_interpolation_pipeline.py", "imports": ["math", "subprocess", "numpy", "torchvision", "cv2", "typing", "os", "tempfile", "torch", "glob"], "module": "modelscope.pipelines.cv.video_frame_interpolation_pipeline"}, "('PIPELINES', 'image-object-detection', 'tbs-detection')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/tbs_detection_pipeline.py", "imports": ["colorsys", "numpy", "cv2", "typing", "os", "torch", "PIL"], "module": "modelscope.pipelines.cv.tbs_detection_pipeline"}, "('PIPELINES', 'indoor-layout-estimation', 'indoor-layout-estimation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/indoor_layout_estimation_pipeline.py", "imports": ["cv2", "numpy", "typing"], "module": "modelscope.pipelines.cv.indoor_layout_estimation_pipeline"}, "('PIPELINES', 'video-human-matting', 'video-human-matting')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/video_human_matting_pipeline.py", "imports": ["moviepy", "numpy", "cv2", "typing", "os", "torch"], "module": "modelscope.pipelines.cv.video_human_matting_pipeline"}, "('PIPELINES', 'skin-retouching', 'unet-skin-retouching')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/skin_retouching_pipeline.py", "imports": ["numpy", "torchvision", "cv2", "typing", "os", "torch", "tensorflow", "PIL"], "module": "modelscope.pipelines.cv.skin_retouching_pipeline"}, "('PIPELINES', 'video-deinterlace', 'video-deinterlace')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/video_deinterlace_pipeline.py", "imports": ["math", "subprocess", "numpy", "torchvision", "cv2", "typing", "os", "tempfile", "torch"], "module": "modelscope.pipelines.cv.video_deinterlace_pipeline"}, "('PIPELINES', 'image-to-image-generation', 'image-to-image-generation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_to_image_generate_pipeline.py", "imports": ["numpy", "torchvision", "cv2", "typing", "os", "torch", "PIL"], "module": "modelscope.pipelines.cv.image_to_image_generate_pipeline"}, "('PIPELINES', 'face-attribute-recognition', 'resnet34-face-attribute-recognition-fairface')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/face_attribute_recognition_pipeline.py", "imports": ["numpy", "typing", "cv2", "os", "torch", "PIL"], "module": "modelscope.pipelines.cv.face_attribute_recognition_pipeline"}, "('PIPELINES', 'image-segmentation', 'm2fp-image-human-parsing')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_human_parsing_pipeline.py", "imports": ["torch", "numpy", "torchvision", "typing"], "module": "modelscope.pipelines.cv.image_human_parsing_pipeline"}, "('PIPELINES', 'video-depth-estimation', 'video-depth-estimation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/video_depth_estimation_pipeline.py", "imports": ["typing"], "module": "modelscope.pipelines.cv.video_depth_estimation_pipeline"}, "('PIPELINES', 'image-body-reshaping', 'flow-based-body-reshaping')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_body_reshaping_pipeline.py", "imports": ["typing"], "module": "modelscope.pipelines.cv.image_body_reshaping_pipeline"}, "('PIPELINES', 'face-detection', 'resnet101-face-detection-cvpr22papermogface')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/mog_face_detection_pipeline.py", "imports": ["os", "numpy", "typing"], "module": "modelscope.pipelines.cv.mog_face_detection_pipeline"}, "('PIPELINES', 'face-recognition', 'resnet-face-recognition-facemask')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/mask_face_recognition_pipeline.py", "imports": ["numpy", "typing", "cv2", "os", "torch", "collections", "PIL"], "module": "modelscope.pipelines.cv.mask_face_recognition_pipeline"}, "('PIPELINES', 'product-segmentation', 'product-segmentation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/product_segmentation_pipeline.py", "imports": ["numpy", "typing"], "module": "modelscope.pipelines.cv.product_segmentation_pipeline"}, "('PIPELINES', 'semantic-segmentation', 'ddpm-image-semantic-segmentation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/ddpm_semantic_segmentation_pipeline.py", "imports": ["torch", "torchvision", "typing"], "module": "modelscope.pipelines.cv.ddpm_semantic_segmentation_pipeline"}, "('PIPELINES', 'video-embedding', 'cmdssl-r2p1d_video_embedding')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/cmdssl_video_embedding_pipeline.py", "imports": ["numpy", "torchvision", "typing", "os", "decord", "torch", "PIL"], "module": "modelscope.pipelines.cv.cmdssl_video_embedding_pipeline"}, "('PIPELINES', 'image-classification', 'image-classification')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_classification_pipeline.py", "imports": ["torch", "numpy", "typing"], "module": "modelscope.pipelines.cv.image_classification_pipeline"}, "('PIPELINES', 'image-classification', 'vit-base_image-classification_ImageNet-labels')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_classification_pipeline.py", "imports": ["torch", "numpy", "typing"], "module": "modelscope.pipelines.cv.image_classification_pipeline"}, "('PIPELINES', 'image-classification', 'vit-base_image-classification_Dailylife-labels')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_classification_pipeline.py", "imports": ["torch", "numpy", "typing"], "module": "modelscope.pipelines.cv.image_classification_pipeline"}, "('PIPELINES', 'image-classification', 'nextvit-small_image-classification_Dailylife-labels')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_classification_pipeline.py", "imports": ["torch", "numpy", "typing"], "module": "modelscope.pipelines.cv.image_classification_pipeline"}, "('PIPELINES', 'image-classification', 'convnext-base_image-classification_garbage')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_classification_pipeline.py", "imports": ["torch", "numpy", "typing"], "module": "modelscope.pipelines.cv.image_classification_pipeline"}, "('PIPELINES', 'image-classification', 'common-image-classification')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_classification_pipeline.py", "imports": ["torch", "numpy", "typing"], "module": "modelscope.pipelines.cv.image_classification_pipeline"}, "('PIPELINES', 'image-classification', 'easyrobust-classification')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_classification_pipeline.py", "imports": ["torch", "numpy", "typing"], "module": "modelscope.pipelines.cv.image_classification_pipeline"}, "('PIPELINES', 'image-classification', 'bnext-small_image-classification_ImageNet-labels')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_classification_pipeline.py", "imports": ["torch", "numpy", "typing"], "module": "modelscope.pipelines.cv.image_classification_pipeline"}, "('PIPELINES', 'referring-video-object-segmentation', 'referring-video-object-segmentation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/referring_video_object_segmentation_pipeline.py", "imports": ["moviepy", "einops", "numpy", "torchvision", "typing", "tqdm", "tempfile", "torch", "PIL"], "module": "modelscope.pipelines.cv.referring_video_object_segmentation_pipeline"}, "('PIPELINES', 'motion-generation', 'mdm-motion-generation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/motion_generation_pipeline.py", "imports": ["numpy", "typing", "os", "tempfile", "torch"], "module": "modelscope.pipelines.cv.motion_generation_pipeline"}, "('PIPELINES', 'image-inpainting', 'fft-inpainting')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_inpainting_pipeline.py", "imports": ["numpy", "cv2", "typing", "torch", "PIL"], "module": "modelscope.pipelines.cv.image_inpainting_pipeline"}, "('PIPELINES', 'image-quality-assessment-degradation', 'image-quality-assessment-degradation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_quality_assessment_degradation_pipeline.py", "imports": ["math", "numpy", "torchvision", "cv2", "typing", "tempfile", "torch"], "module": "modelscope.pipelines.cv.image_quality_assessment_degradation_pipeline"}, "('PIPELINES', 'face-recognition', 'ir50-face-recognition-arcface')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/arc_face_recognition_pipeline.py", "imports": ["numpy", "typing", "cv2", "os", "torch", "PIL"], "module": "modelscope.pipelines.cv.arc_face_recognition_pipeline"}, "('PIPELINES', 'panorama-depth-estimation', 'panorama-depth-estimation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/panorama_depth_estimation_pipeline.py", "imports": ["numpy", "cv2", "typing", "torch", "PIL"], "module": "modelscope.pipelines.cv.panorama_depth_estimation_pipeline"}, "('PIPELINES', 'video-colorization', 'video-colorization')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/video_colorization_pipeline.py", "imports": ["subprocess", "numpy", "torchvision", "cv2", "typing", "os", "tempfile", "torch", "PIL"], "module": "modelscope.pipelines.cv.video_colorization_pipeline"}, "('PIPELINES', 'image-segmentation', 'cascade-mask-rcnn-swin-image-instance-segmentation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_instance_segmentation_pipeline.py", "imports": ["numpy", "cv2", "typing", "os", "torch", "PIL"], "module": "modelscope.pipelines.cv.image_instance_segmentation_pipeline"}, "('PIPELINES', 'movie-scene-segmentation', 'resnet50-bert-movie-scene-segmentation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/movie_scene_segmentation_pipeline.py", "imports": ["torch", "typing"], "module": "modelscope.pipelines.cv.movie_scene_segmentation_pipeline"}, "('PIPELINES', 'action-detection', 'ResNetC3D-action-detection')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/action_detection_pipeline.py", "imports": ["os", "math", "typing"], "module": "modelscope.pipelines.cv.action_detection_pipeline"}, "('PIPELINES', 'product-retrieval-embedding', 'resnet50-product-retrieval-embedding')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/product_retrieval_embedding_pipeline.py", "imports": ["numpy", "torchvision", "typing", "cv2", "os", "torch", "PIL"], "module": "modelscope.pipelines.cv.product_retrieval_embedding_pipeline"}, "('PIPELINES', 'image-demoireing', 'uhdm-image-demoireing')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_restoration_pipeline.py", "imports": ["typing"], "module": "modelscope.pipelines.cv.image_restoration_pipeline"}, "('PIPELINES', 'video-object-detection', 'cspnet_realtime-video-object-detection_streamyolo')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/realtime_video_object_detection_pipeline.py", "imports": ["numpy", "json", "torchvision", "cv2", "typing", "os", "torch", "PIL"], "module": "modelscope.pipelines.cv.realtime_video_object_detection_pipeline"}, "('PIPELINES', 'live-category', 'live-category')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/live_category_pipeline.py", "imports": ["numpy", "torchvision", "typing", "os", "decord", "torch", "PIL"], "module": "modelscope.pipelines.cv.live_category_pipeline"}, "('PIPELINES', 'video-stabilization', 'video-stabilization')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/video_stabilization_pipeline.py", "imports": ["math", "subprocess", "numpy", "cv2", "typing", "os", "tempfile", "torch", "glob"], "module": "modelscope.pipelines.cv.video_stabilization_pipeline"}, "('PIPELINES', 'video-category', 'video-category')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/video_category_pipeline.py", "imports": ["numpy", "json", "torchvision", "typing", "os", "decord", "torch", "PIL"], "module": "modelscope.pipelines.cv.video_category_pipeline"}, "('PIPELINES', 'face-detection', 'manual-face-detection-ulfd')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/ulfd_face_detection_pipeline.py", "imports": ["numpy", "typing", "cv2", "os", "torch", "PIL"], "module": "modelscope.pipelines.cv.ulfd_face_detection_pipeline"}, "('PIPELINES', 'video-multi-object-tracking', 'video-multi-object-tracking')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/video_multi_object_tracking_pipeline.py", "imports": ["os", "torch", "typing"], "module": "modelscope.pipelines.cv.video_multi_object_tracking_pipeline"}, "('PIPELINES', 'face-recognition', 'manual-face-recognition-frir')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/face_recognition_onnx_ir_pipeline.py", "imports": ["numpy", "typing", "cv2", "os", "onnxruntime", "torch", "PIL"], "module": "modelscope.pipelines.cv.face_recognition_onnx_ir_pipeline"}, "('PIPELINES', 'image-multi-view-depth-estimation', 'image-multi-view-depth-estimation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_mvs_depth_estimation_pipeline.py", "imports": ["os", "shutil", "tempfile", "typing"], "module": "modelscope.pipelines.cv.image_mvs_depth_estimation_pipeline"}, "('PIPELINES', 'image-portrait-enhancement', 'gpen-image-portrait-enhancement')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_portrait_enhancement_pipeline.py", "imports": ["math", "scipy", "numpy", "cv2", "typing", "torch", "PIL"], "module": "modelscope.pipelines.cv.image_portrait_enhancement_pipeline"}, "('PIPELINES', 'crowd-counting', 'hrnet-crowd-counting')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/crowd_counting_pipeline.py", "imports": ["math", "numpy", "torchvision", "typing", "torch", "PIL"], "module": "modelscope.pipelines.cv.crowd_counting_pipeline"}, "('PIPELINES', 'face-detection', 'resnet50-face-detection-retinaface')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/retina_face_detection_pipeline.py", "imports": ["numpy", "typing", "cv2", "os", "torch", "PIL"], "module": "modelscope.pipelines.cv.retina_face_detection_pipeline"}, "('PIPELINES', 'human-reconstruction', 'human-reconstruction')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/human_reconstruction_pipeline.py", "imports": ["shutil", "numpy", "typing", "os", "torch", "trimesh"], "module": "modelscope.pipelines.cv.human_reconstruction_pipeline"}, "('PIPELINES', 'general-recognition', 'resnet101-general-recognition')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/general_recognition_pipeline.py", "imports": ["numpy", "torchvision", "typing", "cv2", "os", "torch", "PIL"], "module": "modelscope.pipelines.cv.general_recognition_pipeline"}, "('PIPELINES', 'face-recognition', 'ir101-face-recognition-cfglint')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/face_recognition_pipeline.py", "imports": ["numpy", "typing", "cv2", "os", "torch", "PIL"], "module": "modelscope.pipelines.cv.face_recognition_pipeline"}, "('PIPELINES', 'video-embedding', 'hicossl-s3dg-video_embedding')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/hicossl_video_embedding_pipeline.py", "imports": ["os", "torch", "math", "typing"], "module": "modelscope.pipelines.cv.hicossl_video_embedding_pipeline"}, "('PIPELINES', 'face-detection', 'manual-face-detection-mtcnn')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/mtcnn_face_detection_pipeline.py", "imports": ["os", "torch", "typing"], "module": "modelscope.pipelines.cv.mtcnn_face_detection_pipeline"}, "('PIPELINES', 'domain-specific-object-detection', 'tinynas-detection')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/tinynas_detection_pipeline.py", "imports": ["typing"], "module": "modelscope.pipelines.cv.tinynas_detection_pipeline"}, "('PIPELINES', 'image-object-detection', 'tinynas-detection')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/tinynas_detection_pipeline.py", "imports": ["typing"], "module": "modelscope.pipelines.cv.tinynas_detection_pipeline"}, "('PIPELINES', 'semantic-segmentation', 'u2net-salient-detection')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_salient_detection_pipeline.py", "imports": ["typing"], "module": "modelscope.pipelines.cv.image_salient_detection_pipeline"}, "('PIPELINES', 'semantic-segmentation', 'res2net-salient-detection')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_salient_detection_pipeline.py", "imports": ["typing"], "module": "modelscope.pipelines.cv.image_salient_detection_pipeline"}, "('PIPELINES', 'semantic-segmentation', 'res2net-camouflaged-detection')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_salient_detection_pipeline.py", "imports": ["typing"], "module": "modelscope.pipelines.cv.image_salient_detection_pipeline"}, "('PIPELINES', 'face-human-hand-detection', 'face-human-hand-detection')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/face_human_hand_detection_pipeline.py", "imports": ["numpy", "typing"], "module": "modelscope.pipelines.cv.face_human_hand_detection_pipeline"}, "('PIPELINES', 'image-deblurring', 'nafnet-image-deblur')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_deblur_pipeline.py", "imports": ["torch", "torchvision", "typing"], "module": "modelscope.pipelines.cv.image_deblur_pipeline"}, "('PIPELINES', 'virtual-try-on', 'virtual-try-on')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/virtual_try_on_pipeline.py", "imports": ["numpy", "typing", "cv2", "os", "torch", "PIL"], "module": "modelscope.pipelines.cv.virtual_try_on_pipeline"}, "('PIPELINES', 'video-object-segmentation', 'video-object-segmentation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/video_object_segmentation_pipeline.py", "imports": ["numpy", "torchvision", "typing", "os", "torch", "PIL"], "module": "modelscope.pipelines.cv.video_object_segmentation_pipeline"}, "('PIPELINES', 'face-recognition', 'manual-face-recognition-frfm')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/face_recognition_onnx_fm_pipeline.py", "imports": ["numpy", "typing", "cv2", "os", "onnxruntime", "torch", "PIL"], "module": "modelscope.pipelines.cv.face_recognition_onnx_fm_pipeline"}, "('PIPELINES', 'object-detection-3d', 'object-detection-3d-depe')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/object_detection_3d_pipeline.py", "imports": ["numpy", "cv2", "typing", "os", "tempfile", "torch", "PIL"], "module": "modelscope.pipelines.cv.object_detection_3d_pipeline"}, "('PIPELINES', 'video-instance-segmentation', 'video-instance-segmentation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/video_instance_segmentation_pipeline.py", "imports": ["numpy", "cv2", "typing", "os", "torch", "mmcv", "tqdm"], "module": "modelscope.pipelines.cv.video_instance_segmentation_pipeline"}, "('PIPELINES', 'bad-image-detecting', 'bad-image-detecting')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/bad_image_detecting_pipeline.py", "imports": ["torch", "numpy", "typing"], "module": "modelscope.pipelines.cv.bad_image_detecting_pipeline"}, "('PIPELINES', 'pointcloud-sceneflow-estimation', 'pointcloud-sceneflow-estimation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/pointcloud_sceneflow_estimation_pipeline.py", "imports": ["torch", "numpy", "plyfile", "typing"], "module": "modelscope.pipelines.cv.pointcloud_sceneflow_estimation_pipeline"}, "('PIPELINES', 'image-segmentation', 'maskdino-swin-image-instance-segmentation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/maskdino_instance_segmentation_pipeline.py", "imports": ["torch", "torchvision", "typing"], "module": "modelscope.pipelines.cv.maskdino_instance_segmentation_pipeline"}, "('PIPELINES', 'image-quality-assessment-mos', 'image-quality-assessment-mos')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_quality_assessment_mos_pipeline.py", "imports": ["math", "numpy", "torchvision", "cv2", "typing", "tempfile", "torch"], "module": "modelscope.pipelines.cv.image_quality_assessment_mos_pipeline"}, "('PIPELINES', 'face-liveness', 'manual-face-liveness-flir')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/face_liveness_ir_pipeline.py", "imports": ["numpy", "typing", "cv2", "os", "onnxruntime", "torch", "PIL"], "module": "modelscope.pipelines.cv.face_liveness_ir_pipeline"}, "('PIPELINES', 'face-liveness', 'manual-face-liveness-flxc')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/face_liveness_xc_pipeline.py", "imports": ["numpy", "typing", "cv2", "os", "onnxruntime", "torch", "PIL"], "module": "modelscope.pipelines.cv.face_liveness_xc_pipeline"}, "('PIPELINES', 'image-driving-perception', 'yolopv2_image-driving-percetion_bdd100k')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_driving_perception_pipeline.py", "imports": ["os", "cv2", "numpy", "typing"], "module": "modelscope.pipelines.cv.image_driving_perception_pipeline"}, "('PIPELINES', 'facial-expression-recognition', 'vgg19-facial-expression-recognition-fer')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/facial_expression_recognition_pipeline.py", "imports": ["numpy", "typing", "cv2", "os", "torch", "PIL"], "module": "modelscope.pipelines.cv.facial_expression_recognition_pipeline"}, "('PIPELINES', 'hand-static', 'hand-static')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/hand_static_pipeline.py", "imports": ["numpy", "typing"], "module": "modelscope.pipelines.cv.hand_static_pipeline"}, "('PIPELINES', 'video-panoptic-segmentation', 'video-panoptic-segmentation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/video_panoptic_segmentation_pipeline.py", "imports": ["numpy", "cv2", "typing", "os", "torch", "mmcv", "tqdm"], "module": "modelscope.pipelines.cv.video_panoptic_segmentation_pipeline"}, "('PIPELINES', 'video-super-resolution', 'realbasicvsr-video-super-resolution')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/video_super_resolution_pipeline.py", "imports": ["math", "subprocess", "numpy", "torchvision", "cv2", "typing", "os", "tempfile", "torch"], "module": "modelscope.pipelines.cv.video_super_resolution_pipeline"}, "('PIPELINES', 'license-plate-detection', 'resnet18-license-plate-detection')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/license_plate_detection_pipeline.py", "imports": ["math", "numpy", "cv2", "typing", "os", "torch", "PIL"], "module": "modelscope.pipelines.cv.license_plate_detection_pipeline"}, "('PIPELINES', 'image-colorization', 'ddcolor-image-colorization')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/ddcolor_image_colorization_pipeline.py", "imports": ["numpy", "torchvision", "cv2", "typing", "torch"], "module": "modelscope.pipelines.cv.ddcolor_image_colorization_pipeline"}, "('PIPELINES', 'vision-efficient-tuning', 'vision-efficient-tuning')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/vision_efficient_tuning_pipeline.py", "imports": ["torch", "numpy", "torchvision", "typing"], "module": "modelscope.pipelines.cv.vision_efficient_tuning_pipeline"}, "('PIPELINES', 'face-detection', 'resnet-face-detection-scrfd10gkps')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/face_detection_pipeline.py", "imports": ["numpy", "typing", "cv2", "os", "torch", "PIL"], "module": "modelscope.pipelines.cv.face_detection_pipeline"}, "('PIPELINES', 'image-style-transfer', 'AAMS-style-transfer')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_style_transfer_pipeline.py", "imports": ["os", "cv2", "numpy", "typing"], "module": "modelscope.pipelines.cv.image_style_transfer_pipeline"}, "('PIPELINES', 'video-text-retrieval', 'vop-video-text-retrieval-se')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/vop_retrieval_se_pipeline.py", "imports": ["numpy", "gzip", "typing", "os", "torch"], "module": "modelscope.pipelines.cv.vop_retrieval_se_pipeline"}, "('PIPELINES', 'image-reid-person', 'passvitb-image-reid-person')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_reid_person_pipeline.py", "imports": ["math", "torchvision", "typing", "os", "torch", "PIL"], "module": "modelscope.pipelines.cv.image_reid_person_pipeline"}, "('PIPELINES', 'human-detection', 'resnet18-human-detection')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_detection_pipeline.py", "imports": ["numpy", "typing"], "module": "modelscope.pipelines.cv.image_detection_pipeline"}, "('PIPELINES', 'image-object-detection', 'vit-object-detection')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_detection_pipeline.py", "imports": ["numpy", "typing"], "module": "modelscope.pipelines.cv.image_detection_pipeline"}, "('PIPELINES', 'image-object-detection', 'abnormal-object-detection')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_detection_pipeline.py", "imports": ["numpy", "typing"], "module": "modelscope.pipelines.cv.image_detection_pipeline"}, "('PIPELINES', 'ocr-detection', 'resnet18-ocr-detection')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/ocr_detection_pipeline.py", "imports": ["math", "numpy", "typing", "cv2", "os", "tf_slim", "torch", "tensorflow"], "module": "modelscope.pipelines.cv.ocr_detection_pipeline"}, "('PIPELINES', 'image-classification', 'tinynas-classification')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/tinynas_classification_pipeline.py", "imports": ["math", "torchvision", "typing", "os", "torch"], "module": "modelscope.pipelines.cv.tinynas_classification_pipeline"}, "('PIPELINES', 'image-colorization', 'unet-image-colorization')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_colorization_pipeline.py", "imports": ["numpy", "torchvision", "cv2", "typing", "torch", "PIL"], "module": "modelscope.pipelines.cv.image_colorization_pipeline"}, "('PIPELINES', 'card-detection', 'resnet-card-detection-scrfd34gkps')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/card_detection_pipeline.py", "imports": ["typing"], "module": "modelscope.pipelines.cv.card_detection_pipeline"}, "('PIPELINES', 'video-summarization', 'googlenet_pgl_video_summarization')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/video_summarization_pipeline.py", "imports": ["numpy", "cv2", "typing", "os", "torch", "tqdm"], "module": "modelscope.pipelines.cv.video_summarization_pipeline"}, "('PIPELINES', 'image-fewshot-detection', 'image-fewshot-detection')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_defrcn_fewshot_pipeline.py", "imports": ["os", "numpy", "torch", "typing"], "module": "modelscope.pipelines.cv.image_defrcn_fewshot_pipeline"}, "('PIPELINES', 'face-quality-assessment', 'manual-face-quality-assessment-fqa')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/face_quality_assessment_pipeline.py", "imports": ["numpy", "typing", "cv2", "os", "onnxruntime", "torch", "PIL"], "module": "modelscope.pipelines.cv.face_quality_assessment_pipeline"}, "('PIPELINES', 'image-super-resolution', 'rrdb-image-super-resolution')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_super_resolution_pipeline.py", "imports": ["numpy", "cv2", "typing", "torch", "PIL"], "module": "modelscope.pipelines.cv.image_super_resolution_pipeline"}, "('PIPELINES', 'face-image-generation', 'gan-face-image-generation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/face_image_generation_pipeline.py", "imports": ["numpy", "typing", "cv2", "os", "torch", "PIL"], "module": "modelscope.pipelines.cv.face_image_generation_pipeline"}, "('PIPELINES', 'image-debanding', 'rrdb-image-debanding')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_debanding_pipeline.py", "imports": ["torch", "torchvision", "typing"], "module": "modelscope.pipelines.cv.image_debanding_pipeline"}, "('PIPELINES', 'image-object-detection', 'vidt')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/vidt_pipeline.py", "imports": ["torch", "torchvision", "typing"], "module": "modelscope.pipelines.cv.vidt_pipeline"}, "('PIPELINES', 'image-object-detection', 'easycv-detection')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/easycv_pipelines/detection_pipeline.py", "imports": ["typing"], "module": "modelscope.pipelines.cv.easycv_pipelines.detection_pipeline"}, "('PIPELINES', 'image-object-detection', 'yolox_image-object-detection-auto')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/easycv_pipelines/detection_pipeline.py", "imports": ["typing"], "module": "modelscope.pipelines.cv.easycv_pipelines.detection_pipeline"}, "('PIPELINES', 'domain-specific-object-detection', 'yolox-pai_hand-detection')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/easycv_pipelines/detection_pipeline.py", "imports": ["typing"], "module": "modelscope.pipelines.cv.easycv_pipelines.detection_pipeline"}, "('PIPELINES', 'image-segmentation', 'easycv-segmentation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/easycv_pipelines/segmentation_pipeline.py", "imports": ["numpy", "typing"], "module": "modelscope.pipelines.cv.easycv_pipelines.segmentation_pipeline"}, "('PIPELINES', 'face-2d-keypoints', 'mobilenet_face-2d-keypoints_alignment')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/easycv_pipelines/face_2d_keypoints_pipeline.py", "imports": ["math", "numpy", "typing", "cv2", "copy"], "module": "modelscope.pipelines.cv.easycv_pipelines.face_2d_keypoints_pipeline"}, "('PIPELINES', 'human-wholebody-keypoint', 'hrnetw48_human-wholebody-keypoint_image')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/easycv_pipelines/human_wholebody_keypoint_pipeline.py", "imports": ["os", "typing"], "module": "modelscope.pipelines.cv.easycv_pipelines.human_wholebody_keypoint_pipeline"}, "('PIPELINES', 'face-emotion', 'face-emotion')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/face_emotion_pipeline.py", "imports": ["numpy", "typing"], "module": "modelscope.pipelines.cv.face_emotion_pipeline"}, "('PIPELINES', 'nerf-recon-acc', 'nerf-recon-acc')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/nerf_recon_acc_pipeline.py", "imports": ["typing"], "module": "modelscope.pipelines.cv.nerf_recon_acc_pipeline"}, "('PIPELINES', 'shop-segmentation', 'shop-segmentation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/shop_segmentation_pipleline.py", "imports": ["typing"], "module": "modelscope.pipelines.cv.shop_segmentation_pipleline"}, "('PIPELINES', 'face-recognition', 'ir-face-recognition-rts')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/face_recognition_ood_pipeline.py", "imports": ["numpy", "typing", "cv2", "os", "torch", "PIL"], "module": "modelscope.pipelines.cv.face_recognition_ood_pipeline"}, "('PIPELINES', 'document-vl-embedding', 'document-vl-embedding')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/multi_modal/document_vl_embedding_pipeline.py", "imports": ["torch", "typing"], "module": "modelscope.pipelines.multi_modal.document_vl_embedding_pipeline"}, "('PIPELINES', 'text2sql', 'ofa-text2sql')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/multi_modal/text2sql_pipeline.py", "imports": ["torch", "typing"], "module": "modelscope.pipelines.multi_modal.text2sql_pipeline"}, "('PIPELINES', 'visual-entailment', 'visual-entailment')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/multi_modal/visual_entailment_pipeline.py", "imports": ["torch", "typing"], "module": "modelscope.pipelines.multi_modal.visual_entailment_pipeline"}, "('PIPELINES', 'ocr-recognition', 'ofa-ocr-recognition')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/multi_modal/ocr_recognition_pipeline.py", "imports": ["torch", "typing"], "module": "modelscope.pipelines.multi_modal.ocr_recognition_pipeline"}, "('PIPELINES', 'text-to-video-synthesis', 'latent-text-to-video-synthesis')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/multi_modal/text_to_video_synthesis_pipeline.py", "imports": ["einops", "typing", "cv2", "tempfile", "torch"], "module": "modelscope.pipelines.multi_modal.text_to_video_synthesis_pipeline"}, "('PIPELINES', 'text-to-image-synthesis', 'diffusers-stable-diffusion')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/multi_modal/diffusers_wrapped/stable_diffusion/stable_diffusion_pipeline.py", "imports": ["diffusers", "numpy", "typing", "cv2", "torch", "PIL"], "module": "modelscope.pipelines.multi_modal.diffusers_wrapped.stable_diffusion.stable_diffusion_pipeline"}, "('PIPELINES', 'text-to-image-synthesis', 'chinese-stable-diffusion')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/multi_modal/diffusers_wrapped/stable_diffusion/chinese_stable_diffusion_pipeline.py", "imports": ["diffusers", "numpy", "transformers", "typing", "cv2", "torch", "PIL"], "module": "modelscope.pipelines.multi_modal.diffusers_wrapped.stable_diffusion.chinese_stable_diffusion_pipeline"}, "('PIPELINES', 'sudoku', 'ofa-sudoku')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/multi_modal/sudoku_pipeline.py", "imports": ["torch", "typing"], "module": "modelscope.pipelines.multi_modal.sudoku_pipeline"}, "('PIPELINES', 'video-temporal-grounding', 'soonet-video-temporal-grounding')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/multi_modal/soonet_video_temporal_grounding_pipeline.py", "imports": ["numpy", "torchvision", "typing", "os", "torch"], "module": "modelscope.pipelines.multi_modal.soonet_video_temporal_grounding_pipeline"}, "('PIPELINES', 'video-captioning', 'video-captioning')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/multi_modal/video_captioning_pipeline.py", "imports": ["torch", "typing"], "module": "modelscope.pipelines.multi_modal.video_captioning_pipeline"}, "('PIPELINES', 'multi-modal-similarity', 'multi-modal-similarity')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/multi_modal/team_multi_modal_similarity_pipeline.py", "imports": ["typing"], "module": "modelscope.pipelines.multi_modal.team_multi_modal_similarity_pipeline"}, "('PIPELINES', 'visual-question-answering', 'gridvlp-multi-modal-classification')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/multi_modal/gridvlp_pipeline.py", "imports": ["time", "numpy", "json", "transformers", "typing", "os", "traceback", "torch", "PIL"], "module": "modelscope.pipelines.multi_modal.gridvlp_pipeline"}, "('PIPELINES', 'multi-modal-embedding', 'gridvlp-multi-modal-embedding')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/multi_modal/gridvlp_pipeline.py", "imports": ["time", "numpy", "json", "transformers", "typing", "os", "traceback", "torch", "PIL"], "module": "modelscope.pipelines.multi_modal.gridvlp_pipeline"}, "('PIPELINES', 'text-ranking', 'mgeo-ranking')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/multi_modal/mgeo_ranking_pipeline.py", "imports": ["torch", "numpy", "typing"], "module": "modelscope.pipelines.multi_modal.mgeo_ranking_pipeline"}, "('PIPELINES', 'generative-multi-modal-embedding', 'generative-multi-modal-embedding')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/multi_modal/generative_multi_modal_embedding_pipeline.py", "imports": ["typing"], "module": "modelscope.pipelines.multi_modal.generative_multi_modal_embedding_pipeline"}, "('PIPELINES', 'auto-speech-recognition', 'ofa-asr')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/multi_modal/asr_pipeline.py", "imports": ["torch", "typing"], "module": "modelscope.pipelines.multi_modal.asr_pipeline"}, "('PIPELINES', 'visual-question-answering', 'visual-question-answering')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/multi_modal/visual_question_answering_pipeline.py", "imports": ["torch", "typing"], "module": "modelscope.pipelines.multi_modal.visual_question_answering_pipeline"}, "('PIPELINES', 'text-to-image-synthesis', 'text-to-image-synthesis')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/multi_modal/text_to_image_synthesis_pipeline.py", "imports": ["torch", "typing"], "module": "modelscope.pipelines.multi_modal.text_to_image_synthesis_pipeline"}, "('PIPELINES', 'visual-grounding', 'visual-grounding')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/multi_modal/visual_grounding_pipeline.py", "imports": ["torch", "typing"], "module": "modelscope.pipelines.multi_modal.visual_grounding_pipeline"}, "('PIPELINES', 'image-captioning', 'image-captioning')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/multi_modal/image_captioning_pipeline.py", "imports": ["torch", "typing"], "module": "modelscope.pipelines.multi_modal.image_captioning_pipeline"}, "('PIPELINES', 'image-text-retrieval', 'multi-modal-embedding')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/multi_modal/multi_modal_embedding_pipeline.py", "imports": ["typing"], "module": "modelscope.pipelines.multi_modal.multi_modal_embedding_pipeline"}, "('PIPELINES', 'multi-modal-embedding', 'multi-modal-embedding')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/multi_modal/multi_modal_embedding_pipeline.py", "imports": ["typing"], "module": "modelscope.pipelines.multi_modal.multi_modal_embedding_pipeline"}, "('PIPELINES', 'video-multi-modal-embedding', 'video-multi-modal-embedding')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/multi_modal/video_multi_modal_embedding_pipeline.py", "imports": ["typing"], "module": "modelscope.pipelines.multi_modal.video_multi_modal_embedding_pipeline"}, "('PIPELINES', 'text-to-image-synthesis', 'disco_guided_diffusion')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/multi_modal/disco_guided_diffusion_pipeline/disco_guided_diffusion.py", "imports": ["gc", "math", "numpy", "json", "torchvision", "cv2", "os", "importlib", "torch", "clip", "PIL"], "module": "modelscope.pipelines.multi_modal.disco_guided_diffusion_pipeline.disco_guided_diffusion"}, "('PIPELINES', 'image-text-retrieval', 'image-text-retrieval')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/multi_modal/image_text_retrieval_pipeline.py", "imports": ["torch", "typing"], "module": "modelscope.pipelines.multi_modal.image_text_retrieval_pipeline"}, "('PIPELINES', 'video-question-answering', 'video-question-answering')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/multi_modal/video_question_answering_pipeline.py", "imports": ["torch", "typing"], "module": "modelscope.pipelines.multi_modal.video_question_answering_pipeline"}, "('PREPROCESSORS', 'cv', 'load-image')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/image.py", "imports": ["io", "numpy", "cv2", "typing", "PIL"], "module": "modelscope.preprocessors.image"}, "('PREPROCESSORS', 'cv', 'object-detection-tinynas-preprocessor')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/image.py", "imports": ["io", "numpy", "cv2", "typing", "PIL"], "module": "modelscope.preprocessors.image"}, "('PREPROCESSORS', 'cv', 'image-color-enhance-preprocessor')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/image.py", "imports": ["io", "numpy", "cv2", "typing", "PIL"], "module": "modelscope.preprocessors.image"}, "('PREPROCESSORS', 'cv', 'image-denoise-preprocessor')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/image.py", "imports": ["io", "numpy", "cv2", "typing", "PIL"], "module": "modelscope.preprocessors.image"}, "('PREPROCESSORS', 'cv', 'image-deblur-preprocessor')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/image.py", "imports": ["io", "numpy", "cv2", "typing", "PIL"], "module": "modelscope.preprocessors.image"}, "('PREPROCESSORS', 'cv', 'image-portrait-enhancement-preprocessor')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/image.py", "imports": ["io", "numpy", "cv2", "typing", "PIL"], "module": "modelscope.preprocessors.image"}, "('PREPROCESSORS', 'cv', 'image-instance-segmentation-preprocessor')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/image.py", "imports": ["io", "numpy", "cv2", "typing", "PIL"], "module": "modelscope.preprocessors.image"}, "('PREPROCESSORS', 'cv', 'video-summarization-preprocessor')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/image.py", "imports": ["io", "numpy", "cv2", "typing", "PIL"], "module": "modelscope.preprocessors.image"}, "('PREPROCESSORS', 'cv', 'image-classification-bypass-preprocessor')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/image.py", "imports": ["io", "numpy", "cv2", "typing", "PIL"], "module": "modelscope.preprocessors.image"}, "('PREPROCESSORS', 'audio', 'LinearAECAndFbank')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/audio.py", "imports": ["io", "scipy", "numpy", "typing", "os", "torch"], "module": "modelscope.preprocessors.audio"}, "('PREPROCESSORS', 'audio', 'wav-to-scp')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/asr.py", "imports": ["os", "typing"], "module": "modelscope.preprocessors.asr"}, "('PREPROCESSORS', 'audio', 'wav-to-lists')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/kws.py", "imports": ["os", "yaml", "typing"], "module": "modelscope.preprocessors.kws"}, "('PREPROCESSORS', 'multi-modal', 'ofa-tasks-preprocessor')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/multi_modal.py", "imports": ["io", "numpy", "json", "torchvision", "typing", "os", "decord", "torch", "timm", "PIL"], "module": "modelscope.preprocessors.multi_modal"}, "('PREPROCESSORS', 'multi-modal', 'clip-preprocessor')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/multi_modal.py", "imports": ["io", "numpy", "json", "torchvision", "typing", "os", "decord", "torch", "timm", "PIL"], "module": "modelscope.preprocessors.multi_modal"}, "('PREPROCESSORS', 'multi-modal', 'mplug-tasks-preprocessor')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/multi_modal.py", "imports": ["io", "numpy", "json", "torchvision", "typing", "os", "decord", "torch", "timm", "PIL"], "module": "modelscope.preprocessors.multi_modal"}, "('PREPROCESSORS', 'multi-modal', 'vldoc-preprocessor')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/multi_modal.py", "imports": ["io", "numpy", "json", "torchvision", "typing", "os", "decord", "torch", "timm", "PIL"], "module": "modelscope.preprocessors.multi_modal"}, "('PREPROCESSORS', 'multi-modal', 'hitea-tasks-preprocessor')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/multi_modal.py", "imports": ["io", "numpy", "json", "torchvision", "typing", "os", "decord", "torch", "timm", "PIL"], "module": "modelscope.preprocessors.multi_modal"}, "('PREPROCESSORS', 'cv', 'movie-scene-segmentation-preprocessor')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/video.py", "imports": ["math", "uuid", "numpy", "torchvision", "os", "random", "urllib", "tempfile", "decord", "torch"], "module": "modelscope.preprocessors.video"}, "('PREPROCESSORS', 'science', 'unifold-preprocessor')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/science/uni_fold.py", "imports": ["tarfile", "ipdb", "hashlib", "gzip", "re", "requests", "random", "time", "pathlib", "pickle", "numpy", "json", "unittest", "typing", "os", "torch", "logging", "tqdm"], "module": "modelscope.preprocessors.science.uni_fold"}, "('PREPROCESSORS', 'nlp', 'faq-question-answering-preprocessor')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/faq_question_answering_preprocessor.py", "imports": ["torch", "typing"], "module": "modelscope.preprocessors.nlp.faq_question_answering_preprocessor"}, "('PREPROCESSORS', 'nlp', 'translation-evaluation-preprocessor')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/translation_evaluation_preprocessor.py", "imports": ["transformers", "typing"], "module": "modelscope.preprocessors.nlp.translation_evaluation_preprocessor"}, "('PREPROCESSORS', 'nlp', 'document-grounded-dialog-retrieval')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/document_grounded_dialog_retrieval_preprocessor.py", "imports": ["os", "torch", "transformers", "typing"], "module": "modelscope.preprocessors.nlp.document_grounded_dialog_retrieval_preprocessor"}, "('PREPROCESSORS', 'nlp', 'nli-tokenizer')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/text_classification_preprocessor.py", "imports": ["numpy", "typing"], "module": "modelscope.preprocessors.nlp.text_classification_preprocessor"}, "('PREPROCESSORS', 'nlp', 'sen-sim-tokenizer')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/text_classification_preprocessor.py", "imports": ["numpy", "typing"], "module": "modelscope.preprocessors.nlp.text_classification_preprocessor"}, "('PREPROCESSORS', 'nlp', 'bert-seq-cls-tokenizer')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/text_classification_preprocessor.py", "imports": ["numpy", "typing"], "module": "modelscope.preprocessors.nlp.text_classification_preprocessor"}, "('PREPROCESSORS', 'nlp', 'sen-cls-tokenizer')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/text_classification_preprocessor.py", "imports": ["numpy", "typing"], "module": "modelscope.preprocessors.nlp.text_classification_preprocessor"}, "('PREPROCESSORS', 'nlp', 'text-ranking')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/text_ranking_preprocessor.py", "imports": ["transformers", "typing"], "module": "modelscope.preprocessors.nlp.text_ranking_preprocessor"}, "('PREPROCESSORS', 'nlp', 'word-segment-text-to-label-preprocessor')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/token_classification_preprocessor.py", "imports": ["torch", "numpy", "typing"], "module": "modelscope.preprocessors.nlp.token_classification_preprocessor"}, "('PREPROCESSORS', 'nlp', 'ner-tokenizer')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/token_classification_preprocessor.py", "imports": ["torch", "numpy", "typing"], "module": "modelscope.preprocessors.nlp.token_classification_preprocessor"}, "('PREPROCESSORS', 'nlp', 'token-cls-tokenizer')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/token_classification_preprocessor.py", "imports": ["torch", "numpy", "typing"], "module": "modelscope.preprocessors.nlp.token_classification_preprocessor"}, "('PREPROCESSORS', 'nlp', 'sequence-labeling-tokenizer')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/token_classification_preprocessor.py", "imports": ["torch", "numpy", "typing"], "module": "modelscope.preprocessors.nlp.token_classification_preprocessor"}, "('PREPROCESSORS', 'nlp', 'document-segmentation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/document_segmentation_preprocessor.py", "imports": ["typing"], "module": "modelscope.preprocessors.nlp.document_segmentation_preprocessor"}, "('PREPROCESSORS', 'nlp', 'thai-ner-tokenizer')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/token_classification_thai_preprocessor.py", "imports": ["typing"], "module": "modelscope.preprocessors.nlp.token_classification_thai_preprocessor"}, "('PREPROCESSORS', 'nlp', 'thai-wseg-tokenizer')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/token_classification_thai_preprocessor.py", "imports": ["typing"], "module": "modelscope.preprocessors.nlp.token_classification_thai_preprocessor"}, "('PREPROCESSORS', 'nlp', 'sentence-embedding')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/sentence_embedding_preprocessor.py", "imports": ["typing"], "module": "modelscope.preprocessors.nlp.sentence_embedding_preprocessor"}, "('PREPROCESSORS', 'nlp', 'feature-extraction')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/feature_extraction_preprocessor.py", "imports": ["numpy", "typing"], "module": "modelscope.preprocessors.nlp.feature_extraction_preprocessor"}, "('PREPROCESSORS', 'nlp', 'Tokenize')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/bert_seq_cls_tokenizer.py", "imports": ["transformers", "typing"], "module": "modelscope.preprocessors.nlp.bert_seq_cls_tokenizer"}, "('PREPROCESSORS', 'nlp', 'document-grounded-dialog-generate')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/document_grounded_dialog_generate_preprocessor.py", "imports": ["os", "torch", "transformers", "typing"], "module": "modelscope.preprocessors.nlp.document_grounded_dialog_generate_preprocessor"}, "('PREPROCESSORS', 'nlp', 'mgeo-ranking')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/mgeo_ranking_preprocessor.py", "imports": ["torch", "transformers", "typing"], "module": "modelscope.preprocessors.nlp.mgeo_ranking_preprocessor"}, "('PREPROCESSORS', 'nlp', 'fill-mask')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/fill_mask_preprocessor.py", "imports": ["abc", "numpy", "re", "typing", "os", "torch"], "module": "modelscope.preprocessors.nlp.fill_mask_preprocessor"}, "('PREPROCESSORS', 'nlp', 'fill-mask-ponet')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/fill_mask_preprocessor.py", "imports": ["abc", "numpy", "re", "typing", "os", "torch"], "module": "modelscope.preprocessors.nlp.fill_mask_preprocessor"}, "('PREPROCESSORS', 'nlp', 'conversational-text-to-sql')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/space_T_en/conversational_text_to_sql_preprocessor.py", "imports": ["json", "text2sql_lgesql", "typing", "os", "torch"], "module": "modelscope.preprocessors.nlp.space_T_en.conversational_text_to_sql_preprocessor"}, "('PREPROCESSORS', 'nlp', 'viet-ner-tokenizer')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/token_classification_viet_preprocessor.py", "imports": ["torch", "typing"], "module": "modelscope.preprocessors.nlp.token_classification_viet_preprocessor"}, "('PREPROCESSORS', 'nlp', 'mglm-summarization')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/mglm_summarization_preprocessor.py", "imports": ["os", "re", "typing"], "module": "modelscope.preprocessors.nlp.mglm_summarization_preprocessor"}, "('PREPROCESSORS', 'nlp', 'text-error-correction')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/text_error_correction.py", "imports": ["os", "torch", "transformers", "typing"], "module": "modelscope.preprocessors.nlp.text_error_correction"}, "('PREPROCESSORS', 'nlp', 'zero-shot-cls-tokenizer')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/zero_shot_classification_preprocessor.py", "imports": ["typing"], "module": "modelscope.preprocessors.nlp.zero_shot_classification_preprocessor"}, "('PREPROCESSORS', 'nlp', 'table-question-answering-preprocessor')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/space_T_cn/table_question_answering_preprocessor.py", "imports": ["os", "torch", "transformers", "typing"], "module": "modelscope.preprocessors.nlp.space_T_cn.table_question_answering_preprocessor"}, "('PREPROCESSORS', 'nlp', 'text-gen-tokenizer')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/text_generation_preprocessor.py", "imports": ["os", "numpy", "torch", "typing"], "module": "modelscope.preprocessors.nlp.text_generation_preprocessor"}, "('PREPROCESSORS', 'nlp', 'text-gen-jieba-tokenizer')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/text_generation_preprocessor.py", "imports": ["os", "numpy", "torch", "typing"], "module": "modelscope.preprocessors.nlp.text_generation_preprocessor"}, "('PREPROCESSORS', 'nlp', 'sentence-piece')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/text_generation_preprocessor.py", "imports": ["os", "numpy", "torch", "typing"], "module": "modelscope.preprocessors.nlp.text_generation_preprocessor"}, "('PREPROCESSORS', 'nlp', 'text2text-gen-preprocessor')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/text_generation_preprocessor.py", "imports": ["os", "numpy", "torch", "typing"], "module": "modelscope.preprocessors.nlp.text_generation_preprocessor"}, "('PREPROCESSORS', 'nlp', 'dialog-use-preprocessor')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/dialog_classification_use_preprocessor.py", "imports": ["torch", "transformers", "typing"], "module": "modelscope.preprocessors.nlp.dialog_classification_use_preprocessor"}, "('PREPROCESSORS', 'nlp', 'dialog-intent-preprocessor')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/space/dialog_intent_prediction_preprocessor.py", "imports": ["os", "json", "typing"], "module": "modelscope.preprocessors.nlp.space.dialog_intent_prediction_preprocessor"}, "('PREPROCESSORS', 'nlp', 'dialog-modeling-preprocessor')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/space/dialog_modeling_preprocessor.py", "imports": ["os", "typing"], "module": "modelscope.preprocessors.nlp.space.dialog_modeling_preprocessor"}, "('PREPROCESSORS', 'nlp', 'dialog-state-tracking-preprocessor')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/space/dialog_state_tracking_preprocessor.py", "imports": ["typing"], "module": "modelscope.preprocessors.nlp.space.dialog_state_tracking_preprocessor"}, "('PREPROCESSORS', 'nlp', 're-tokenizer')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/relation_extraction_preprocessor.py", "imports": ["transformers", "typing"], "module": "modelscope.preprocessors.nlp.relation_extraction_preprocessor"}, "('PREPROCESSORS', 'nlp', 'siamese-uie-preprocessor')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/siamese_uie_preprocessor.py", "imports": ["transformers", "typing"], "module": "modelscope.preprocessors.nlp.siamese_uie_preprocessor"}, "('PREPROCESSORS', 'nlp', 'document-grounded-dialog-rerank')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/document_grounded_dialog_rerank_preprocessor.py", "imports": ["transformers", "typing", "os", "torch", "copy"], "module": "modelscope.preprocessors.nlp.document_grounded_dialog_rerank_preprocessor"}, "('PREPROCESSORS', 'nlp', 'word-alignment')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/word_alignment_preprocessor.py", "imports": ["itertools", "numpy", "typing", "os", "torch"], "module": "modelscope.preprocessors.nlp.word_alignment_preprocessor"}, "('PREPROCESSORS', 'cv', 'image-quality_assessment-mos-preprocessor')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/cv/image_quality_assessment_mos.py", "imports": ["math", "numpy", "torchvision", "cv2", "typing"], "module": "modelscope.preprocessors.cv.image_quality_assessment_mos"}, "('PREPROCESSORS', 'cv', 'controllable-image-generation-preprocessor')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/cv/controllable_image_generation.py", "imports": ["math", "numpy", "torchvision", "cv2", "typing", "os", "torch", "PIL"], "module": "modelscope.preprocessors.cv.controllable_image_generation"}, "('PREPROCESSORS', 'cv', 'RandomCrop')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/cv/image_classification_preprocessor.py", "imports": ["numpy", "torchvision", "cv2", "typing", "os", "torch", "PIL"], "module": "modelscope.preprocessors.cv.image_classification_preprocessor"}, "('PREPROCESSORS', 'cv', 'RandomResizedCrop')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/cv/image_classification_preprocessor.py", "imports": ["numpy", "torchvision", "cv2", "typing", "os", "torch", "PIL"], "module": "modelscope.preprocessors.cv.image_classification_preprocessor"}, "('PREPROCESSORS', 'cv', 'Resize')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/cv/image_classification_preprocessor.py", "imports": ["numpy", "torchvision", "cv2", "typing", "os", "torch", "PIL"], "module": "modelscope.preprocessors.cv.image_classification_preprocessor"}, "('PREPROCESSORS', 'cv', 'CenterCrop')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/cv/image_classification_preprocessor.py", "imports": ["numpy", "torchvision", "cv2", "typing", "os", "torch", "PIL"], "module": "modelscope.preprocessors.cv.image_classification_preprocessor"}, "('PREPROCESSORS', 'cv', 'RandomHorizontalFlip')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/cv/image_classification_preprocessor.py", "imports": ["numpy", "torchvision", "cv2", "typing", "os", "torch", "PIL"], "module": "modelscope.preprocessors.cv.image_classification_preprocessor"}, "('PREPROCESSORS', 'cv', 'Normalize')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/cv/image_classification_preprocessor.py", "imports": ["numpy", "torchvision", "cv2", "typing", "os", "torch", "PIL"], "module": "modelscope.preprocessors.cv.image_classification_preprocessor"}, "('PREPROCESSORS', 'cv', 'ImageToTensor')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/cv/image_classification_preprocessor.py", "imports": ["numpy", "torchvision", "cv2", "typing", "os", "torch", "PIL"], "module": "modelscope.preprocessors.cv.image_classification_preprocessor"}, "('PREPROCESSORS', 'cv', 'image-classification-preprocessor')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/cv/image_classification_preprocessor.py", "imports": ["numpy", "torchvision", "cv2", "typing", "os", "torch", "PIL"], "module": "modelscope.preprocessors.cv.image_classification_preprocessor"}, "('PREPROCESSORS', 'cv', 'image-quality_assessment-man-preprocessor')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/cv/image_quality_assessment_man.py", "imports": ["math", "numpy", "torchvision", "typing", "torch", "PIL"], "module": "modelscope.preprocessors.cv.image_quality_assessment_man"}, "('PREPROCESSORS', 'cv', 'bad-image-detecting-preprocessor')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/cv/bad_image_detecting_preprocessor.py", "imports": ["math", "numpy", "torchvision", "typing", "torch", "PIL"], "module": "modelscope.preprocessors.cv.bad_image_detecting_preprocessor"}, "('PREPROCESSORS', 'cv', 'image-demoire-preprocessor')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/cv/image_restoration_preprocessor.py", "imports": ["math", "numpy", "torchvision", "typing", "torch", "PIL"], "module": "modelscope.preprocessors.cv.image_restoration_preprocessor"}, "('PREPROCESSORS', 'cv', 'image-classification-mmcv-preprocessor')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/cv/mmcls_preprocessor.py", "imports": ["os", "numpy", "typing"], "module": "modelscope.preprocessors.cv.mmcls_preprocessor"}, "('PREPROCESSORS', 'text-to-speech', 'kantts-data-preprocessor')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/tts.py", "imports": ["os", "kantts", "typing"], "module": "modelscope.preprocessors.tts"}, "('PREPROCESSORS', 'default', 'Compose')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/common.py", "imports": ["time", "numpy", "typing", "torch", "collections"], "module": "modelscope.preprocessors.common"}, "('PREPROCESSORS', 'default', 'ToTensor')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/common.py", "imports": ["time", "numpy", "typing", "torch", "collections"], "module": "modelscope.preprocessors.common"}, "('PREPROCESSORS', 'default', 'Filter')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/common.py", "imports": ["time", "numpy", "typing", "torch", "collections"], "module": "modelscope.preprocessors.common"}, "('PREPROCESSORS', 'default', 'ToNumpy')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/common.py", "imports": ["time", "numpy", "typing", "torch", "collections"], "module": "modelscope.preprocessors.common"}, "('PREPROCESSORS', 'default', 'Rename')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/common.py", "imports": ["time", "numpy", "typing", "torch", "collections"], "module": "modelscope.preprocessors.common"}, "('PREPROCESSORS', 'default', 'Identity')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/common.py", "imports": ["time", "numpy", "typing", "torch", "collections"], "module": "modelscope.preprocessors.common"}, "('LR_SCHEDULER', 'default', 'ConstantWarmup')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/lrscheduler/warmup/warmup.py", "imports": [], "module": "modelscope.trainers.lrscheduler.warmup.warmup"}, "('LR_SCHEDULER', 'default', 'LinearWarmup')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/lrscheduler/warmup/warmup.py", "imports": [], "module": "modelscope.trainers.lrscheduler.warmup.warmup"}, "('LR_SCHEDULER', 'default', 'ExponentialWarmup')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/lrscheduler/warmup/warmup.py", "imports": [], "module": "modelscope.trainers.lrscheduler.warmup.warmup"}, "('OPTIMIZERS', 'default', 'ChildTuningAdamW')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/optimizer/child_tuning_adamw_optimizer.py", "imports": ["math", "numpy", "typing", "types", "torch"], "module": "modelscope.trainers.optimizer.child_tuning_adamw_optimizer"}, "('TRAINERS', 'default', 'dummy')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/base.py", "imports": ["time", "os", "typing", "abc"], "module": "modelscope.trainers.base"}, "('TRAINERS', 'default', 'speech_dfsmn_kws_char_farfield')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/audio/kws_farfield_trainer.py", "imports": ["datetime", "pickle", "math", "numpy", "typing", "os", "torch", "glob"], "module": "modelscope.trainers.audio.kws_farfield_trainer"}, "('TRAINERS', 'default', 'speech_frcrn_ans_cirm_16k')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/audio/ans_trainer.py", "imports": [], "module": "modelscope.trainers.audio.ans_trainer"}, "('TRAINERS', 'default', 'speech-separation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/audio/separation_trainer.py", "imports": ["csv", "numpy", "typing", "os", "torch", "torchaudio", "speechbrain", "tqdm"], "module": "modelscope.trainers.audio.separation_trainer"}, "('TRAINERS', 'default', 'speech-asr-trainer')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/audio/asr_trainer.py", "imports": ["shutil", "json", "typing", "os", "tempfile", "funasr"], "module": "modelscope.trainers.audio.asr_trainer"}, "('TRAINERS', 'default', 'speech-kantts-trainer')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/audio/tts_trainer.py", "imports": ["shutil", "json", "typing", "os", "zipfile", "tempfile"], "module": "modelscope.trainers.audio.tts_trainer"}, "('TRAINERS', 'default', 'speech_kws_fsmn_char_ctc_nearfield')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/audio/kws_nearfield_trainer.py", "imports": ["datetime", "tensorboardX", "re", "yaml", "typing", "os", "torch", "copy"], "module": "modelscope.trainers.audio.kws_nearfield_trainer"}, "('TRAINERS', 'default', 'nlp-base-trainer')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/nlp_trainer.py", "imports": ["os", "numpy", "torch", "typing"], "module": "modelscope.trainers.nlp_trainer"}, "('TRAINERS', 'default', 'nlp-veco-trainer')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/nlp_trainer.py", "imports": ["os", "numpy", "torch", "typing"], "module": "modelscope.trainers.nlp_trainer"}, "('HOOKS', 'default', 'LrSchedulerHook')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/hooks/lr_scheduler_hook.py", "imports": [], "module": "modelscope.trainers.hooks.lr_scheduler_hook"}, "('HOOKS', 'default', 'PlateauLrSchedulerHook')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/hooks/lr_scheduler_hook.py", "imports": [], "module": "modelscope.trainers.hooks.lr_scheduler_hook"}, "('HOOKS', 'default', 'NoneLrSchedulerHook')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/hooks/lr_scheduler_hook.py", "imports": [], "module": "modelscope.trainers.hooks.lr_scheduler_hook"}, "('HOOKS', 'default', 'IterTimerHook')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/hooks/iter_timer_hook.py", "imports": ["time"], "module": "modelscope.trainers.hooks.iter_timer_hook"}, "('HOOKS', 'default', 'OptimizerHook')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/hooks/optimizer/base.py", "imports": ["logging", "torch"], "module": "modelscope.trainers.hooks.optimizer.base"}, "('HOOKS', 'default', 'NoneOptimizerHook')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/hooks/optimizer/base.py", "imports": ["logging", "torch"], "module": "modelscope.trainers.hooks.optimizer.base"}, "('HOOKS', 'default', 'TorchAMPOptimizerHook')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/hooks/optimizer/torch_optimizer_hook.py", "imports": ["logging"], "module": "modelscope.trainers.hooks.optimizer.torch_optimizer_hook"}, "('HOOKS', 'default', 'ApexAMPOptimizerHook')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/hooks/optimizer/apex_optimizer_hook.py", "imports": ["logging", "torch", "packaging"], "module": "modelscope.trainers.hooks.optimizer.apex_optimizer_hook"}, "('HOOKS', 'default', 'MegatronHook')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/hooks/megatron_hook.py", "imports": ["os", "torch", "megatron_util", "copy"], "module": "modelscope.trainers.hooks.megatron_hook"}, "('HOOKS', 'default', 'ClipClampLogitScaleHook')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/hooks/clip_clamp_logit_scale_hook.py", "imports": ["torch"], "module": "modelscope.trainers.hooks.clip_clamp_logit_scale_hook"}, "('HOOKS', 'default', 'DDPHook')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/hooks/ddp_hook.py", "imports": [], "module": "modelscope.trainers.hooks.ddp_hook"}, "('HOOKS', 'default', 'TensorboardHook')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/hooks/logger/tensorboard_hook.py", "imports": ["os", "numpy", "torch"], "module": "modelscope.trainers.hooks.logger.tensorboard_hook"}, "('HOOKS', 'default', 'TextLoggerHook')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/hooks/logger/text_logger_hook.py", "imports": ["datetime", "json", "os", "torch", "collections"], "module": "modelscope.trainers.hooks.logger.text_logger_hook"}, "('HOOKS', 'default', 'EvaluationHook')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/hooks/evaluation_hook.py", "imports": ["collections"], "module": "modelscope.trainers.hooks.evaluation_hook"}, "('HOOKS', 'default', 'CheckpointHook')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/hooks/checkpoint_hook.py", "imports": ["numpy", "re", "packaging", "os", "random", "torch"], "module": "modelscope.trainers.hooks.checkpoint_hook"}, "('HOOKS', 'default', 'BestCkptSaverHook')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/hooks/checkpoint_hook.py", "imports": ["numpy", "re", "packaging", "os", "random", "torch"], "module": "modelscope.trainers.hooks.checkpoint_hook"}, "('HOOKS', 'default', 'LoadCheckpointHook')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/hooks/checkpoint_hook.py", "imports": ["numpy", "re", "packaging", "os", "random", "torch"], "module": "modelscope.trainers.hooks.checkpoint_hook"}, "('HOOKS', 'default', 'EarlyStopHook')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/hooks/early_stop_hook.py", "imports": ["numpy"], "module": "modelscope.trainers.hooks.early_stop_hook"}, "('HOOKS', 'default', 'SparsityHook')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/hooks/compression/sparsity_hook.py", "imports": ["os"], "module": "modelscope.trainers.hooks.compression.sparsity_hook"}, "('HOOKS', 'default', 'DeepspeedHook')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/hooks/deepspeed_hook.py", "imports": ["shutil", "megatron_util", "deepspeed", "os", "torch"], "module": "modelscope.trainers.hooks.deepspeed_hook"}, "('PARALLEL', 'default', 'DistributedDataParallel')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/parallel/builder.py", "imports": ["torch"], "module": "modelscope.trainers.parallel.builder"}, "('TRAINERS', 'default', 'document-grounded-dialog-generate-trainer')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/nlp/document_grounded_dialog_generate_trainer.py", "imports": ["sacrebleu", "json", "re", "transformers", "string", "os", "torch", "rouge", "collections", "tqdm"], "module": "modelscope.trainers.nlp.document_grounded_dialog_generate_trainer"}, "('TRAINERS', 'default', 'csanmt-translation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/nlp/csanmt_translation_trainer.py", "imports": ["time", "tensorflow", "os", "typing"], "module": "modelscope.trainers.nlp.csanmt_translation_trainer"}, "('TRAINERS', 'default', 'document-grounded-dialog-rerank-trainer')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/nlp/document_grounded_dialog_rerank_trainer.py", "imports": ["time", "numpy", "transformers", "typing", "os", "random", "torch"], "module": "modelscope.trainers.nlp.document_grounded_dialog_rerank_trainer"}, "('TRAINERS', 'default', 'nlp-plug-trainer')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/nlp/plug_trainer.py", "imports": ["megatron_util", "typing", "deepspeed", "os", "torch"], "module": "modelscope.trainers.nlp.plug_trainer"}, "('TRAINERS', 'default', 'siamese-uie-trainer')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/nlp/siamese_uie_trainer.py", "imports": ["time", "math", "numpy", "json", "typing", "os", "random", "torch", "collections"], "module": "modelscope.trainers.nlp.siamese_uie_trainer"}, "('TRAINERS', 'default', 'document-grounded-dialog-retrieval-trainer')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/nlp/document_grounded_dialog_retrieval_trainer.py", "imports": ["numpy", "json", "transformers", "os", "torch", "faiss", "tqdm"], "module": "modelscope.trainers.nlp.document_grounded_dialog_retrieval_trainer"}, "('TRAINERS', 'default', 'nlp-gpt3-trainer')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/nlp/gpt3_trainer.py", "imports": ["os", "torch", "typing", "copy"], "module": "modelscope.trainers.nlp.gpt3_trainer"}, "('TRAINERS', 'default', 'table-question-answering-trainer')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/nlp/table_question_answering_trainer.py", "imports": ["time", "numpy", "json", "typing", "os", "torch", "tqdm"], "module": "modelscope.trainers.nlp.table_question_answering_trainer"}, "('TRAINERS', 'default', 'text-generation-trainer')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/nlp/text_generation_trainer.py", "imports": ["collections", "torch"], "module": "modelscope.trainers.nlp.text_generation_trainer"}, "('TRAINERS', 'default', 'nlp-gpt-moe-trainer')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/nlp/gpt_moe_trainer.py", "imports": ["megatron_util", "typing", "os", "torch", "collections"], "module": "modelscope.trainers.nlp.gpt_moe_trainer"}, "('TRAINERS', 'default', 'bert-sentiment-analysis')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/nlp/sequence_classification_trainer.py", "imports": ["time", "numpy", "typing"], "module": "modelscope.trainers.nlp.sequence_classification_trainer"}, "('TRAINERS', 'default', 'nlp-sentence-embedding-trainer')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/nlp/sentence_embedding_trainer.py", "imports": ["time", "numpy", "transformers", "dataclasses", "typing", "torch", "tqdm"], "module": "modelscope.trainers.nlp.sentence_embedding_trainer"}, "('TRAINERS', 'default', 'dialog-modeling-trainer')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/nlp/space/dialog_modeling_trainer.py", "imports": ["time", "os", "numpy", "typing"], "module": "modelscope.trainers.nlp.space.dialog_modeling_trainer"}, "('TRAINERS', 'default', 'dialog-intent-trainer')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/nlp/space/dialog_intent_trainer.py", "imports": ["os", "numpy", "typing"], "module": "modelscope.trainers.nlp.space.dialog_intent_trainer"}, "('TRAINERS', 'default', 'nlp-text-ranking-trainer')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/nlp/text_ranking_trainer.py", "imports": ["time", "numpy", "dataclasses", "typing", "torch", "tqdm"], "module": "modelscope.trainers.nlp.text_ranking_trainer"}, "('TRAINERS', 'default', 'faq-question-answering-trainer')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/nlp/faq_question_answering_trainer.py", "imports": ["contextlib", "distutils", "numpy", "dataclasses", "typing", "torch", "functools", "collections"], "module": "modelscope.trainers.nlp.faq_question_answering_trainer"}, "('TRAINERS', 'default', 'trainer')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/trainer.py", "imports": ["inspect", "distutils", "json", "typing", "os", "torch", "copy", "functools", "collections"], "module": "modelscope.trainers.trainer"}, "('TRAINERS', 'default', 'image-fewshot-detection')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/cv/image_defrcn_fewshot_detection_trainer.py", "imports": ["typing", "os", "torch", "collections", "detectron2"], "module": "modelscope.trainers.cv.image_defrcn_fewshot_detection_trainer"}, "('TRAINERS', 'default', 'image-instance-segmentation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/cv/image_instance_segmentation_trainer.py", "imports": [], "module": "modelscope.trainers.cv.image_instance_segmentation_trainer"}, "('TRAINERS', 'default', 'ocr-recognition')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/cv/ocr_recognition_trainer.py", "imports": ["time", "collections", "torch"], "module": "modelscope.trainers.cv.ocr_recognition_trainer"}, "('TRAINERS', 'default', 'card-detection-scrfd')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/cv/card_detection_scrfd_trainer.py", "imports": [], "module": "modelscope.trainers.cv.card_detection_scrfd_trainer"}, "('TRAINERS', 'default', 'face-detection-scrfd')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/cv/face_detection_scrfd_trainer.py", "imports": ["time", "os", "typing", "copy"], "module": "modelscope.trainers.cv.face_detection_scrfd_trainer"}, "('TRAINERS', 'default', 'image-inpainting')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/cv/image_inpainting_trainer.py", "imports": ["time", "collections", "torch"], "module": "modelscope.trainers.cv.image_inpainting_trainer"}, "('TRAINERS', 'default', 'movie-scene-segmentation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/cv/movie_scene_segmentation_trainer.py", "imports": [], "module": "modelscope.trainers.cv.movie_scene_segmentation_trainer"}, "('TRAINERS', 'default', 'image-classification')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/cv/image_classifition_trainer.py", "imports": ["time", "numpy", "typing", "os", "torch", "copy"], "module": "modelscope.trainers.cv.image_classifition_trainer"}, "('TRAINERS', 'default', 'referring-video-object-segmentation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/cv/referring_video_object_segmentation_trainer.py", "imports": ["os", "torch"], "module": "modelscope.trainers.cv.referring_video_object_segmentation_trainer"}, "('TRAINERS', 'default', 'cartoon-translation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/cv/cartoon_translation_trainer.py", "imports": ["numpy", "typing", "packaging", "os", "tensorflow", "tqdm"], "module": "modelscope.trainers.cv.cartoon_translation_trainer"}, "('TRAINERS', 'default', 'action-detection')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/cv/action_detection_trainer.py", "imports": ["typing", "os", "fvcore", "torch", "detectron2"], "module": "modelscope.trainers.cv.action_detection_trainer"}, "('TRAINERS', 'default', 'nerf-recon-acc')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/cv/nerf_recon_acc_trainer.py", "imports": ["time", "datetime", "numpy", "cv2", "typing", "os", "random", "torch", "glob", "tqdm"], "module": "modelscope.trainers.cv.nerf_recon_acc_trainer"}, "('TRAINERS', 'default', 'ocr-detection-db')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/cv/ocr_detection_db_trainer.py", "imports": ["datetime", "time", "easydict", "math", "numpy", "typing", "os", "copy", "torch", "tqdm"], "module": "modelscope.trainers.cv.ocr_detection_db_trainer"}, "('TRAINERS', 'default', 'vision-efficient-tuning')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/cv/vision_efficient_tuning_trainer.py", "imports": ["typing", "torch"], "module": "modelscope.trainers.cv.vision_efficient_tuning_trainer"}, "('TRAINERS', 'default', 'image-portrait-enhancement')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/cv/image_portrait_enhancement_trainer.py", "imports": ["collections", "torch"], "module": "modelscope.trainers.cv.image_portrait_enhancement_trainer"}, "('TRAINERS', 'default', 'tinynas-damoyolo')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/cv/image_detection_damoyolo_trainer.py", "imports": ["datetime", "time", "easydict", "math", "typing", "os", "torch"], "module": "modelscope.trainers.cv.image_detection_damoyolo_trainer"}, "('HOOKS', 'default', 'AddLrLogHook')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/easycv/utils/hooks.py", "imports": [], "module": "modelscope.trainers.easycv.utils.hooks"}, "('METRICS', 'default', 'EasyCVMetric')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/easycv/utils/metric.py", "imports": ["torch", "numpy", "typing", "itertools"], "module": "modelscope.trainers.easycv.utils.metric"}, "('TRAINERS', 'default', 'easycv')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/easycv/trainer.py", "imports": ["typing", "easycv", "torch", "functools", "copy"], "module": "modelscope.trainers.easycv.trainer"}, "('TRAINERS', 'default', 'mplug')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/multi_modal/mplug/mplug_trainer.py", "imports": ["torch", "collections", "typing"], "module": "modelscope.trainers.multi_modal.mplug.mplug_trainer"}, "('TRAINERS', 'default', 'image-classification-team')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/multi_modal/team/team_trainer.py", "imports": ["sklearn", "numpy", "typing", "os", "torch", "collections"], "module": "modelscope.trainers.multi_modal.team.team_trainer"}, "('TRAINERS', 'default', 'mgeo-ranking-trainer')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/multi_modal/mgeo_ranking_trainer.py", "imports": ["torch", "dataclasses", "typing"], "module": "modelscope.trainers.multi_modal.mgeo_ranking_trainer"}, "('TRAINERS', 'default', 'clip-multi-modal-embedding')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/multi_modal/clip/clip_trainer.py", "imports": ["os", "torch", "math", "typing"], "module": "modelscope.trainers.multi_modal.clip.clip_trainer"}, "('TRAINERS', 'default', 'ofa')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/multi_modal/ofa/ofa_trainer.py", "imports": ["shutil", "math", "json", "typing", "os", "tempfile", "torch", "functools"], "module": "modelscope.trainers.multi_modal.ofa.ofa_trainer"}, "('CUSTOM_DATASETS', 'image-deblurring', 'RedsDataset')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/reds_image_deblurring_dataset.py", "imports": ["numpy", "cv2"], "module": "modelscope.msdatasets.dataset_cls.custom_datasets.reds_image_deblurring_dataset"}, "('CUSTOM_DATASETS', 'image-inpainting', 'FFTInpainting')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/image_inpainting/image_inpainting_dataset.py", "imports": ["albumentations", "enum", "numpy", "cv2", "os", "glob"], "module": "modelscope.msdatasets.dataset_cls.custom_datasets.image_inpainting.image_inpainting_dataset"}, "('CUSTOM_DATASETS', 'movie-scene-segmentation', 'resnet50-bert')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/movie_scene_segmentation/movie_scene_segmentation_dataset.py", "imports": ["json", "torchvision", "os", "random", "torch", "copy"], "module": "modelscope.msdatasets.dataset_cls.custom_datasets.movie_scene_segmentation.movie_scene_segmentation_dataset"}, "('CUSTOM_DATASETS', 'image-deblurring', 'GoproDataset')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/gopro_image_deblurring_dataset.py", "imports": ["numpy", "cv2"], "module": "modelscope.msdatasets.dataset_cls.custom_datasets.gopro_image_deblurring_dataset"}, "('CUSTOM_DATASETS', 'image-portrait-enhancement', 'PairedDataset')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/image_portrait_enhancement/image_portrait_enhancement_dataset.py", "imports": ["numpy", "cv2"], "module": "modelscope.msdatasets.dataset_cls.custom_datasets.image_portrait_enhancement.image_portrait_enhancement_dataset"}, "('CUSTOM_DATASETS', 'image-classification', 'ClsDataset')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/image_classification/classification_dataset.py", "imports": ["easycv"], "module": "modelscope.msdatasets.dataset_cls.custom_datasets.image_classification.classification_dataset"}, "('CUSTOM_DATASETS', 'language-guided-video-summarization', 'clip-it-language-guided-video-summarization')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/language_guided_video_summarization_dataset.py", "imports": ["numpy", "json", "os", "torch", "h5py"], "module": "modelscope.msdatasets.dataset_cls.custom_datasets.language_guided_video_summarization_dataset"}, "('CUSTOM_DATASETS', 'image-denoising', 'SiddDataset')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/sidd_image_denoising/sidd_image_denoising_dataset.py", "imports": ["numpy", "cv2"], "module": "modelscope.msdatasets.dataset_cls.custom_datasets.sidd_image_denoising.sidd_image_denoising_dataset"}, "('CUSTOM_DATASETS', 'referring-video-object-segmentation', 'swinT-referring-video-object-segmentation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/referring_video_object_segmentation/referring_video_object_segmentation_dataset.py", "imports": ["pandas", "pycocotools", "numpy", "json", "torchvision", "os", "torch", "h5py", "glob", "tqdm"], "module": "modelscope.msdatasets.dataset_cls.custom_datasets.referring_video_object_segmentation.referring_video_object_segmentation_dataset"}, "('CUSTOM_DATASETS', 'image-quality-assessment-mos', 'image-quality-assessment-mos')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/image_quality_assmessment_mos/image_quality_assessment_mos_dataset.py", "imports": [], "module": "modelscope.msdatasets.dataset_cls.custom_datasets.image_quality_assmessment_mos.image_quality_assessment_mos_dataset"}, "('CUSTOM_DATASETS', 'video-super-resolution', 'real-basicvsr')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/video_super_resolution/video_super_resolution_dataset.py", "imports": ["torch", "numpy", "collections", "cv2"], "module": "modelscope.msdatasets.dataset_cls.custom_datasets.video_super_resolution.video_super_resolution_dataset"}, "('CUSTOM_DATASETS', 'image-object-detection', 'DetDataset')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/object_detection/detection_dataset.py", "imports": ["easycv"], "module": "modelscope.msdatasets.dataset_cls.custom_datasets.object_detection.detection_dataset"}, "('CUSTOM_DATASETS', 'image-segmentation', 'DetDataset')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/object_detection/detection_dataset.py", "imports": ["easycv"], "module": "modelscope.msdatasets.dataset_cls.custom_datasets.object_detection.detection_dataset"}, "('CUSTOM_DATASETS', 'image-object-detection', 'DetImagesMixDataset')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/object_detection/detection_dataset.py", "imports": ["easycv"], "module": "modelscope.msdatasets.dataset_cls.custom_datasets.object_detection.detection_dataset"}, "('CUSTOM_DATASETS', 'domain-specific-object-detection', 'DetImagesMixDataset')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/object_detection/detection_dataset.py", "imports": ["easycv"], "module": "modelscope.msdatasets.dataset_cls.custom_datasets.object_detection.detection_dataset"}, "('CUSTOM_DATASETS', 'image-segmentation', 'cascade_mask_rcnn_swin')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/image_instance_segmentation_coco_dataset.py", "imports": ["os", "numpy", "pycocotools"], "module": "modelscope.msdatasets.dataset_cls.custom_datasets.image_instance_segmentation_coco_dataset"}, "('CUSTOM_DATASETS', 'image-colorization', 'ddcolor')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/image_colorization/image_colorization_dataset.py", "imports": ["cv2", "numpy", "torch"], "module": "modelscope.msdatasets.dataset_cls.custom_datasets.image_colorization.image_colorization_dataset"}, "('CUSTOM_DATASETS', 'text-ranking', 'mgeo')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/mgeo_ranking_dataset.py", "imports": ["random", "torch", "json", "typing"], "module": "modelscope.msdatasets.dataset_cls.custom_datasets.mgeo_ranking_dataset"}, "('CUSTOM_DATASETS', 'image-quality-assessment-degradation', 'image-quality-assessment-degradation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/image_quality_assessment_degradation/image_quality_assessment_degradation_dataset.py", "imports": ["torchvision"], "module": "modelscope.msdatasets.dataset_cls.custom_datasets.image_quality_assessment_degradation.image_quality_assessment_degradation_dataset"}, "('CUSTOM_DATASETS', 'face-2d-keypoints', 'FaceKeypointDataset')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/face_2d_keypoins/face_2d_keypoints_dataset.py", "imports": ["easycv"], "module": "modelscope.msdatasets.dataset_cls.custom_datasets.face_2d_keypoins.face_2d_keypoints_dataset"}, "('CUSTOM_DATASETS', 'text-ranking', 'bert')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/text_ranking_dataset.py", "imports": ["random", "torch", "typing"], "module": "modelscope.msdatasets.dataset_cls.custom_datasets.text_ranking_dataset"}, "('CUSTOM_DATASETS', 'sentence-embedding', 'bert')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/text_ranking_dataset.py", "imports": ["random", "torch", "typing"], "module": "modelscope.msdatasets.dataset_cls.custom_datasets.text_ranking_dataset"}, "('CUSTOM_DATASETS', 'bad-image-detecting', 'bad-image-detecting')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/bad_image_detecting/bad_image_detecting_dataset.py", "imports": [], "module": "modelscope.msdatasets.dataset_cls.custom_datasets.bad_image_detecting.bad_image_detecting_dataset"}, "('CUSTOM_DATASETS', 'video-frame-interpolation', 'video-frame-interpolation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/video_frame_interpolation/video_frame_interpolation_dataset.py", "imports": ["cv2", "numpy", "torch"], "module": "modelscope.msdatasets.dataset_cls.custom_datasets.video_frame_interpolation.video_frame_interpolation_dataset"}, "('CUSTOM_DATASETS', 'nli', 'veco')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/veco_dataset.py", "imports": ["numpy", "datasets", "typing"], "module": "modelscope.msdatasets.dataset_cls.custom_datasets.veco_dataset"}, "('CUSTOM_DATASETS', 'human-wholebody-keypoint', 'WholeBodyCocoTopDownDataset')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/human_wholebody_keypoint/human_wholebody_keypoint_dataset.py", "imports": ["easycv"], "module": "modelscope.msdatasets.dataset_cls.custom_datasets.human_wholebody_keypoint.human_wholebody_keypoint_dataset"}, "('CUSTOM_DATASETS', 'image-segmentation', 'SegDataset')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/image_semantic_segmentation/segmentation_dataset.py", "imports": ["easycv"], "module": "modelscope.msdatasets.dataset_cls.custom_datasets.image_semantic_segmentation.segmentation_dataset"}, "('CUSTOM_DATASETS', 'hand-2d-keypoints', 'HandCocoWholeBodyDataset')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/hand_2d_keypoints/hand_2d_keypoints_dataset.py", "imports": ["easycv"], "module": "modelscope.msdatasets.dataset_cls.custom_datasets.hand_2d_keypoints.hand_2d_keypoints_dataset"}, "('CUSTOM_DATASETS', 'ocr-recognition', 'OCRRecognition')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/ocr_recognition_dataset.py", "imports": ["six", "numpy", "json", "cv2", "os", "torch", "lmdb", "PIL"], "module": "modelscope.msdatasets.dataset_cls.custom_datasets.ocr_recognition_dataset"}, "('CUSTOM_DATASETS', 'video-stabilization', 'video-stabilization')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/video_stabilization/video_stabilization_dataset.py", "imports": [], "module": "modelscope.msdatasets.dataset_cls.custom_datasets.video_stabilization.video_stabilization_dataset"}, "('EXPORTERS', 'translation', 'csanmt-translation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/exporters/nlp/csanmt_for_translation_exporter.py", "imports": ["os", "tensorflow", "typing"], "module": "modelscope.exporters.nlp.csanmt_for_translation_exporter"}, "('EXPORTERS', 'text-classification', 'bert')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/exporters/nlp/sbert_for_sequence_classification_exporter.py", "imports": ["collections", "typing", "torch"], "module": "modelscope.exporters.nlp.sbert_for_sequence_classification_exporter"}, "('EXPORTERS', 'text-classification', 'structbert')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/exporters/nlp/sbert_for_sequence_classification_exporter.py", "imports": ["collections", "typing", "torch"], "module": "modelscope.exporters.nlp.sbert_for_sequence_classification_exporter"}, "('EXPORTERS', 'sentence-similarity', 'bert')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/exporters/nlp/sbert_for_sequence_classification_exporter.py", "imports": ["collections", "typing", "torch"], "module": "modelscope.exporters.nlp.sbert_for_sequence_classification_exporter"}, "('EXPORTERS', 'sentiment-classification', 'bert')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/exporters/nlp/sbert_for_sequence_classification_exporter.py", "imports": ["collections", "typing", "torch"], "module": "modelscope.exporters.nlp.sbert_for_sequence_classification_exporter"}, "('EXPORTERS', 'nli', 'bert')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/exporters/nlp/sbert_for_sequence_classification_exporter.py", "imports": ["collections", "typing", "torch"], "module": "modelscope.exporters.nlp.sbert_for_sequence_classification_exporter"}, "('EXPORTERS', 'sentence-similarity', 'structbert')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/exporters/nlp/sbert_for_sequence_classification_exporter.py", "imports": ["collections", "typing", "torch"], "module": "modelscope.exporters.nlp.sbert_for_sequence_classification_exporter"}, "('EXPORTERS', 'sentiment-classification', 'structbert')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/exporters/nlp/sbert_for_sequence_classification_exporter.py", "imports": ["collections", "typing", "torch"], "module": "modelscope.exporters.nlp.sbert_for_sequence_classification_exporter"}, "('EXPORTERS', 'nli', 'structbert')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/exporters/nlp/sbert_for_sequence_classification_exporter.py", "imports": ["collections", "typing", "torch"], "module": "modelscope.exporters.nlp.sbert_for_sequence_classification_exporter"}, "('EXPORTERS', 'transformer-crf', 'transformer-crf')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/exporters/nlp/model_for_token_classification_exporter.py", "imports": ["torch", "collections", "typing"], "module": "modelscope.exporters.nlp.model_for_token_classification_exporter"}, "('EXPORTERS', 'token-classification', 'transformer-crf')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/exporters/nlp/model_for_token_classification_exporter.py", "imports": ["torch", "collections", "typing"], "module": "modelscope.exporters.nlp.model_for_token_classification_exporter"}, "('EXPORTERS', 'named-entity-recognition', 'transformer-crf')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/exporters/nlp/model_for_token_classification_exporter.py", "imports": ["torch", "collections", "typing"], "module": "modelscope.exporters.nlp.model_for_token_classification_exporter"}, "('EXPORTERS', 'part-of-speech', 'transformer-crf')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/exporters/nlp/model_for_token_classification_exporter.py", "imports": ["torch", "collections", "typing"], "module": "modelscope.exporters.nlp.model_for_token_classification_exporter"}, "('EXPORTERS', 'word-segmentation', 'transformer-crf')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/exporters/nlp/model_for_token_classification_exporter.py", "imports": ["torch", "collections", "typing"], "module": "modelscope.exporters.nlp.model_for_token_classification_exporter"}, "('EXPORTERS', 'zero-shot-classification', 'bert')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/exporters/nlp/sbert_for_zero_shot_classification_exporter.py", "imports": ["collections", "typing"], "module": "modelscope.exporters.nlp.sbert_for_zero_shot_classification_exporter"}, "('EXPORTERS', 'zero-shot-classification', 'structbert')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/exporters/nlp/sbert_for_zero_shot_classification_exporter.py", "imports": ["collections", "typing"], "module": "modelscope.exporters.nlp.sbert_for_zero_shot_classification_exporter"}, "('EXPORTERS', 'image-object-detection', 'tinynas-damoyolo')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/exporters/cv/object_detection_damoyolo_exporter.py", "imports": ["numpy", "typing", "os", "functools", "torch", "onnx"], "module": "modelscope.exporters.cv.object_detection_damoyolo_exporter"}, "('EXPORTERS', 'default', 'cartoon-translation')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/exporters/cv/cartoon_translation_exporter.py", "imports": ["os", "tensorflow", "typing", "packaging"], "module": "modelscope.exporters.cv.cartoon_translation_exporter"}, "('EXPORTERS', 'face-detection', 'scrfd')": {"filepath": "/usr/local/lib/python3.9/dist-packages/modelscope/exporters/cv/face_detection_scrfd_exporter.py", "imports": ["numpy", "typing", "os", "functools", "torch", "onnx"], "module": "modelscope.exporters.cv.face_detection_scrfd_exporter"}}, "requirements": {"modelscope.models.builder": [], "modelscope.models.audio.itn.generic_inverse_text_processing": ["os", "typing"], "modelscope.models.audio.sv.generic_speaker_verification": ["os", "typing"], "modelscope.models.audio.sv.ecapa_tdnn": ["math", "typing", "os", "torch", "torchaudio"], "modelscope.models.audio.ans.denoise_net": ["torch"], "modelscope.models.audio.ans.unet": ["torch"], "modelscope.models.audio.ans.frcrn": ["os", "torch", "typing"], "modelscope.models.audio.ans.complex_nn": ["torch"], "modelscope.models.audio.ans.conv_stft": ["scipy", "numpy", "torch"], "modelscope.models.audio.ans.layers.activations": ["torch"], "modelscope.models.audio.ans.layers.uni_deep_fsmn": ["numpy", "torch"], "modelscope.models.audio.ans.layers.affine_transform": ["torch"], "modelscope.models.audio.ans.layers.layer_base": ["numpy", "six", "torch", "abc"], "modelscope.models.audio.ans.se_module_complex": ["torch"], "modelscope.models.audio.tts.sambert_hifi": ["datetime", "shutil", "matplotlib", "numpy", "json", "yaml", "os", "zipfile", "wave", "__future__"], "modelscope.models.audio.tts.voice": ["time", "pickle", "threading", "numpy", "json", "yaml", "os", "torch", "kantts", "collections"], "modelscope.models.audio.punc.generic_punctuation": ["os", "typing"], "modelscope.models.audio.separation.mossformer_conv_module": ["torch"], "modelscope.models.audio.separation.mossformer": ["os", "torch", "typing", "copy"], "modelscope.models.audio.separation.mossformer_block": ["torch"], "modelscope.models.audio.separation.layer_norm": ["__future__", "torch"], "modelscope.models.audio.kws.farfield.model": ["os", "tempfile", "typing"], "modelscope.models.audio.kws.farfield.fsmn": ["numpy", "torch"], "modelscope.models.audio.kws.farfield.model_def": ["enum", "math", "struct"], "modelscope.models.audio.kws.farfield.fsmn_sele_v2": ["torch"], "modelscope.models.audio.kws.generic_key_word_spotting": ["os", "typing"], "modelscope.models.audio.kws.nearfield.model": ["typing", "os", "tempfile", "torch", "sys"], "modelscope.models.audio.kws.nearfield.fsmn": ["torch", "numpy", "typing"], "modelscope.models.audio.kws.nearfield.cmvn": ["numpy", "re", "torch"], "modelscope.models.audio.asr.wenet_automatic_speech_recognition": ["os", "wenetruntime", "json", "typing"], "modelscope.models.audio.asr.generic_automatic_speech_recognition": ["os", "typing"], "modelscope.models.audio.aec.network.se_net": ["torch"], "modelscope.models.audio.aec.network.modulation_loss": ["torch", "math", "torchaudio"], "modelscope.models.audio.aec.network.loss": ["torch"], "modelscope.models.audio.aec.layers.activations": ["torch"], "modelscope.models.audio.aec.layers.deep_fsmn": ["numpy", "torch"], "modelscope.models.audio.aec.layers.uni_deep_fsmn": ["numpy", "torch"], "modelscope.models.audio.aec.layers.affine_transform": ["numpy", "torch"], "modelscope.models.audio.aec.layers.layer_base": ["numpy", "re", "torch", "abc"], "modelscope.models.science.unifold.model": ["os", "argparse", "torch", "typing"], "modelscope.models.science.unifold.data.data_ops": ["itertools", "operator", "numpy", "typing", "unicore", "torch", "functools"], "modelscope.models.science.unifold.data.process": ["torch", "numpy", "typing"], "modelscope.models.science.unifold.data.utils": ["pickle", "scipy", "numpy", "json", "gzip", "typing", "functools", "copy"], "modelscope.models.science.unifold.data.protein": ["io", "numpy", "dataclasses", "typing", "Bio"], "modelscope.models.science.unifold.data.msa_pairing": ["pandas", "scipy", "numpy", "typing", "collections"], "modelscope.models.science.unifold.data.residue_constants": ["os", "functools", "numpy", "collections", "unicore", "typing"], "modelscope.models.science.unifold.data.process_multimer": ["numpy", "collections", "typing"], "modelscope.models.science.unifold.msa.templates": ["datetime", "abc", "numpy", "re", "dataclasses", "typing", "os", "functools", "glob", "absl"], "modelscope.models.science.unifold.msa.parsers": ["itertools", "typing", "collections", "re", "dataclasses", "string"], "modelscope.models.science.unifold.msa.utils": ["os", "json", "typing", "absl"], "modelscope.models.science.unifold.msa.mmcif": ["io", "functools", "absl", "collections", "dataclasses", "typing", "Bio"], "modelscope.models.science.unifold.msa.pipeline": ["os", "numpy", "typing", "absl"], "modelscope.models.science.unifold.msa.tools.hmmsearch": ["os", "subprocess", "typing", "absl"], "modelscope.models.science.unifold.msa.tools.hhsearch": ["subprocess", "typing", "os", "glob", "absl"], "modelscope.models.science.unifold.msa.tools.hmmbuild": ["os", "subprocess", "re", "absl"], "modelscope.models.science.unifold.msa.tools.utils": ["contextlib", "time", "shutil", "tempfile", "typing", "absl"], "modelscope.models.science.unifold.msa.tools.kalign": ["os", "subprocess", "typing", "absl"], "modelscope.models.science.unifold.msa.tools.jackhmmer": ["subprocess", "typing", "os", "urllib", "concurrent", "glob", "absl"], "modelscope.models.science.unifold.msa.tools.hhblits": ["subprocess", "typing", "os", "glob", "absl"], "modelscope.models.science.unifold.msa.msa_identifiers": ["re", "dataclasses", "typing"], "modelscope.models.science.unifold.modules.embedders": ["torch", "unicore", "typing"], "modelscope.models.science.unifold.modules.confidence": ["torch", "typing"], "modelscope.models.science.unifold.modules.alphafold": ["torch", "unicore"], "modelscope.models.science.unifold.modules.auxillary_heads": ["torch", "unicore", "typing"], "modelscope.models.science.unifold.modules.frame": ["__future__", "numpy", "typing", "torch"], "modelscope.models.science.unifold.modules.evoformer": ["torch", "functools", "unicore", "typing"], "modelscope.models.science.unifold.modules.attentions": ["torch", "functools", "unicore", "typing"], "modelscope.models.science.unifold.modules.structure_module": ["torch", "typing", "math", "unicore"], "modelscope.models.science.unifold.modules.template": ["math", "typing", "unicore", "torch", "functools"], "modelscope.models.science.unifold.modules.featurization": ["torch", "unicore", "typing"], "modelscope.models.science.unifold.modules.triangle_multiplication": ["torch", "functools", "unicore", "typing"], "modelscope.models.science.unifold.modules.common": ["functools", "typing", "unicore", "torch"], "modelscope.models.science.unifold.config": ["ml_collections", "typing", "copy"], "modelscope.models.science.unifold.dataset": ["numpy", "json", "typing", "unicore", "os", "copy", "torch", "logging", "ml_collections"], "modelscope.models.nlp.xlm_roberta.backbone": ["transformers", "math", "torch", "packaging"], "modelscope.models.nlp.xlm_roberta.configuration": ["collections", "transformers", "typing"], "modelscope.models.nlp.bart.text_error_correction": ["os", "torch", "typing"], "modelscope.models.nlp.T5.backbone": ["math", "transformers", "typing", "os", "torch", "copy", "warnings"], "modelscope.models.nlp.T5.configuration": ["transformers", "typing"], "modelscope.models.nlp.T5.text2text_generation": ["transformers", "typing", "torch", "copy", "warnings"], "modelscope.models.nlp.unite.modeling_unite": ["math", "numpy", "transformers", "dataclasses", "typing", "packaging", "torch", "warnings"], "modelscope.models.nlp.unite.configuration_unite": ["enum"], "modelscope.models.nlp.gpt_neo.backbone": ["transformers"], "modelscope.models.nlp.gpt2.backbone": ["transformers"], "modelscope.models.nlp.heads.torch_pretrain_head": ["torch", "transformers", "typing"], "modelscope.models.nlp.heads.text_classification_head": ["torch", "typing"], "modelscope.models.nlp.heads.crf_head": ["torch", "transformers", "typing"], "modelscope.models.nlp.heads.infromation_extraction_head": ["torch"], "modelscope.models.nlp.heads.text_ranking_head": ["torch", "typing"], "modelscope.models.nlp.heads.token_classification_head": ["torch", "typing"], "modelscope.models.nlp.heads.fill_mask_head": ["torch", "transformers", "typing"], "modelscope.models.nlp.heads.text_generation_head": ["torch", "typing"], "modelscope.models.nlp.plug.backbone": ["math", "megatron_util", "torch", "__future__", "logging"], "modelscope.models.nlp.plug.generator": ["torch"], "modelscope.models.nlp.plug.distributed_plug": ["torch", "typing", "megatron_util"], "modelscope.models.nlp.plug.configuration": ["json", "transformers", "copy"], "modelscope.models.nlp.plug.AnnealingLR": ["math", "torch"], "modelscope.models.nlp.hf_transformers.backbone": ["transformers"], "modelscope.models.nlp.codegeex.codegeex": ["math", "torch"], "modelscope.models.nlp.codegeex.tokenizer": ["torch", "transformers", "typing"], "modelscope.models.nlp.codegeex.inference": ["torch", "typing"], "modelscope.models.nlp.codegeex.codegeex_for_code_generation": ["torch", "typing", "copy"], "modelscope.models.nlp.codegeex.codegeex_for_code_translation": ["torch", "typing", "copy"], "modelscope.models.nlp.space_T_en.text_to_sql": ["os", "torch", "text2sql_lgesql", "typing"], "modelscope.models.nlp.palm_v2.configuration": ["transformers"], "modelscope.models.nlp.palm_v2.dureader_eval": ["math", "numpy", "json", "re", "zipfile", "copy", "rouge", "sys", "collections", "argparse"], "modelscope.models.nlp.palm_v2.text_generation": ["math", "subprocess", "numpy", "json", "transformers", "dataclasses", "typing", "os", "copy", "torch", "codecs"], "modelscope.models.nlp.dgds.backbone": ["os", "__future__", "transformers", "torch"], "modelscope.models.nlp.dgds.document_grounded_dialog_retrieval": ["os", "torch", "typing"], "modelscope.models.nlp.dgds.document_grounded_dialog_rerank": ["os", "torch", "typing"], "modelscope.models.nlp.dgds.document_grounded_dialog_generate": ["os", "torch", "typing"], "modelscope.models.nlp.lstm.backbone": ["torch"], "modelscope.models.nlp.lstm.token_classification": [], "modelscope.models.nlp.bloom.backbone": ["transformers"], "modelscope.models.nlp.task_models.text_classification": ["numpy", "typing"], "modelscope.models.nlp.task_models.feature_extraction": ["numpy", "typing"], "modelscope.models.nlp.task_models.task_model": ["abc", "re", "typing", "os", "torch", "collections"], "modelscope.models.nlp.task_models.information_extraction": ["numpy", "typing"], "modelscope.models.nlp.task_models.token_classification": ["torch", "typing"], "modelscope.models.nlp.task_models.fill_mask": ["torch", "numpy", "typing"], "modelscope.models.nlp.task_models.text_generation": ["numpy", "transformers", "typing"], "modelscope.models.nlp.task_models.text_ranking": ["numpy", "typing"], "modelscope.models.nlp.peer.backbone": ["math", "transformers", "dataclasses", "typing", "torch"], "modelscope.models.nlp.peer.text_classification": ["torch", "copy"], "modelscope.models.nlp.peer.configuration": ["transformers"], "modelscope.models.nlp.peer.sas_utils": ["random", "nltk", "torch", "numpy"], "modelscope.models.nlp.deberta_v2.backbone": ["torch", "collections", "transformers", "typing"], "modelscope.models.nlp.deberta_v2.configuration": ["transformers"], "modelscope.models.nlp.deberta_v2.tokenization": ["os", "unicodedata", "transformers", "typing", "sentencepiece"], "modelscope.models.nlp.deberta_v2.fill_mask": ["torch", "transformers", "typing"], "modelscope.models.nlp.deberta_v2.tokenization_fast": ["os", "shutil", "transformers", "typing"], "modelscope.models.nlp.megatron_bert.backbone": ["transformers", "math", "torch"], "modelscope.models.nlp.megatron_bert.configuration": ["collections", "transformers", "typing"], "modelscope.models.nlp.megatron_bert.fill_mask": ["transformers", "torch"], "modelscope.models.nlp.structbert.backbone": ["math", "transformers", "dataclasses", "typing", "packaging", "torch"], "modelscope.models.nlp.structbert.text_classification": ["torch"], "modelscope.models.nlp.structbert.configuration": ["transformers"], "modelscope.models.nlp.structbert.adv_utils": ["torch"], "modelscope.models.nlp.structbert.faq_question_answering": ["math", "typing", "os", "torch", "collections"], "modelscope.models.nlp.structbert.token_classification": ["torch"], "modelscope.models.nlp.structbert.fill_mask": ["transformers", "torch"], "modelscope.models.nlp.plug_mental.backbone": ["math", "transformers", "dataclasses", "typing", "packaging", "torch"], "modelscope.models.nlp.plug_mental.text_classification": ["torch"], "modelscope.models.nlp.plug_mental.configuration": ["transformers"], "modelscope.models.nlp.plug_mental.adv_utils": ["torch"], "modelscope.models.nlp.gpt3.backbone": ["math", "transformers", "typing", "os", "torch", "addict"], "modelscope.models.nlp.gpt3.configuration": ["transformers", "torch"], "modelscope.models.nlp.gpt3.distributed_gpt3": ["math", "megatron_util", "transformers", "typing", "os", "torch", "collections"], "modelscope.models.nlp.gpt3.tokenizer": ["tokenizers", "typing"], "modelscope.models.nlp.gpt3.text_generation": ["torch", "collections", "transformers", "typing"], "modelscope.models.nlp.csanmt.translation": ["tensorflow", "collections", "math", "typing"], "modelscope.models.nlp.space_T_cn.backbone": ["tarfile", "shutil", "math", "numpy", "os", "tempfile", "copy", "torch", "__future__"], "modelscope.models.nlp.space_T_cn.table_question_answering": ["numpy", "transformers", "typing", "os", "torch"], "modelscope.models.nlp.space_T_cn.configuration": ["__future__", "json", "logging", "copy"], "modelscope.models.nlp.mglm.mglm_for_text_summarization": ["megatron_util", "numpy", "typing", "os", "random", "torch"], "modelscope.models.nlp.mglm.generation_utils": ["collections", "typing", "torch", "abc"], "modelscope.models.nlp.mglm.blocklm_utils": ["random", "math", "megatron_util", "torch", "copy", "scipy", "numpy"], "modelscope.models.nlp.mglm.configure_data": ["bisect", "itertools", "megatron_util", "numpy", "os", "random", "torch", "copy"], "modelscope.models.nlp.mglm.utils": ["time", "megatron_util", "subprocess", "numpy", "json", "os", "random", "torch"], "modelscope.models.nlp.mglm.tasks.eval_utils": ["datetime", "time", "sklearn", "megatron_util", "typing", "os", "random", "finetune_glm", "torch", "utils", "collections", "tasks"], "modelscope.models.nlp.mglm.tasks.data_utils": ["pickle", "megatron_util", "numpy", "json", "re", "typing", "copy", "torch"], "modelscope.models.nlp.mglm.tasks.language_model.finetune": ["finetune_glm", "math", "megatron_util", "torch", "functools", "pretrain_glm", "tasks"], "modelscope.models.nlp.mglm.tasks.language_model.detokenizer": ["re"], "modelscope.models.nlp.mglm.tasks.language_model.dataset": ["bisect", "math", "itertools", "numpy", "json", "torch", "utils", "tasks"], "modelscope.models.nlp.mglm.tasks.superglue.pvp": ["math", "abc", "numpy", "string", "typing", "random", "copy", "utils", "collections", "tasks"], "modelscope.models.nlp.mglm.tasks.superglue.finetune": ["finetune_glm", "collections", "tasks"], "modelscope.models.nlp.mglm.tasks.superglue.evaluate": ["tasks", "typing", "functools", "__future__", "collections", "re", "string"], "modelscope.models.nlp.mglm.tasks.superglue.dataset": ["abc", "re", "random", "copy", "utils", "glob", "csv", "pandas", "numpy", "json", "typing", "os", "data_utils", "torch", "collections", "tqdm"], "modelscope.models.nlp.mglm.tasks.seq2seq.finetune": ["finetune_glm", "megatron_util", "torch", "functools", "pretrain_glm", "collections", "tasks"], "modelscope.models.nlp.mglm.tasks.seq2seq.evaluate": ["datetime", "random", "generation_utils", "megatron_util", "torch", "rouge_score", "string"], "modelscope.models.nlp.mglm.tasks.seq2seq.dataset": ["numpy", "json", "os", "random", "data_utils", "torch", "utils", "tasks", "tqdm"], "modelscope.models.nlp.mglm.arguments": ["json", "deepspeed", "os", "torch", "argparse"], "modelscope.models.nlp.mglm.train_utils": ["torch", "apex", "deepspeed", "megatron_util"], "modelscope.models.nlp.mglm.process_grid": ["os", "statistics", "sys", "json", "glob"], "modelscope.models.nlp.mglm.run_test": ["sys", "test"], "modelscope.models.nlp.mglm.data_utils.wordpiece": ["unicodedata", "io", "os", "__future__", "collections", "logging"], "modelscope.models.nlp.mglm.data_utils.file_utils": ["io", "hashlib", "botocore", "requests", "boto3", "functools", "pathlib", "shutil", "json", "os", "urllib", "tempfile", "sys", "__future__", "logging", "tqdm"], "modelscope.models.nlp.mglm.data_utils.lazy_loader": ["time", "os", "pickle", "itertools", "torch", "numpy", "mmap"], "modelscope.models.nlp.mglm.data_utils.tokenization": ["itertools", "csv", "regex", "sentencepiece", "os", "random", "torch", "nltk", "collections"], "modelscope.models.nlp.mglm.data_utils.datasets": ["bisect", "math", "random", "nltk", "time", "csv", "itertools", "pandas", "operator", "numpy", "json", "os", "torch", "tqdm"], "modelscope.models.nlp.mglm.data_utils.corpora": ["multiprocessing", "json", "os", "random", "queue", "torch", "collections", "tqdm"], "modelscope.models.nlp.mglm.data_utils.extraction": ["os", "nltk", "json", "glob"], "modelscope.models.nlp.mglm.data_utils.sp_tokenizer": ["os"], "modelscope.models.nlp.mglm.data_utils.tokenization_gpt2": ["io", "regex", "json", "os", "functools", "sys", "__future__", "logging"], "modelscope.models.nlp.mglm.data_utils.samplers": ["os", "math", "torch", "sys", "numpy"], "modelscope.models.nlp.mglm.model.transformer": ["math", "megatron_util", "torch", "apex", "deepspeed"], "modelscope.models.nlp.mglm.model.modeling_bert": ["tarfile", "shutil", "math", "megatron_util", "json", "apex", "os", "data_utils", "tempfile", "copy", "torch", "__future__", "logging"], "modelscope.models.nlp.mglm.model.prompt": ["random", "torch"], "modelscope.models.nlp.mglm.model.downstream": ["torch"], "modelscope.models.nlp.mglm.model.distributed": ["torch", "megatron_util"], "modelscope.models.nlp.mglm.model.modeling_glm": ["torch", "megatron_util"], "modelscope.models.nlp.mglm.test.test_rel_shift": ["matplotlib", "numpy", "learning_rates", "torch"], "modelscope.models.nlp.mglm.test.test_block": ["random", "numpy", "blocklm_utils", "argparse"], "modelscope.models.nlp.veco.backbone": ["transformers"], "modelscope.models.nlp.veco.text_classification": ["transformers"], "modelscope.models.nlp.veco.configuration": ["transformers"], "modelscope.models.nlp.veco.token_classification": ["transformers", "torch"], "modelscope.models.nlp.veco.fill_mask": ["transformers"], "modelscope.models.nlp.ponet.backbone": ["math", "distutils", "transformers", "packaging", "torch"], "modelscope.models.nlp.ponet.document_segmentation": ["torch", "typing"], "modelscope.models.nlp.ponet.configuration": ["transformers"], "modelscope.models.nlp.ponet.tokenization": ["transformers", "typing"], "modelscope.models.nlp.ponet.fill_mask": ["transformers", "torch"], "modelscope.models.nlp.fid_plug.backbone": ["math", "numpy", "transformers", "dataclasses", "typing", "os", "torch", "copy"], "modelscope.models.nlp.fid_plug.configuration": ["transformers"], "modelscope.models.nlp.fid_plug.text_generation": ["os", "io", "transformers", "torch"], "modelscope.models.nlp.space.dialog_modeling": ["os", "typing"], "modelscope.models.nlp.space.dialog_state_tracking": ["torch", "transformers", "typing"], "modelscope.models.nlp.space.configuration": [], "modelscope.models.nlp.space.modules.functions": ["numpy", "torch"], "modelscope.models.nlp.space.modules.embedder": ["torch"], "modelscope.models.nlp.space.modules.multihead_attention": ["torch"], "modelscope.models.nlp.space.modules.transformer_block": ["torch"], "modelscope.models.nlp.space.modules.feedforward": ["torch"], "modelscope.models.nlp.space.model.generator": ["numpy", "math", "torch"], "modelscope.models.nlp.space.model.gen_unified_transformer": ["torch"], "modelscope.models.nlp.space.model.model_base": ["os", "torch"], "modelscope.models.nlp.space.model.tokenization_space": ["transformers"], "modelscope.models.nlp.space.model.unified_transformer": ["numpy", "torch"], "modelscope.models.nlp.space.model.intent_unified_transformer": ["torch"], "modelscope.models.nlp.space.dialog_intent_prediction": ["os", "typing"], "modelscope.models.nlp.bert.backbone": ["transformers", "math", "torch", "packaging"], "modelscope.models.nlp.bert.text_classification": [], "modelscope.models.nlp.bert.document_segmentation": ["torch", "typing"], "modelscope.models.nlp.bert.sentence_embedding": ["torch"], "modelscope.models.nlp.bert.configuration": ["collections", "transformers", "typing"], "modelscope.models.nlp.bert.siamese_uie": ["torch", "copy"], "modelscope.models.nlp.bert.word_alignment": ["torch"], "modelscope.models.nlp.bert.token_classification": [], "modelscope.models.nlp.bert.fill_mask": [], "modelscope.models.nlp.bert.text_ranking": [], "modelscope.models.nlp.use.transformer": ["math", "torch"], "modelscope.models.nlp.use.user_satisfaction_estimation": ["numpy", "transformers", "typing", "os", "torch"], "modelscope.models.nlp.gpt_moe.backbone": ["math", "transformers", "typing", "os", "torch", "addict"], "modelscope.models.nlp.gpt_moe.distributed_gpt_moe": ["torch", "transformers", "math", "megatron_util"], "modelscope.models.nlp.gpt_moe.configuration": ["transformers", "torch"], "modelscope.models.nlp.gpt_moe.checkpointing": ["os", "torch", "megatron_util"], "modelscope.models.nlp.gpt_moe.moe.utils": ["torch", "typing"], "modelscope.models.nlp.gpt_moe.moe.mappings": ["torch", "megatron_util"], "modelscope.models.nlp.gpt_moe.moe.sharded_moe": ["math", "megatron_util", "tutel", "scipy", "apex", "typing", "torch"], "modelscope.models.nlp.gpt_moe.moe.layer": ["megatron_util", "typing", "torch"], "modelscope.models.nlp.gpt_moe.moe.experts": ["torch", "copy"], "modelscope.models.nlp.gpt_moe.tokenizer": ["tokenizers"], "modelscope.models.nlp.gpt_moe.text_generation": ["transformers", "typing"], "modelscope.models.cv.face_reconstruction.utils": ["math", "array", "scipy", "numpy", "numba", "cv2", "os", "torch", "argparse", "PIL"], "modelscope.models.cv.face_reconstruction.models.facelandmark.nets.large_eyeball_net": ["torch"], "modelscope.models.cv.face_reconstruction.models.facelandmark.nets.large_base_lmks_net": ["torch"], "modelscope.models.cv.face_reconstruction.models.facelandmark.large_model_infer": ["os", "numpy", "torch", "cv2"], "modelscope.models.cv.face_reconstruction.models.facelandmark.large_base_lmks_infer": ["numpy", "torch"], "modelscope.models.cv.face_reconstruction.models.opt": [], "modelscope.models.cv.face_reconstruction.models.facerecon_model": ["os", "numpy", "torch", "cv2"], "modelscope.models.cv.face_reconstruction.models.bfm": ["os", "scipy", "numpy", "torch"], "modelscope.models.cv.face_reconstruction.models.networks": ["os", "kornia", "typing", "torch"], "modelscope.models.cv.face_reconstruction.models.losses": ["numpy", "clip", "kornia", "torch"], "modelscope.models.cv.face_reconstruction.models.nv_diffrast": ["nvdiffrast", "numpy", "typing", "torch", "warnings"], "modelscope.models.cv.panorama_depth_estimation.unifuse_model": ["os", "numpy", "torchvision", "torch"], "modelscope.models.cv.panorama_depth_estimation.networks.layers": ["numpy", "torch"], "modelscope.models.cv.panorama_depth_estimation.networks.mobilenet": ["torch"], "modelscope.models.cv.panorama_depth_estimation.networks.unifuse": ["__future__", "numpy", "collections", "torch"], "modelscope.models.cv.panorama_depth_estimation.networks.util": ["scipy", "numpy", "cv2"], "modelscope.models.cv.panorama_depth_estimation.networks.equi": ["__future__", "numpy", "collections", "torch"], "modelscope.models.cv.panorama_depth_estimation.networks.resnet": ["torch"], "modelscope.models.cv.video_panoptic_segmentation.head.semantic_fpn_wrapper": ["mmdet", "mmcv", "torch"], "modelscope.models.cv.video_panoptic_segmentation.head.kernel_head": ["mmcv", "torch"], "modelscope.models.cv.video_panoptic_segmentation.head.kernel_updator": ["mmcv", "torch"], "modelscope.models.cv.video_panoptic_segmentation.head.kernel_update_head": ["numpy", "mmdet", "mmcv", "torch"], "modelscope.models.cv.video_panoptic_segmentation.head.mask": ["torch", "__future__", "pycocotools", "numpy", "cv2"], "modelscope.models.cv.video_panoptic_segmentation.head.kernel_iter_head": ["mmdet", "torch"], "modelscope.models.cv.video_panoptic_segmentation.head.track_heads": ["numpy", "mmcv", "torch"], "modelscope.models.cv.video_panoptic_segmentation.video_k_net": ["numpy", "mmdet", "mmcv", "torch"], "modelscope.models.cv.video_panoptic_segmentation.backbone.swin_transformer": ["numpy", "mmdet", "timm", "torch"], "modelscope.models.cv.video_panoptic_segmentation.backbone.swin_checkpoint": ["os", "pkgutil", "torch", "importlib", "collections", "torchvision"], "modelscope.models.cv.video_panoptic_segmentation.track.quasi_dense_embed_tracker": ["mmdet", "mmcv", "torch"], "modelscope.models.cv.video_panoptic_segmentation.visualizer": ["numpy", "hashlib", "cv2"], "modelscope.models.cv.video_panoptic_segmentation.neck.fpn": ["mmcv", "torch"], "modelscope.models.cv.image_matching.loftr_quadtree.loftr_module.transformer": ["einops", "math", "torch", "copy", "timm"], "modelscope.models.cv.image_matching.loftr_quadtree.loftr_module.fine_preprocess": ["einops", "torch"], "modelscope.models.cv.image_matching.loftr_quadtree.loftr_module.quadtree_attention": ["timm", "torch"], "modelscope.models.cv.image_matching.loftr_quadtree.loftr_module.linear_attention": ["torch"], "modelscope.models.cv.image_matching.loftr_quadtree.utils.position_encoding": ["math", "torch"], "modelscope.models.cv.image_matching.loftr_quadtree.utils.coarse_matching": ["einops", "torch"], "modelscope.models.cv.image_matching.loftr_quadtree.utils.fine_matching": ["kornia", "math", "torch"], "modelscope.models.cv.image_matching.loftr_quadtree.loftr": ["einops", "torch"], "modelscope.models.cv.image_matching.loftr_quadtree.backbone.resnet_fpn": ["torch"], "modelscope.models.cv.image_matching.utils.misc": ["yacs"], "modelscope.models.cv.image_matching.quadtree_attention_model": ["pathlib", "numpy", "cv2", "os", "torch"], "modelscope.models.cv.image_matching.config.default": ["yacs"], "modelscope.models.cv.image_denoise.nafnet.arch_util": ["torch"], "modelscope.models.cv.image_denoise.nafnet.NAFNet_arch": ["numpy", "torch"], "modelscope.models.cv.image_denoise.nafnet_for_image_denoise": ["os", "torch", "typing"], "modelscope.models.cv.image_to_image_generation.model": ["math", "torch"], "modelscope.models.cv.image_to_image_generation.data.transforms": ["random", "math", "torchvision", "PIL"], "modelscope.models.cv.image_to_image_generation.models.autoencoder": ["math", "torch"], "modelscope.models.cv.image_to_image_generation.models.clip": ["math", "torch"], "modelscope.models.cv.image_to_image_generation.ops.diffusion": ["math", "torch"], "modelscope.models.cv.image_to_image_generation.ops.losses": ["math", "torch"], "modelscope.models.cv.face_emotion.efficient.model": ["torch"], "modelscope.models.cv.face_emotion.efficient.utils": ["math", "functools", "torch", "collections", "re"], "modelscope.models.cv.face_emotion.emotion_model": ["os", "sys", "torch"], "modelscope.models.cv.face_emotion.emotion_infer": ["torchvision", "torch", "PIL"], "modelscope.models.cv.face_emotion.face_alignment.face_align": ["numpy", "cv2", "os", "sys", "PIL"], "modelscope.models.cv.face_emotion.face_alignment.face": ["os", "tensorflow", "numpy", "cv2"], "modelscope.models.cv.image_inpainting.base": ["torch", "typing"], "modelscope.models.cv.image_inpainting.model": ["os", "torch", "typing"], "modelscope.models.cv.image_inpainting.default": ["bisect", "torch"], "modelscope.models.cv.image_inpainting.modules.perceptual": ["torchvision", "torch"], "modelscope.models.cv.image_inpainting.modules.inception": ["torchvision", "torch"], "modelscope.models.cv.image_inpainting.modules.ffc": ["numpy", "kornia", "torch"], "modelscope.models.cv.image_inpainting.modules.feature_matching": ["torch", "typing"], "modelscope.models.cv.image_inpainting.modules.adversarial": ["torch", "typing"], "modelscope.models.cv.image_inpainting.modules.ade20k.base": ["os", "torch"], "modelscope.models.cv.image_inpainting.modules.ade20k.resnet": ["os", "math", "torch"], "modelscope.models.cv.image_inpainting.modules.pix2pixhd": ["torch", "functools", "numpy", "collections", "logging"], "modelscope.models.cv.image_inpainting.refinement": ["numpy", "cv2", "kornia", "torch", "tqdm"], "modelscope.models.cv.hand_static.hand_model": ["numpy", "torchvision", "cv2", "os", "torch", "sys", "PIL"], "modelscope.models.cv.hand_static.networks": ["os", "torchvision", "torch"], "modelscope.models.cv.image_depth_estimation_bts.networks.utils": ["math", "torch"], "modelscope.models.cv.image_depth_estimation_bts.networks.encoder": ["torchvision", "torch"], "modelscope.models.cv.image_depth_estimation_bts.networks.decoder": ["torch"], "modelscope.models.cv.image_depth_estimation_bts.networks.bts_model": ["torch"], "modelscope.models.cv.image_depth_estimation_bts.depth_estimation_bts_model": ["os", "torch"], "modelscope.models.cv.language_guided_video_summarization.transformer.modules": ["torch"], "modelscope.models.cv.language_guided_video_summarization.transformer.layers": ["torch"], "modelscope.models.cv.language_guided_video_summarization.transformer.models": ["numpy", "torch"], "modelscope.models.cv.language_guided_video_summarization.transformer.sub_layers": ["numpy", "torch"], "modelscope.models.cv.language_guided_video_summarization.summarizer": ["videofeatures_clipit", "numpy", "bmt_clipit", "typing", "os", "torch", "argparse"], "modelscope.models.cv.image_human_parsing.parsing_utils": ["torch", "numpy", "copy", "PIL"], "modelscope.models.cv.image_human_parsing.m2fp_net": ["os", "torch", "typing"], "modelscope.models.cv.image_human_parsing.m2fp.m2fp_encoder": ["torch", "numpy", "typing"], "modelscope.models.cv.image_human_parsing.m2fp.m2fp_decoder": ["torch"], "modelscope.models.cv.image_human_parsing.backbone.deeplab_resnet": ["numpy", "torch"], "modelscope.models.cv.movie_scene_segmentation.model": ["einops", "math", "numpy", "torchvision", "typing", "os", "shotdetect_scenedetect_lgss", "torch", "PIL"], "modelscope.models.cv.movie_scene_segmentation.get_model": [], "modelscope.models.cv.movie_scene_segmentation.utils.save_op": ["os", "subprocess", "numpy", "cv2", "tqdm"], "modelscope.models.cv.movie_scene_segmentation.utils.trn": ["transformers", "torch"], "modelscope.models.cv.movie_scene_segmentation.utils.shot_encoder": ["torch", "typing"], "modelscope.models.cv.movie_scene_segmentation.utils.head": ["torch"], "modelscope.models.cv.image_mvs_depth_estimation.module": ["torch"], "modelscope.models.cv.image_mvs_depth_estimation.depth_filter": ["numpy", "cv2", "os", "plyfile", "PIL"], "modelscope.models.cv.image_mvs_depth_estimation.general_eval_dataset": ["os", "torch", "sys", "numpy", "re", "cv2", "PIL"], "modelscope.models.cv.image_mvs_depth_estimation.utils": ["random", "numpy", "torchvision", "torch"], "modelscope.models.cv.image_mvs_depth_estimation.cas_mvsnet": ["torch"], "modelscope.models.cv.image_mvs_depth_estimation.colmap2mvsnet": ["shutil", "multiprocessing", "numpy", "cv2", "os", "struct", "functools", "__future__", "collections"], "modelscope.models.cv.image_mvs_depth_estimation.casmvs_model": ["easydict", "numpy", "cv2", "os", "torch"], "modelscope.models.cv.image_panoptic_segmentation.panseg_model": ["os", "torch"], "modelscope.models.cv.image_panoptic_segmentation.r50_panseg_model": ["easycv"], "modelscope.models.cv.image_portrait_enhancement.losses.model_irse": ["torch"], "modelscope.models.cv.image_portrait_enhancement.losses.helpers": ["collections", "torch"], "modelscope.models.cv.image_portrait_enhancement.losses.losses": ["torch"], "modelscope.models.cv.image_portrait_enhancement.gpen": ["math", "itertools", "operator", "random", "functools", "torch"], "modelscope.models.cv.image_portrait_enhancement.eqface.fqa": ["os", "numpy", "torch", "cv2"], "modelscope.models.cv.image_portrait_enhancement.eqface.model_resnet": ["torch"], "modelscope.models.cv.image_portrait_enhancement.retinaface.utils": ["numpy", "itertools", "math", "torch"], "modelscope.models.cv.image_portrait_enhancement.retinaface.models.retinaface": ["collections", "torchvision", "torch"], "modelscope.models.cv.image_portrait_enhancement.retinaface.models.net": ["time", "torchvision", "torch"], "modelscope.models.cv.image_portrait_enhancement.retinaface.detection": ["os", "numpy", "torch", "cv2"], "modelscope.models.cv.image_portrait_enhancement.align_faces": ["numpy", "skimage", "cv2"], "modelscope.models.cv.image_portrait_enhancement.image_portrait_enhancement": ["os", "torch", "math", "typing"], "modelscope.models.cv.facial_expression_recognition.fer.facial_expression_recognition": ["numpy", "cv2", "os", "torch", "PIL"], "modelscope.models.cv.facial_expression_recognition.fer.vgg": ["torch"], "modelscope.models.cv.facial_expression_recognition.fer.transforms": ["types", "torch", "numpy", "numbers", "PIL"], "modelscope.models.cv.image_classification.mmcls_model": ["os"], "modelscope.models.cv.image_classification.utils": ["os", "math", "itertools", "torch", "numpy", "collections", "mmcls"], "modelscope.models.cv.image_classification.resnet50_cc": ["math", "torchvision", "os", "torch", "collections"], "modelscope.models.cv.image_classification.backbones.beit_v2": ["einops", "math", "itertools", "typing", "os", "torch", "functools", "warnings", "collections", "mmcv", "mmcls"], "modelscope.models.cv.image_classification.backbones.nextvit": ["einops", "math", "itertools", "typing", "os", "torch", "functools", "warnings", "collections", "mmcv", "mmcls"], "modelscope.models.cv.video_depth_estimation.geometry.pose": ["torch"], "modelscope.models.cv.video_depth_estimation.geometry.camera": ["torch", "functools"], "modelscope.models.cv.video_depth_estimation.geometry.camera_utils": ["torch"], "modelscope.models.cv.video_depth_estimation.geometry.pose_utils": ["numpy", "torch"], "modelscope.models.cv.video_depth_estimation.configs.default_config": ["os", "yacs"], "modelscope.models.cv.video_depth_estimation.networks.optim.update": ["torch"], "modelscope.models.cv.video_depth_estimation.networks.optim.extractor": ["torchvision", "torch"], "modelscope.models.cv.video_depth_estimation.networks.depth_pose.depth_pose_net": ["torch", "functools"], "modelscope.models.cv.video_depth_estimation.networks.layers.resnet.layers": ["__future__", "torch"], "modelscope.models.cv.video_depth_estimation.networks.layers.resnet.pose_decoder": ["__future__", "collections", "torch"], "modelscope.models.cv.video_depth_estimation.networks.layers.resnet.depth_decoder": ["__future__", "numpy", "collections", "torch"], "modelscope.models.cv.video_depth_estimation.networks.layers.resnet.resnet_encoder": ["__future__", "numpy", "torchvision", "torch"], "modelscope.models.cv.video_depth_estimation.models.model_wrapper": ["numpy", "random", "importlib", "torch", "collections"], "modelscope.models.cv.video_depth_estimation.models.sup_model_mf": [], "modelscope.models.cv.video_depth_estimation.models.sfm_model_mf": ["random", "torch"], "modelscope.models.cv.video_depth_estimation.models.model_utils": [], "modelscope.models.cv.video_depth_estimation.models.model_checkpoint": ["os", "numpy", "re", "torch"], "modelscope.models.cv.video_depth_estimation.dro_model": ["numpy", "cv2", "os", "torch", "glob", "tqdm"], "modelscope.models.cv.video_depth_estimation.utils.misc": ["termcolor"], "modelscope.models.cv.video_depth_estimation.utils.horovod": ["horovod"], "modelscope.models.cv.video_depth_estimation.utils.image_gt": ["torch", "functools", "cv2", "PIL"], "modelscope.models.cv.video_depth_estimation.utils.augmentations": ["numpy", "torchvision", "cv2", "random", "PIL"], "modelscope.models.cv.video_depth_estimation.utils.image": ["numpy", "cv2", "os", "torch", "functools", "PIL"], "modelscope.models.cv.video_depth_estimation.utils.config": ["datetime", "os", "yacs", "torch"], "modelscope.models.cv.video_depth_estimation.utils.depth": ["matplotlib", "numpy", "torchvision", "torch"], "modelscope.models.cv.video_depth_estimation.utils.types": ["yacs", "numpy", "torch"], "modelscope.models.cv.video_depth_estimation.utils.load": ["inspect", "os", "torch", "importlib", "warnings", "collections", "logging"], "modelscope.models.cv.face_recognition.align_face": ["numpy", "skimage", "cv2"], "modelscope.models.cv.face_recognition.torchkit.rts_backbone": ["os", "collections", "math", "torch"], "modelscope.models.cv.face_recognition.torchkit.backbone.model_irse": ["collections", "torch"], "modelscope.models.cv.face_recognition.torchkit.backbone.model_resnet": ["torch"], "modelscope.models.cv.face_recognition.torchkit.backbone.facemask_backbone": ["collections", "torch"], "modelscope.models.cv.face_recognition.torchkit.backbone.arcface_backbone": ["torch"], "modelscope.models.cv.face_recognition.torchkit.backbone.common": ["torch"], "modelscope.models.cv.video_summarization.kts.cpd_nonlin": ["numpy"], "modelscope.models.cv.video_summarization.kts.cpd_auto": ["numpy"], "modelscope.models.cv.video_summarization.summarizer": ["os", "numpy", "torch", "typing"], "modelscope.models.cv.video_summarization.pgl_sum": ["math", "torch"], "modelscope.models.cv.video_summarization.base_model": ["cv2", "numpy", "torch"], "modelscope.models.cv.image_driving_perception.utils": ["time", "numpy", "torchvision", "torch"], "modelscope.models.cv.image_driving_perception.preprocessor": ["torch", "cv2", "numpy", "typing"], "modelscope.models.cv.image_driving_perception.image_driving_percetion_model": ["numpy", "cv2", "typing", "os", "torch"], "modelscope.models.cv.pointcloud_sceneflow_estimation.pointnet2_utils": ["torch", "pointnet2_cuda", "typing"], "modelscope.models.cv.pointcloud_sceneflow_estimation.rcp_model": ["os", "numpy", "torch"], "modelscope.models.cv.pointcloud_sceneflow_estimation.sf_rcp": ["torch"], "modelscope.models.cv.pointcloud_sceneflow_estimation.common": ["math", "torch"], "modelscope.models.cv.human_reconstruction.Reconstruction": ["numpy", "torchvision", "cv2", "typing", "os", "torch", "skimage", "PIL"], "modelscope.models.cv.human_reconstruction.utils": ["os", "numpy", "torch", "mcubes"], "modelscope.models.cv.human_reconstruction.models.geometry": ["torch"], "modelscope.models.cv.human_reconstruction.models.PixToMesh": ["torch"], "modelscope.models.cv.human_reconstruction.models.Surface_head": ["torch"], "modelscope.models.cv.human_reconstruction.models.networks": ["numpy", "functools", "torch"], "modelscope.models.cv.human_reconstruction.models.Res_backbone": ["numpy", "torch"], "modelscope.models.cv.human_reconstruction.models.Embedding": ["torch"], "modelscope.models.cv.human_reconstruction.models.detectors": ["numpy", "torch"], "modelscope.models.cv.human_reconstruction.models.human_segmenter": ["tensorflow", "numpy", "cv2"], "modelscope.models.cv.table_recognition.lineless_table_process": ["torch", "numpy", "shapely", "cv2"], "modelscope.models.cv.table_recognition.modules.lore_processor": ["os", "math", "torch", "copy", "numpy"], "modelscope.models.cv.table_recognition.modules.lore_detector": ["os", "math", "torch", "copy", "numpy"], "modelscope.models.cv.table_recognition.model_lore": ["math", "numpy", "typing", "os", "torch", "copy"], "modelscope.models.cv.referring_video_object_segmentation.model": ["os", "torch", "typing"], "modelscope.models.cv.referring_video_object_segmentation.utils.backbone": ["einops", "torchvision", "torch"], "modelscope.models.cv.referring_video_object_segmentation.utils.misc": ["typing", "pickle", "torchvision", "torch"], "modelscope.models.cv.referring_video_object_segmentation.utils.segmentation": ["torch", "typing"], "modelscope.models.cv.referring_video_object_segmentation.utils.swin_transformer": ["einops", "timm", "functools", "torch", "operator", "numpy"], "modelscope.models.cv.referring_video_object_segmentation.utils.matcher": ["scipy", "torch"], "modelscope.models.cv.referring_video_object_segmentation.utils.multimodal_transformer": ["einops", "transformers", "typing", "os", "torch", "copy"], "modelscope.models.cv.referring_video_object_segmentation.utils.postprocessing": ["pycocotools", "einops", "numpy", "torch"], "modelscope.models.cv.referring_video_object_segmentation.utils.mttr": ["einops", "torch"], "modelscope.models.cv.referring_video_object_segmentation.utils.criterion": ["torch"], "modelscope.models.cv.referring_video_object_segmentation.utils.position_encoding_2d": ["math", "torch"], "modelscope.models.cv.video_super_resolution.basicvsr_net": ["torch"], "modelscope.models.cv.video_super_resolution.real_basicvsr_net": ["torch"], "modelscope.models.cv.video_super_resolution.msrresnet_lite_model": ["os", "torch", "functools", "typing"], "modelscope.models.cv.video_super_resolution.real_basicvsr_for_video_super_resolution": ["os", "torch", "typing"], "modelscope.models.cv.video_super_resolution.common": ["torch"], "modelscope.models.cv.image_depth_estimation.newcrfs_model": ["os", "numpy", "torch"], "modelscope.models.cv.image_depth_estimation.networks.uper_crf_head": ["mmcv", "torch"], "modelscope.models.cv.image_depth_estimation.networks.newcrf_utils": ["os", "pkgutil", "torch", "importlib", "warnings", "collections", "torchvision"], "modelscope.models.cv.image_depth_estimation.networks.newcrf_depth": ["torch"], "modelscope.models.cv.image_depth_estimation.networks.swin_transformer": ["numpy", "timm", "torch"], "modelscope.models.cv.image_depth_estimation.networks.newcrf_layers": ["numpy", "timm", "torch"], "modelscope.models.cv.video_human_matting.model": ["numpy", "torchvision", "typing", "os", "torch"], "modelscope.models.cv.video_human_matting.models.deep_guided_filter": ["torch"], "modelscope.models.cv.video_human_matting.models.matting": ["torch", "typing"], "modelscope.models.cv.video_human_matting.models.lraspp": ["torch"], "modelscope.models.cv.video_human_matting.models.decoder": ["torch", "typing"], "modelscope.models.cv.video_human_matting.models.effv2": ["torch"], "modelscope.models.cv.object_detection.mmdet_ms.roi_heads.mask_heads.fcn_mask_head": ["numpy", "mmdet", "torch", "warnings", "mmcv"], "modelscope.models.cv.object_detection.mmdet_ms.roi_heads.bbox_heads.convfc_bbox_head": ["mmdet", "torch"], "modelscope.models.cv.object_detection.mmdet_ms.backbones.vit": ["math", "mmdet", "torch", "functools", "timm"], "modelscope.models.cv.object_detection.mmdet_ms.necks.fpn": ["mmdet", "mmcv", "torch"], "modelscope.models.cv.object_detection.mmdet_ms.utils.convModule_norm": ["mmcv"], "modelscope.models.cv.object_detection.mmdet_ms.utils.checkpoint": ["time", "io", "pkgutil", "torchvision", "os", "tempfile", "torch", "importlib", "warnings", "collections", "mmcv"], "modelscope.models.cv.object_detection.mmdet_ms.dense_heads.anchor_head": ["mmdet"], "modelscope.models.cv.object_detection.mmdet_ms.dense_heads.rpn_head": ["mmdet", "mmcv", "torch", "copy"], "modelscope.models.cv.object_detection.yolox_pai": ["easycv"], "modelscope.models.cv.object_detection.dino": ["easycv"], "modelscope.models.cv.object_detection.mmdet_model": ["os", "numpy", "torch"], "modelscope.models.cv.stream_yolo.realtime_video_detector": ["time", "numpy", "json", "cv2", "os", "torch", "logging", "argparse", "tqdm"], "modelscope.models.cv.stream_yolo.data.data_augment": ["random", "numpy", "math", "cv2"], "modelscope.models.cv.stream_yolo.models.tal_head": ["torch"], "modelscope.models.cv.stream_yolo.models.network_blocks": ["torch"], "modelscope.models.cv.stream_yolo.models.darknet": ["torch"], "modelscope.models.cv.stream_yolo.models.streamyolo": ["torch"], "modelscope.models.cv.stream_yolo.models.dfp_pafpn": ["torch"], "modelscope.models.cv.stream_yolo.utils.boxes": ["torchvision", "torch"], "modelscope.models.cv.stream_yolo.utils.format": ["math"], "modelscope.models.cv.stream_yolo.exp.yolox_base": ["os", "random", "torch"], "modelscope.models.cv.stream_yolo.exp.base_exp": ["torch", "abc"], "modelscope.models.cv.stream_yolo.exp.default.streamyolo": ["os", "sys", "torch"], "modelscope.models.cv.stream_yolo.exp.build": ["os", "sys"], "modelscope.models.cv.image_reid_person.pass_model": ["os", "enum", "torch"], "modelscope.models.cv.image_reid_person.transreid_model": ["collections", "itertools", "functools", "torch"], "modelscope.models.cv.ocr_recognition.model": ["os", "torch"], "modelscope.models.cv.ocr_recognition.modules.convnextvit": ["torch"], "modelscope.models.cv.ocr_recognition.modules.crnn": ["torch"], "modelscope.models.cv.ocr_recognition.modules.timm_tinyc": ["math", "itertools", "torch", "copy", "functools", "collections", "logging"], "modelscope.models.cv.ocr_recognition.modules.vitstr": ["torch", "functools", "copy", "__future__", "logging"], "modelscope.models.cv.ocr_recognition.modules.convnext": ["torch"], "modelscope.models.cv.ocr_recognition.preprocessor": ["numpy", "cv2", "os", "torch", "PIL"], "modelscope.models.cv.vision_middleware.backbone": ["math", "numpy", "typing", "os", "torch", "collections"], "modelscope.models.cv.vision_middleware.model": ["os", "torch", "json", "typing"], "modelscope.models.cv.vision_middleware.head": ["numpy", "mmcv", "torch", "abc"], "modelscope.models.cv.vision_middleware.vim": ["einops", "math", "torch"], "modelscope.models.cv.image_colorization.unet.unet": ["numpy", "torch"], "modelscope.models.cv.image_colorization.unet.utils": ["enum", "functools", "torch"], "modelscope.models.cv.image_colorization.ddcolor.ddcolor": ["torch"], "modelscope.models.cv.image_colorization.ddcolor.ddcolor_for_image_colorization": ["numpy", "typing", "os", "torch", "copy"], "modelscope.models.cv.image_colorization.ddcolor.utils.unet": ["enum", "collections", "torch"], "modelscope.models.cv.image_colorization.ddcolor.utils.position_encoding": ["math", "torch"], "modelscope.models.cv.image_colorization.ddcolor.utils.transformer_utils": ["typing", "torch"], "modelscope.models.cv.image_colorization.ddcolor.utils.convnext": ["timm", "torch"], "modelscope.models.cv.image_colorization.ddcolor.utils.vgg": ["os", "collections", "torchvision", "torch"], "modelscope.models.cv.image_colorization.ddcolor.loss": ["torch"], "modelscope.models.cv.video_single_object_tracking.tracker.procontext": ["torch", "copy"], "modelscope.models.cv.video_single_object_tracking.tracker.ostrack": ["torch"], "modelscope.models.cv.video_single_object_tracking.models.ostrack.base_backbone": ["timm", "torch"], "modelscope.models.cv.video_single_object_tracking.models.ostrack.utils": ["torch"], "modelscope.models.cv.video_single_object_tracking.models.ostrack.vit_ce": ["torch", "timm", "functools"], "modelscope.models.cv.video_single_object_tracking.models.ostrack.ostrack": ["torch"], "modelscope.models.cv.video_single_object_tracking.models.layers.patch_embed": ["timm", "torch"], "modelscope.models.cv.video_single_object_tracking.models.layers.attn": ["torch"], "modelscope.models.cv.video_single_object_tracking.models.layers.head": ["torch"], "modelscope.models.cv.video_single_object_tracking.models.layers.attn_blocks": ["timm", "math", "torch"], "modelscope.models.cv.video_single_object_tracking.models.procontext.utils": ["torch"], "modelscope.models.cv.video_single_object_tracking.models.procontext.vit_ce": ["torch", "timm", "functools"], "modelscope.models.cv.video_single_object_tracking.models.procontext.procontext": ["torch"], "modelscope.models.cv.video_single_object_tracking.utils.utils": ["cv2", "math", "torch", "numpy", "typing"], "modelscope.models.cv.video_single_object_tracking.config.ostrack": ["easydict"], "modelscope.models.cv.image_defrcn_fewshot.evaluation.pascal_voc_evaluation": ["os", "tempfile", "numpy", "collections", "detectron2"], "modelscope.models.cv.image_defrcn_fewshot.evaluation.coco_evaluation": ["contextlib", "io", "tabulate", "itertools", "pycocotools", "numpy", "json", "os", "fvcore", "copy", "torch", "collections", "logging", "detectron2"], "modelscope.models.cv.image_defrcn_fewshot.evaluation.evaluator": ["datetime", "time", "torch", "logging", "detectron2"], "modelscope.models.cv.image_defrcn_fewshot.models.roi_heads": ["detectron2", "torch"], "modelscope.models.cv.image_defrcn_fewshot.models.fast_rcnn": ["numpy", "fvcore", "detectron2", "torch"], "modelscope.models.cv.image_defrcn_fewshot.models.defrcn": ["os", "torch", "detectron2", "typing"], "modelscope.models.cv.image_defrcn_fewshot.models.calibration_layer": ["torch", "sklearn", "detectron2", "cv2"], "modelscope.models.cv.image_defrcn_fewshot.models.gdl": ["torch"], "modelscope.models.cv.image_defrcn_fewshot.models.resnet": ["torchvision", "torch"], "modelscope.models.cv.image_defrcn_fewshot.utils.configuration_mapper": ["detectron2"], "modelscope.models.cv.image_defrcn_fewshot.utils.register_data": [], "modelscope.models.cv.image_defrcn_fewshot.utils.voc_register": ["os", "fvcore", "numpy", "detectron2", "xml"], "modelscope.models.cv.image_defrcn_fewshot.utils.model_surgery_op": ["os", "argparse", "torch"], "modelscope.models.cv.image_defrcn_fewshot.utils.coco_register": ["contextlib", "os", "io", "fvcore", "pycocotools", "detectron2"], "modelscope.models.cv.image_defrcn_fewshot.utils.requirements_check": ["packaging", "importlib", "sys", "importlib_metadata", "collections"], "modelscope.models.cv.image_defrcn_fewshot.defrcn_for_fewshot": ["os", "torch", "typing"], "modelscope.models.cv.object_detection_3d.depe.mmdet3d_plugin.models.detectors.petr3d": ["numpy", "mmdet", "torch", "mmcv", "mmdet3d"], "modelscope.models.cv.object_detection_3d.depe.mmdet3d_plugin.models.backbones.vovnet": ["mmdet", "collections", "mmcv", "torch"], "modelscope.models.cv.object_detection_3d.depe.mmdet3d_plugin.models.necks.cp_fpn": ["mmdet", "mmcv", "torch"], "modelscope.models.cv.object_detection_3d.depe.mmdet3d_plugin.models.utils.petr_transformer": ["math", "torch", "copy", "warnings", "mmdet", "mmcv", "typing"], "modelscope.models.cv.object_detection_3d.depe.mmdet3d_plugin.models.utils.positional_encoding": ["mmcv", "math", "torch"], "modelscope.models.cv.object_detection_3d.depe.mmdet3d_plugin.models.dense_heads.depth_net": ["mmdet", "mmcv", "torch"], "modelscope.models.cv.object_detection_3d.depe.mmdet3d_plugin.models.dense_heads.petrv2_dednhead": ["math", "numpy", "mmdet", "torch", "copy", "mmcv", "mmdet3d"], "modelscope.models.cv.object_detection_3d.depe.mmdet3d_plugin.core.bbox.assigners.hungarian_assigner_3d": ["scipy", "mmdet", "torch"], "modelscope.models.cv.object_detection_3d.depe.mmdet3d_plugin.core.bbox.coders.nms_free_coder": ["mmdet", "torch"], "modelscope.models.cv.object_detection_3d.depe.mmdet3d_plugin.core.bbox.util": ["numpy", "mmdet3d", "torch"], "modelscope.models.cv.object_detection_3d.depe.mmdet3d_plugin.core.bbox.match_costs.match_cost": ["mmdet", "torch"], "modelscope.models.cv.object_detection_3d.depe.mmdet3d_plugin.datasets.pipelines.transform_3d": ["torch", "copy", "numpy", "mmdet", "mmcv", "mmdet3d", "PIL"], "modelscope.models.cv.object_detection_3d.depe.mmdet3d_plugin.datasets.pipelines.loading": ["numpy", "mmdet", "mmcv"], "modelscope.models.cv.object_detection_3d.depe.mmdet3d_plugin.datasets.nuscenes_dataset": ["numpy", "mmdet", "mmdet3d"], "modelscope.models.cv.object_detection_3d.depe.result_vis": ["pickle", "numpy", "json", "pyquaternion", "cv2", "os", "mmdet3d", "argparse"], "modelscope.models.cv.object_detection_3d.depe.depe_detect": ["os", "numpy", "torch", "typing"], "modelscope.models.cv.image_debanding.rrdb.rrdb_image_debanding": ["os", "torch", "typing"], "modelscope.models.cv.text_driven_segmentation.lseg_blocks": ["torch"], "modelscope.models.cv.text_driven_segmentation.lseg_net": ["numpy", "torch"], "modelscope.models.cv.text_driven_segmentation.simple_tokenizer": ["os", "functools", "regex", "html", "gzip", "ftfy"], "modelscope.models.cv.text_driven_segmentation.model": ["numpy", "collections", "typing", "torch"], "modelscope.models.cv.text_driven_segmentation.lseg_base": ["torch"], "modelscope.models.cv.text_driven_segmentation.lseg_model": ["numpy", "json", "typing", "os", "torch", "PIL"], "modelscope.models.cv.text_driven_segmentation.lseg_vit": ["types", "timm", "math", "torch"], "modelscope.models.cv.text_driven_segmentation.clip": ["pkg_resources", "hashlib", "torchvision", "typing", "os", "urllib", "tqdm", "torch", "warnings", "PIL"], "modelscope.models.cv.video_deinterlace.deinterlace_arch": ["torch"], "modelscope.models.cv.video_deinterlace.UNet_for_video_deinterlace": ["os", "torch", "typing", "copy"], "modelscope.models.cv.video_deinterlace.models.deep_fourier_upsampling": ["numpy", "torch"], "modelscope.models.cv.video_deinterlace.models.utils": ["torch"], "modelscope.models.cv.video_deinterlace.models.enh": ["torch"], "modelscope.models.cv.video_deinterlace.models.fre": ["torch"], "modelscope.models.cv.video_deinterlace.models.archs": ["numpy", "torch"], "modelscope.models.cv.image_quality_assessment_degradation.image_quality_assessment_degradation": ["os", "torch", "typing"], "modelscope.models.cv.image_quality_assessment_degradation.degradation_model": ["time", "torch", "collections", "numpy", "json", "torchvision", "cv2"], "modelscope.models.cv.image_instance_segmentation.model": ["os", "torch", "typing"], "modelscope.models.cv.image_instance_segmentation.backbones.swin_transformer": ["numpy", "timm", "torch"], "modelscope.models.cv.image_instance_segmentation.cascade_mask_rcnn_swin": ["os", "collections", "torch"], "modelscope.models.cv.image_instance_segmentation.maskdino.position_encoding": ["math", "torch"], "modelscope.models.cv.image_instance_segmentation.maskdino.utils": ["math", "torch", "copy"], "modelscope.models.cv.image_instance_segmentation.maskdino.maskdino_decoder": ["torch"], "modelscope.models.cv.image_instance_segmentation.maskdino.ms_deform_attn": ["math", "torch", "warnings", "__future__", "mmcv"], "modelscope.models.cv.image_instance_segmentation.maskdino.dino_decoder": ["torch", "typing"], "modelscope.models.cv.image_instance_segmentation.maskdino.maskdino_encoder": ["torch", "numpy", "typing"], "modelscope.models.cv.image_instance_segmentation.postprocess_utils": ["itertools", "pycocotools", "numpy", "cv2", "torch"], "modelscope.models.cv.image_instance_segmentation.maskdino_model": ["os", "torch", "typing"], "modelscope.models.cv.image_instance_segmentation.datasets.transforms": ["os", "numpy"], "modelscope.models.cv.image_instance_segmentation.maskdino_swin": ["os", "torch"], "modelscope.models.cv.image_quality_assessment_man.image_quality_assessment_man": ["os", "torch", "typing"], "modelscope.models.cv.image_quality_assessment_man.swin": ["einops", "math", "itertools", "torch", "warnings", "collections"], "modelscope.models.cv.image_quality_assessment_man.maniqa": ["einops", "timm", "torch"], "modelscope.models.cv.controllable_image_generation.annotator.mlsd.mbv2_mlsd_large": ["os", "sys", "torch"], "modelscope.models.cv.controllable_image_generation.annotator.mlsd.utils": ["os", "numpy", "torch", "cv2"], "modelscope.models.cv.controllable_image_generation.annotator.midas.utils": ["torch", "sys", "numpy", "re", "cv2"], "modelscope.models.cv.controllable_image_generation.annotator.midas.midas.blocks": ["torch"], "modelscope.models.cv.controllable_image_generation.annotator.midas.midas.midas_net_custom": ["torch"], "modelscope.models.cv.controllable_image_generation.annotator.midas.midas.dpt_depth": ["torch"], "modelscope.models.cv.controllable_image_generation.annotator.midas.midas.midas_net": ["torch"], "modelscope.models.cv.controllable_image_generation.annotator.midas.midas.base_model": ["torch"], "modelscope.models.cv.controllable_image_generation.annotator.midas.midas.transforms": ["numpy", "math", "cv2"], "modelscope.models.cv.controllable_image_generation.annotator.midas.midas.vit": ["types", "timm", "math", "torch"], "modelscope.models.cv.controllable_image_generation.annotator.midas.api": ["os", "torch", "torchvision", "cv2"], "modelscope.models.cv.controllable_image_generation.annotator.openpose.model": ["collections", "torch"], "modelscope.models.cv.controllable_image_generation.annotator.openpose.util": ["matplotlib", "numpy", "math", "cv2"], "modelscope.models.cv.controllable_image_generation.annotator.openpose.hand": ["time", "math", "matplotlib", "scipy", "numpy", "json", "cv2", "torch", "skimage"], "modelscope.models.cv.controllable_image_generation.annotator.openpose.body": ["time", "math", "matplotlib", "scipy", "numpy", "torchvision", "cv2", "torch"], "modelscope.models.cv.controllable_image_generation.annotator.annotator": ["einops", "numpy", "cv2", "os", "torch", "mmcv", "mmseg"], "modelscope.models.cv.controllable_image_generation.controlnet": ["einops", "control_ldm", "math", "numpy", "cv2", "typing", "os", "random", "tempfile", "torch", "sys", "PIL"], "modelscope.models.cv.motion_generation.model": [], "modelscope.models.cv.motion_generation.modules.rotation2xyz": ["torch"], "modelscope.models.cv.motion_generation.modules.cfg_sampler": ["torch", "copy"], "modelscope.models.cv.motion_generation.modules.smpl": ["contextlib", "os", "smplx", "torch", "numpy"], "modelscope.models.cv.motion_generation.modules.respace": ["numpy", "torch"], "modelscope.models.cv.motion_generation.modules.gaussian_diffusion": ["math", "torch", "copy", "enum", "numpy"], "modelscope.models.cv.motion_generation.modules.mdm": ["numpy", "clip", "torch"], "modelscope.models.cv.crowd_counting.cc_model": ["os", "torch", "typing"], "modelscope.models.cv.crowd_counting.hrnet_aspp_relu": ["numpy", "os", "functools", "torch", "logging"], "modelscope.models.cv.salient_detection.models.modules": ["torch"], "modelscope.models.cv.salient_detection.models.senet": ["torch"], "modelscope.models.cv.salient_detection.models.utils": ["torch"], "modelscope.models.cv.salient_detection.models.u2net": ["torch"], "modelscope.models.cv.salient_detection.models.backbone.Res2Net_v1b": ["math", "torch"], "modelscope.models.cv.salient_detection.salient_model": ["torchvision", "cv2", "os", "torch", "PIL"], "modelscope.models.cv.video_streaming_perception.longshortnet.longshortnet": ["time", "numpy", "json", "cv2", "os", "torch", "logging", "argparse", "tqdm"], "modelscope.models.cv.video_streaming_perception.longshortnet.models.dfp_pafpn_long": ["collections", "torch"], "modelscope.models.cv.video_streaming_perception.longshortnet.models.longshort": ["torch"], "modelscope.models.cv.video_streaming_perception.longshortnet.models.dfp_pafpn_short": ["collections", "torch"], "modelscope.models.cv.video_streaming_perception.longshortnet.models.longshort_backbone_neck": ["torch"], "modelscope.models.cv.video_streaming_perception.longshortnet.exp.longshortnet_base": [], "modelscope.models.cv.video_object_segmentation.mod_resnet": ["collections", "math", "torch"], "modelscope.models.cv.video_object_segmentation.inference_core": ["torch"], "modelscope.models.cv.video_object_segmentation.modules": ["torchvision", "torch"], "modelscope.models.cv.video_object_segmentation.aggregate": ["torch"], "modelscope.models.cv.video_object_segmentation.eval_network": ["torch"], "modelscope.models.cv.video_object_segmentation.model": ["os", "torch", "typing"], "modelscope.models.cv.video_object_segmentation.inference_memory_bank": ["math", "torch"], "modelscope.models.cv.video_object_segmentation.cbam": ["torch"], "modelscope.models.cv.video_object_segmentation.network": ["math", "torch"], "modelscope.models.cv.tinynas_detection.tinynas_damoyolo": [], "modelscope.models.cv.tinynas_detection.detector": ["os", "pickle", "torchvision", "torch"], "modelscope.models.cv.tinynas_detection.utils": ["os", "easydict", "shutil", "tempfile", "importlib", "sys"], "modelscope.models.cv.tinynas_detection.tinynas_detector": [], "modelscope.models.cv.tinynas_detection.damo.augmentations.box_level_augs.gaussian_maps": ["math", "torch"], "modelscope.models.cv.tinynas_detection.damo.augmentations.box_level_augs.color_augs": ["random", "torch"], "modelscope.models.cv.tinynas_detection.damo.augmentations.box_level_augs.box_level_augs": ["random", "numpy"], "modelscope.models.cv.tinynas_detection.damo.augmentations.box_level_augs.geometric_augs": ["random", "torchvision", "torch", "copy"], "modelscope.models.cv.tinynas_detection.damo.augmentations.scale_aware_aug": ["copy"], "modelscope.models.cv.tinynas_detection.damo.apis.detector_inference": ["os", "torch", "tqdm"], "modelscope.models.cv.tinynas_detection.damo.apis.detector_evaluater": ["os", "torch"], "modelscope.models.cv.tinynas_detection.damo.detectors.detector": ["torch"], "modelscope.models.cv.tinynas_detection.damo.structures.boxlist_ops": ["torch"], "modelscope.models.cv.tinynas_detection.damo.structures.image_list": ["__future__", "torch"], "modelscope.models.cv.tinynas_detection.damo.structures.bounding_box": ["torch"], "modelscope.models.cv.tinynas_detection.damo.base_models.losses.distill_loss": ["torch"], "modelscope.models.cv.tinynas_detection.damo.base_models.losses.gfocal_loss": ["functools", "torch"], "modelscope.models.cv.tinynas_detection.damo.base_models.backbones.tinynas_res": ["torch"], "modelscope.models.cv.tinynas_detection.damo.base_models.backbones.tinynas_csp": ["torch"], "modelscope.models.cv.tinynas_detection.damo.base_models.backbones.darknet": ["torch"], "modelscope.models.cv.tinynas_detection.damo.base_models.heads.gfocal_v2_tiny": ["numpy", "functools", "torch"], "modelscope.models.cv.tinynas_detection.damo.base_models.heads.zero_head": ["torch"], "modelscope.models.cv.tinynas_detection.damo.base_models.necks.giraffe_config": ["collections", "networkx"], "modelscope.models.cv.tinynas_detection.damo.base_models.necks.giraffe_fpn": ["math", "numpy", "typing", "torch", "functools", "timm", "collections"], "modelscope.models.cv.tinynas_detection.damo.base_models.necks.giraffe_fpn_btn": ["torch"], "modelscope.models.cv.tinynas_detection.damo.base_models.core.base_ops": ["math", "torch"], "modelscope.models.cv.tinynas_detection.damo.base_models.core.repvgg_block": ["numpy", "torch"], "modelscope.models.cv.tinynas_detection.damo.base_models.core.utils": ["torch", "functools"], "modelscope.models.cv.tinynas_detection.damo.base_models.core.weight_init": ["numpy", "torch"], "modelscope.models.cv.tinynas_detection.damo.base_models.core.neck_ops": ["numpy", "torch"], "modelscope.models.cv.tinynas_detection.damo.base_models.core.ota_assigner": ["warnings", "torch"], "modelscope.models.cv.tinynas_detection.damo.base_models.core.ops": ["numpy", "torch"], "modelscope.models.cv.tinynas_detection.damo.utils.boxes": ["numpy", "torchvision", "torch"], "modelscope.models.cv.tinynas_detection.damo.utils.model_utils": ["time", "thop", "math", "torch", "copy"], "modelscope.models.cv.tinynas_detection.damo.utils.scheduler": ["math"], "modelscope.models.cv.cmdssl_video_embedding.resnet3d": ["torch"], "modelscope.models.cv.cmdssl_video_embedding.c3d": ["torch"], "modelscope.models.cv.cmdssl_video_embedding.resnet2p1d": ["torch"], "modelscope.models.cv.facial_landmark_confidence.flc.manual_landmark_net": ["math", "torch"], "modelscope.models.cv.facial_landmark_confidence.flc.facial_landmark_confidence": ["numpy", "cv2", "os", "torch", "PIL"], "modelscope.models.cv.image_face_fusion.facegan.gan_wrap": ["numpy", "torchvision", "cv2", "os", "torch", "PIL"], "modelscope.models.cv.image_face_fusion.facegan.model": ["random", "math", "torch"], "modelscope.models.cv.image_face_fusion.facegan.op.fused_act": ["torch"], "modelscope.models.cv.image_face_fusion.facegan.op.upfirdn2d": ["collections", "torch"], "modelscope.models.cv.image_face_fusion.facegan.op.conv2d_gradfix": ["contextlib", "torch", "warnings"], "modelscope.models.cv.image_face_fusion.network.model_irse": ["collections", "torch"], "modelscope.models.cv.image_face_fusion.network.aad_layer": ["torch"], "modelscope.models.cv.image_face_fusion.network.facerecon_model": ["os", "torch", "typing"], "modelscope.models.cv.image_face_fusion.network.bfm": ["os", "scipy", "numpy", "torch"], "modelscope.models.cv.image_face_fusion.network.dense_motion": ["torch"], "modelscope.models.cv.image_face_fusion.network.aei_flow_net": ["torch"], "modelscope.models.cv.image_face_fusion.network.ops": ["torch"], "modelscope.models.cv.image_face_fusion.image_face_fusion": ["numpy", "torchvision", "typing", "cv2", "os", "torch", "collections", "PIL"], "modelscope.models.cv.image_face_fusion.facelib.align_trans": ["numpy", "cv2"], "modelscope.models.cv.image_face_fusion.facelib.matlab_cp2tform": ["numpy"], "modelscope.models.cv.image_paintbyexample.model": ["omegaconf", "typing", "os", "paint_ldm", "torch"], "modelscope.models.cv.image_deblur.nafnet_for_image_deblur": ["os", "torch", "typing"], "modelscope.models.cv.image_restoration.demoire_models.nets": ["torch"], "modelscope.models.cv.image_restoration.image_restoration_model": ["os", "numpy", "torch", "cv2"], "modelscope.models.cv.ocr_detection.model": ["os", "numpy", "torch", "typing"], "modelscope.models.cv.ocr_detection.utils": ["numpy", "pyclipper", "cv2", "shapely"], "modelscope.models.cv.ocr_detection.modules.seg_detector_loss": ["sys", "torch"], "modelscope.models.cv.ocr_detection.modules.dbnet": ["os", "math", "torch", "sys", "collections"], "modelscope.models.cv.ocr_detection.preprocessor": ["math", "numpy", "cv2", "typing", "os", "torch", "PIL"], "modelscope.models.cv.vop_retrieval.backbone": ["numpy", "hashlib", "typing", "os", "urllib", "torch", "warnings", "collections", "tqdm"], "modelscope.models.cv.vop_retrieval.model": ["os", "torch"], "modelscope.models.cv.vop_retrieval.basic_utils": ["shutil", "pickle", "numpy", "torchvision", "cv2", "os", "random", "zipfile", "torch", "ujson", "collections", "PIL"], "modelscope.models.cv.vop_retrieval.model_se": ["os", "torch"], "modelscope.models.cv.vop_retrieval.tokenization_clip": ["os", "torch", "functools", "regex", "html", "gzip", "ftfy"], "modelscope.models.cv.cartoon.utils": ["os", "random", "tensorflow", "numpy", "cv2"], "modelscope.models.cv.cartoon.facelib.face_detector": ["time", "tensorflow", "numpy", "cv2"], "modelscope.models.cv.cartoon.facelib.face_landmark": ["tensorflow", "numpy", "cv2"], "modelscope.models.cv.cartoon.facelib.facer": ["time", "numpy", "cv2"], "modelscope.models.cv.cartoon.facelib.config": ["os", "numpy", "easydict"], "modelscope.models.cv.cartoon.facelib.LK.lk": ["numpy"], "modelscope.models.cv.cartoon.mtcnn_pytorch.src.align_trans": ["numpy", "cv2"], "modelscope.models.cv.cartoon.mtcnn_pytorch.src.matlab_cp2tform": ["numpy"], "modelscope.models.cv.cartoon.model_tf": ["tensorflow", "typing"], "modelscope.models.cv.cartoon.network": ["tensorflow"], "modelscope.models.cv.cartoon.loss": ["scipy", "numpy", "os", "joblib", "tensorflow", "skimage"], "modelscope.models.cv.tinynas_classfication.super_res_idwexkx": ["uuid", "torch"], "modelscope.models.cv.tinynas_classfication.super_blocks": ["uuid", "torch"], "modelscope.models.cv.tinynas_classfication.global_utils": [], "modelscope.models.cv.tinynas_classfication.super_res_kxkx": ["uuid", "torch"], "modelscope.models.cv.tinynas_classfication.master_net": ["torch"], "modelscope.models.cv.tinynas_classfication.plain_net_utils": ["torch"], "modelscope.models.cv.tinynas_classfication.model_zoo": [], "modelscope.models.cv.tinynas_classfication.basic_blocks": ["uuid", "numpy", "torch"], "modelscope.models.cv.tinynas_classfication.super_res_k1kxk1": ["uuid", "torch"], "modelscope.models.cv.face_attribute_recognition.fair_face.face_attribute_recognition": ["numpy", "torchvision", "cv2", "os", "torch", "PIL"], "modelscope.models.cv.nerf_recon_acc.nerf_recon_acc": ["time", "numpy", "cv2", "os", "torch", "glob", "tqdm"], "modelscope.models.cv.nerf_recon_acc.network.segmenter": ["tensorflow", "numpy"], "modelscope.models.cv.nerf_recon_acc.network.utils": ["gc", "torch", "collections", "numpy", "tinycudann", "mcubes"], "modelscope.models.cv.nerf_recon_acc.network.nerf": ["nerfacc", "numpy", "tinycudann", "torch"], "modelscope.models.cv.nerf_recon_acc.dataloader.read_write_model": ["os", "struct", "numpy", "collections", "argparse"], "modelscope.models.cv.nerf_recon_acc.dataloader.nerf_dataset": ["math", "numpy", "json", "torchvision", "os", "torch", "PIL"], "modelscope.models.cv.nerf_recon_acc.nerf_preprocess": ["subprocess", "numpy", "cv2", "typing", "os", "tensorflow", "glob"], "modelscope.models.cv.image_body_reshaping.model": ["torch"], "modelscope.models.cv.image_body_reshaping.slim_utils": ["os", "random", "math", "torch", "numpy", "numba", "cv2"], "modelscope.models.cv.image_body_reshaping.pose_estimator.model": ["collections", "torch"], "modelscope.models.cv.image_body_reshaping.pose_estimator.util": ["numpy"], "modelscope.models.cv.image_body_reshaping.pose_estimator.body": ["math", "scipy", "numpy", "cv2", "torch"], "modelscope.models.cv.image_body_reshaping.image_body_reshaping": ["numpy", "cv2", "typing", "os", "torch"], "modelscope.models.cv.image_body_reshaping.person_info": ["torch", "numpy", "cv2", "copy"], "modelscope.models.cv.body_2d_keypoints.hrnet_basic_modules": ["torch"], "modelscope.models.cv.body_2d_keypoints.hrnet_v2": ["os", "numpy", "torch"], "modelscope.models.cv.body_2d_keypoints.w48": [], "modelscope.models.cv.image_binary_quant_classification.binary_quant_model": ["os", "collections", "torch"], "modelscope.models.cv.image_binary_quant_classification.bnext": ["numpy", "torch"], "modelscope.models.cv.video_inpainting.inpainting_model": ["torch", "numpy", "math", "torchvision"], "modelscope.models.cv.video_inpainting.inpainting": ["time", "os", "torch", "numpy", "torchvision", "cv2", "PIL"], "modelscope.models.cv.robust_image_classification.easyrobust_model": ["os", "torch"], "modelscope.models.cv.face_human_hand_detection.ghost_pan": ["math", "torch"], "modelscope.models.cv.face_human_hand_detection.det_infer": ["torch", "numpy", "cv2"], "modelscope.models.cv.face_human_hand_detection.utils": ["torch"], "modelscope.models.cv.face_human_hand_detection.shufflenetv2": ["torch"], "modelscope.models.cv.face_human_hand_detection.one_stage_detector": ["torch"], "modelscope.models.cv.face_human_hand_detection.nanodet_plus_head": ["math", "numpy", "torchvision", "cv2", "torch"], "modelscope.models.cv.bad_image_detecting.bad_image_detecting": ["numpy", "torchvision", "typing", "os", "torch"], "modelscope.models.cv.face_detection.mtcnn.models.box_utils": ["numpy", "PIL"], "modelscope.models.cv.face_detection.mtcnn.models.detector": ["os", "numpy", "torch", "PIL"], "modelscope.models.cv.face_detection.mtcnn.models.get_nets": ["numpy", "collections", "torch"], "modelscope.models.cv.face_detection.mtcnn.models.first_stage": ["numpy", "math", "torch", "PIL"], "modelscope.models.cv.face_detection.mogface.models.utils": ["numpy", "itertools", "math", "torch"], "modelscope.models.cv.face_detection.mogface.models.mogprednet": ["math", "torch"], "modelscope.models.cv.face_detection.mogface.models.mogface": ["torch"], "modelscope.models.cv.face_detection.mogface.models.detectors": ["os", "numpy", "torch", "cv2"], "modelscope.models.cv.face_detection.mogface.models.resnet": ["torch"], "modelscope.models.cv.face_detection.scrfd.mmdet_patch.models.detectors.base": ["torch", "abc", "collections", "numpy", "mmdet", "mmcv"], "modelscope.models.cv.face_detection.scrfd.mmdet_patch.models.detectors.single_stage": ["mmdet", "torch"], "modelscope.models.cv.face_detection.scrfd.mmdet_patch.models.detectors.scrfd": ["mmdet", "torch"], "modelscope.models.cv.face_detection.scrfd.mmdet_patch.models.detectors.tinymog": ["mmdet", "torch"], "modelscope.models.cv.face_detection.scrfd.mmdet_patch.models.backbones.mobilenet": ["mmdet", "mmcv", "torch"], "modelscope.models.cv.face_detection.scrfd.mmdet_patch.models.backbones.resnet": ["mmdet", "mmcv", "torch"], "modelscope.models.cv.face_detection.scrfd.mmdet_patch.models.dense_heads.scrfd_head": ["numpy", "mmdet", "mmcv", "torch"], "modelscope.models.cv.face_detection.scrfd.mmdet_patch.core.bbox.transforms": ["numpy", "torch"], "modelscope.models.cv.face_detection.scrfd.mmdet_patch.core.post_processing.bbox_nms": ["torch"], "modelscope.models.cv.face_detection.scrfd.mmdet_patch.datasets.retinaface": ["numpy", "mmdet"], "modelscope.models.cv.face_detection.scrfd.mmdet_patch.datasets.pipelines.formating": ["numpy", "mmdet", "mmcv", "torch"], "modelscope.models.cv.face_detection.scrfd.mmdet_patch.datasets.pipelines.loading": ["os", "mmdet", "numpy", "pycocotools"], "modelscope.models.cv.face_detection.scrfd.mmdet_patch.datasets.pipelines.auto_augment": ["copy", "numpy", "mmdet", "mmcv", "cv2"], "modelscope.models.cv.face_detection.scrfd.mmdet_patch.datasets.pipelines.transforms": ["numpy", "mmdet", "mmcv"], "modelscope.models.cv.face_detection.scrfd.scrfd_detect": ["numpy", "typing", "os", "torch", "copy"], "modelscope.models.cv.face_detection.scrfd.preprocessor": ["numpy", "typing", "PIL"], "modelscope.models.cv.face_detection.scrfd.tinymog_detect": ["os", "torch", "typing", "copy"], "modelscope.models.cv.face_detection.ulfd_slim.vision.box_utils": ["math", "torch"], "modelscope.models.cv.face_detection.ulfd_slim.vision.mb_tiny": ["torch"], "modelscope.models.cv.face_detection.ulfd_slim.vision.ssd.mb_tiny_fd": ["torch"], "modelscope.models.cv.face_detection.ulfd_slim.vision.ssd.ssd": ["torch", "numpy", "collections", "typing"], "modelscope.models.cv.face_detection.ulfd_slim.vision.ssd.data_preprocessing": [], "modelscope.models.cv.face_detection.ulfd_slim.vision.ssd.predictor": ["torch"], "modelscope.models.cv.face_detection.ulfd_slim.vision.ssd.fd_config": ["numpy"], "modelscope.models.cv.face_detection.ulfd_slim.vision.transforms": ["numpy", "types", "torch", "cv2"], "modelscope.models.cv.face_detection.ulfd_slim.detection": ["os", "numpy", "torch", "cv2"], "modelscope.models.cv.face_detection.peppa_pig_face.face_detector": ["tensorflow", "numpy", "cv2"], "modelscope.models.cv.face_detection.peppa_pig_face.face_landmark": ["tensorflow", "numpy", "cv2"], "modelscope.models.cv.face_detection.peppa_pig_face.facer": ["numpy", "cv2"], "modelscope.models.cv.face_detection.peppa_pig_face.LK.lk": ["numpy"], "modelscope.models.cv.face_detection.retinaface.utils": ["numpy", "itertools", "math", "torch"], "modelscope.models.cv.face_detection.retinaface.models.retinaface": ["collections", "torchvision", "torch"], "modelscope.models.cv.face_detection.retinaface.models.net": ["time", "torchvision", "torch"], "modelscope.models.cv.face_detection.retinaface.detection": ["torch", "numpy", "cv2"], "modelscope.models.cv.video_frame_interpolation.interp_model.transformer_layers": ["math", "torch", "functools", "sys", "timm"], "modelscope.models.cv.video_frame_interpolation.interp_model.flow_reversal": ["torch"], "modelscope.models.cv.video_frame_interpolation.interp_model.UNet": ["torch"], "modelscope.models.cv.video_frame_interpolation.interp_model.IFNet_swin": ["numpy", "timm", "torch"], "modelscope.models.cv.video_frame_interpolation.interp_model.refinenet_arch": ["numpy", "torch"], "modelscope.models.cv.video_frame_interpolation.flow_model.raft": ["numpy", "torch"], "modelscope.models.cv.video_frame_interpolation.flow_model.update": ["torch"], "modelscope.models.cv.video_frame_interpolation.flow_model.extractor": ["torch"], "modelscope.models.cv.video_frame_interpolation.flow_model.corr": ["torch"], "modelscope.models.cv.video_frame_interpolation.VFINet_for_video_frame_interpolation": ["os", "torch", "typing", "copy"], "modelscope.models.cv.video_frame_interpolation.utils.utils": ["scipy", "numpy", "torch"], "modelscope.models.cv.video_frame_interpolation.utils.scene_change_detection": ["numpy", "torch"], "modelscope.models.cv.video_frame_interpolation.VFINet_arch": ["torch"], "modelscope.models.cv.action_detection.action_detection_onnx": ["shutil", "uuid", "subprocess", "numpy", "cv2", "os", "onnxruntime", "urllib", "tempfile"], "modelscope.models.cv.action_detection.modules.action_detection_pytorch": ["typing", "fvcore", "torch", "logging", "detectron2"], "modelscope.models.cv.action_detection.modules.resnet": ["detectron2", "torch"], "modelscope.models.cv.video_multi_object_tracking.tracker.multitracker": ["numpy", "collections", "torch"], "modelscope.models.cv.video_multi_object_tracking.tracker.basetrack": ["numpy", "collections"], "modelscope.models.cv.video_multi_object_tracking.tracker.matching": ["lap", "scipy", "numpy"], "modelscope.models.cv.video_multi_object_tracking.models.model": ["torch"], "modelscope.models.cv.video_multi_object_tracking.models.decode": ["torch"], "modelscope.models.cv.video_multi_object_tracking.models.yolo": ["math", "torch", "copy"], "modelscope.models.cv.video_multi_object_tracking.models.common": ["torch"], "modelscope.models.cv.video_multi_object_tracking.utils.utils": ["cv2", "numpy", "torch"], "modelscope.models.cv.video_multi_object_tracking.utils.image": ["numpy", "cv2"], "modelscope.models.cv.video_multi_object_tracking.utils.kalman_filter": ["scipy", "numpy"], "modelscope.models.cv.video_multi_object_tracking.utils.visualization": ["numpy", "cv2"], "modelscope.models.cv.face_generation.op.fused_act": ["os", "torch"], "modelscope.models.cv.face_generation.op.upfirdn2d": ["os", "collections", "torch"], "modelscope.models.cv.face_generation.op.conv2d_gradfix": ["contextlib", "torch", "warnings"], "modelscope.models.cv.face_generation.stylegan2": ["math", "operator", "random", "functools", "torch"], "modelscope.models.cv.vidt.backbone": ["os", "math", "torch", "timm", "numpy"], "modelscope.models.cv.vidt.fpn_fusion": ["torch"], "modelscope.models.cv.vidt.deformable_transformer": ["math", "torch", "copy", "warnings", "timm"], "modelscope.models.cv.vidt.model": ["os", "torch"], "modelscope.models.cv.vidt.head": ["math", "torch", "copy"], "modelscope.models.cv.body_3d_keypoints.hdformer.backbone": ["torch"], "modelscope.models.cv.body_3d_keypoints.hdformer.block": ["einops", "math", "torch"], "modelscope.models.cv.body_3d_keypoints.hdformer.skeleton": ["numpy"], "modelscope.models.cv.body_3d_keypoints.hdformer.hdformer_detector": ["os", "numpy", "torch", "typing"], "modelscope.models.cv.body_3d_keypoints.hdformer.directed_graph": ["sys", "numpy", "typing"], "modelscope.models.cv.body_3d_keypoints.hdformer.hdformer": ["torch"], "modelscope.models.cv.body_3d_keypoints.cannonical_pose.body_3d_pose": ["numpy", "typing", "os", "torch", "logging"], "modelscope.models.cv.body_3d_keypoints.cannonical_pose.canonical_pose_modules": ["torch"], "modelscope.models.cv.virual_tryon.sdafnet": ["random", "numpy", "torch"], "modelscope.models.cv.vision_efficient_tuning.backbone": ["torch", "functools"], "modelscope.models.cv.vision_efficient_tuning.vision_efficient_tuning": ["os", "collections", "torch"], "modelscope.models.cv.vision_efficient_tuning.timm_weight_init": ["warnings", "math", "torch"], "modelscope.models.cv.vision_efficient_tuning.model": ["torch", "typing"], "modelscope.models.cv.vision_efficient_tuning.petl": ["torchvision", "collections", "math", "torch"], "modelscope.models.cv.vision_efficient_tuning.head": ["torch"], "modelscope.models.cv.vision_efficient_tuning.timm_vision_transformer": ["math", "itertools", "torch", "functools", "collections", "logging"], "modelscope.models.cv.vision_efficient_tuning.timm_helpers": ["typing", "itertools", "math", "torch"], "modelscope.models.cv.image_quality_assessment_mos.censeo_ivqa_model": ["torch"], "modelscope.models.cv.image_quality_assessment_mos.image_quality_assessment_mos": ["os", "torch", "typing"], "modelscope.models.cv.image_quality_assessment_mos.backbones.resnet": ["os", "torch"], "modelscope.models.cv.image_quality_assessment_mos.heads.simple_head": ["torch"], "modelscope.models.cv.open_vocabulary_detection_vild.vild": ["scipy", "numpy", "typing", "os", "torch", "tensorflow", "clip"], "modelscope.models.cv.human_wholebody_keypoint.human_wholebody_keypoint": ["easycv"], "modelscope.models.cv.image_semantic_segmentation.ddpm_segmentation_model": ["os", "torch", "ddpm_guided_diffusion", "typing"], "modelscope.models.cv.image_semantic_segmentation.ddpm_seg.feature_extractors": ["torch", "typing"], "modelscope.models.cv.image_semantic_segmentation.ddpm_seg.utils": ["random", "numpy", "torch", "PIL"], "modelscope.models.cv.image_semantic_segmentation.ddpm_seg.pixel_classifier": ["numpy", "os", "torch", "collections", "PIL"], "modelscope.models.cv.image_semantic_segmentation.ddpm_seg.data_util": [], "modelscope.models.cv.image_semantic_segmentation.vit_adapter.models.decode_heads.base_decode_head": ["mmdet", "mmcv", "torch", "abc"], "modelscope.models.cv.image_semantic_segmentation.vit_adapter.models.decode_heads.mask2former_head_from_mmseg": ["mmdet", "mmcv", "torch", "copy"], "modelscope.models.cv.image_semantic_segmentation.vit_adapter.models.segmentors.encoder_decoder_mask2former": ["mmdet", "torch"], "modelscope.models.cv.image_semantic_segmentation.vit_adapter.models.segmentors.base_segmentor": ["torch", "abc", "warnings", "numpy", "collections", "mmcv"], "modelscope.models.cv.image_semantic_segmentation.vit_adapter.models.backbone.adapter_modules": ["functools", "torch", "timm", "mmdet", "logging"], "modelscope.models.cv.image_semantic_segmentation.vit_adapter.models.backbone.beit_adapter": ["math", "mmdet", "torch", "timm", "logging"], "modelscope.models.cv.image_semantic_segmentation.vit_adapter.models.backbone.base.beit": ["math", "functools", "torch", "timm", "mmdet", "mmcv"], "modelscope.models.cv.image_semantic_segmentation.vit_adapter.utils.builder": ["mmcv"], "modelscope.models.cv.image_semantic_segmentation.vit_adapter.utils.data_process_func": ["mmdet", "mmcv"], "modelscope.models.cv.image_semantic_segmentation.vit_adapter.utils.seg_func": ["warnings", "torch"], "modelscope.models.cv.image_semantic_segmentation.pan_merge.base_panoptic_fusion_head": ["mmdet", "mmcv", "abc"], "modelscope.models.cv.image_semantic_segmentation.pan_merge.maskformer_semantic_head": ["mmdet", "torch"], "modelscope.models.cv.image_semantic_segmentation.semantic_seg_model": ["os", "numpy", "torch"], "modelscope.models.cv.image_semantic_segmentation.segformer": ["easycv"], "modelscope.models.cv.shop_segmentation.neck_fpn": ["timm", "mmcv", "torch"], "modelscope.models.cv.shop_segmentation.head_fpn": ["numpy", "mmcv", "timm", "torch"], "modelscope.models.cv.shop_segmentation.utils": ["regex", "html", "gzip", "typing", "os", "torch", "functools", "ftfy"], "modelscope.models.cv.shop_segmentation.shop_seg_model": ["numpy", "json", "typing", "os", "torch", "PIL"], "modelscope.models.cv.shop_segmentation.shop_seg_base": ["torch"], "modelscope.models.cv.shop_segmentation.models": ["timm", "collections", "math", "torch"], "modelscope.models.cv.shop_segmentation.common": ["warnings", "torch"], "modelscope.models.cv.super_resolution.ecbsr_model": ["os", "torch", "typing"], "modelscope.models.cv.super_resolution.rrdbnet_arch": ["torch"], "modelscope.models.cv.super_resolution.ecb": ["torch"], "modelscope.models.cv.super_resolution.arch_util": ["math", "itertools", "torch", "warnings", "collections", "torchvision"], "modelscope.models.cv.indoor_layout_estimation.networks.misc.fourier": ["scipy", "numpy", "PIL"], "modelscope.models.cv.indoor_layout_estimation.networks.misc.post_proc": ["scipy", "numpy", "sklearn"], "modelscope.models.cv.indoor_layout_estimation.networks.misc.panostretch": ["scipy", "numpy", "functools"], "modelscope.models.cv.indoor_layout_estimation.networks.utils": ["numpy", "torch"], "modelscope.models.cv.indoor_layout_estimation.networks.panovit": ["numpy", "torch"], "modelscope.models.cv.indoor_layout_estimation.networks.backbone.resnet_DA": ["torchvision", "torch"], "modelscope.models.cv.indoor_layout_estimation.networks.backbone.vit_horizon_pry_image": ["numpy", "timm", "torch"], "modelscope.models.cv.indoor_layout_estimation.networks.modality.layout": ["shapely", "math", "scipy", "numpy", "torch"], "modelscope.models.cv.indoor_layout_estimation.panovit": ["os", "numpy", "yacs", "torch"], "modelscope.models.cv.action_recognition.s3dg": ["torch"], "modelscope.models.cv.action_recognition.temporal_patch_shift_transformer": ["einops", "abc", "operator", "numpy", "torchvision", "functools", "torch", "timm"], "modelscope.models.cv.action_recognition.models": ["torch"], "modelscope.models.cv.action_recognition.tada_convnext": ["math", "torch"], "modelscope.models.cv.hand_2d_keypoints.hand_2d_keypoints": ["easycv"], "modelscope.models.cv.easycv_base": ["easycv"], "modelscope.models.cv.image_skychange.ptsemseg.unet": ["torch"], "modelscope.models.cv.image_skychange.ptsemseg.BlockModules": ["torch"], "modelscope.models.cv.image_skychange.ptsemseg.hrnet_backnone": ["os", "numpy", "logging", "torch"], "modelscope.models.cv.image_skychange.ptsemseg.hrnet_super_and_ocr": ["numpy", "__future__", "torch"], "modelscope.models.cv.image_skychange.skychange": ["pdb", "numpy", "json", "numbers", "torchvision", "cv2", "os", "torch", "collections", "PIL"], "modelscope.models.cv.image_skychange.skychange_model": ["time", "math", "pdb", "json", "cv2", "typing", "os", "torch", "collections"], "modelscope.models.cv.image_skychange.preprocessor": ["pdb", "numpy", "json", "numbers", "torchvision", "cv2", "typing", "torch"], "modelscope.models.cv.abnormal_object_detection.mmdet_ms.roi_head.mask_scoring_roi_head": ["mmdet", "torch"], "modelscope.models.cv.abnormal_object_detection.mmdet_ms.roi_head.roi_extractors.single_level_roi_extractor": ["mmdet", "mmcv", "torch"], "modelscope.models.cv.abnormal_object_detection.mmdet_model": ["os", "numpy", "torch"], "modelscope.models.cv.image_color_enhance.adaint.adaint": ["numbers", "torchvision", "typing", "os", "torch"], "modelscope.models.cv.image_color_enhance.deeplpf.deeplpf_image_color_enhance": ["os", "torch", "typing"], "modelscope.models.cv.image_color_enhance.deeplpf.deeplpfnet": ["matplotlib", "math", "torch"], "modelscope.models.cv.image_color_enhance.csrnet": ["functools", "math", "torch"], "modelscope.models.cv.image_color_enhance.image_color_enhance": ["os", "torch", "typing"], "modelscope.models.cv.face_2d_keypoints.face_2d_keypoints_align": ["easycv"], "modelscope.models.cv.video_stabilization.DUT.MotionPro": ["math", "numpy", "cv2", "os", "torch"], "modelscope.models.cv.video_stabilization.DUT.rf_det_module": ["torch"], "modelscope.models.cv.video_stabilization.DUT.DUT_raft": ["torch", "sys", "numpy", "cv2"], "modelscope.models.cv.video_stabilization.DUT.rf_det_so": ["torch"], "modelscope.models.cv.video_stabilization.DUT.RAFT.raft": ["numpy", "torch"], "modelscope.models.cv.video_stabilization.DUT.RAFT.update": ["torch"], "modelscope.models.cv.video_stabilization.DUT.RAFT.extractor": ["torch"], "modelscope.models.cv.video_stabilization.DUT.RAFT.corr": ["alt_cuda_corr", "torch"], "modelscope.models.cv.video_stabilization.DUT.config": ["__future__", "easydict"], "modelscope.models.cv.video_stabilization.DUT.Smoother": ["numpy", "math", "torch"], "modelscope.models.cv.video_stabilization.utils.image_utils": ["skimage", "torch"], "modelscope.models.cv.video_stabilization.utils.IterativeSmooth": ["os", "numpy", "math", "torch"], "modelscope.models.cv.video_stabilization.utils.MedianFilter": ["torch", "numpy", "math", "cv2"], "modelscope.models.cv.video_stabilization.utils.RAFTUtils": ["scipy", "numpy", "torch"], "modelscope.models.cv.video_stabilization.utils.ProjectionUtils": ["torch", "numpy", "math", "cv2"], "modelscope.models.cv.video_stabilization.utils.WarpUtils": ["numpy", "torch", "tqdm"], "modelscope.models.cv.video_stabilization.utils.math_utils": ["numpy", "torch"], "modelscope.models.cv.video_stabilization.DUTRAFTStabilizer": ["math", "numpy", "cv2", "typing", "os", "tempfile", "torch", "sys"], "modelscope.models.cv.video_instance_segmentation.video_knet": ["mmdet", "torch"], "modelscope.models.cv.video_instance_segmentation.head.kernel_frame_iter_head": ["mmdet", "mmcv", "torch"], "modelscope.models.cv.video_instance_segmentation.head.kernel_head": ["mmdet", "mmcv", "torch"], "modelscope.models.cv.video_instance_segmentation.head.kernel_updator": ["mmcv", "torch"], "modelscope.models.cv.video_instance_segmentation.head.kernel_update_head": ["numpy", "mmdet", "mmcv", "torch"], "modelscope.models.cv.video_instance_segmentation.head.kernel_iter_head": ["mmdet", "torch"], "modelscope.models.cv.video_instance_segmentation.utils": ["numpy", "mmdet", "torch"], "modelscope.models.cv.video_instance_segmentation.track.mask_hungarian_assigner": ["scipy", "numpy", "mmdet", "torch"], "modelscope.models.cv.video_instance_segmentation.track.kernel_update_head": ["numpy", "mmdet", "mmcv", "torch"], "modelscope.models.cv.video_instance_segmentation.neck.msdeformattn_decoder": ["mmdet", "mmcv", "torch"], "modelscope.models.cv.product_segmentation.net": ["torch"], "modelscope.models.cv.product_segmentation.seg_infer": ["torch", "numpy", "cv2", "PIL"], "modelscope.models.cv.skin_retouching.weights_init": ["torch"], "modelscope.models.cv.skin_retouching.utils": ["time", "einops", "cv2", "torch", "numpy", "typing"], "modelscope.models.cv.skin_retouching.retinaface.box_utils": ["torch", "numpy", "typing"], "modelscope.models.cv.skin_retouching.retinaface.prior_box": ["torch", "math", "itertools"], "modelscope.models.cv.skin_retouching.retinaface.utils": ["pathlib", "typing", "torch", "numpy", "re", "cv2"], "modelscope.models.cv.skin_retouching.retinaface.predict_single": ["albumentations", "numpy", "torchvision", "typing", "torch"], "modelscope.models.cv.skin_retouching.retinaface.net": ["torch", "typing"], "modelscope.models.cv.skin_retouching.retinaface.network": ["torch", "torchvision", "typing"], "modelscope.models.cv.skin_retouching.detection_model.detection_unet_in": ["torch"], "modelscope.models.cv.skin_retouching.detection_model.detection_module": ["torch"], "modelscope.models.cv.skin_retouching.unet_deploy": ["warnings", "torch"], "modelscope.models.cv.skin_retouching.inpainting_model.inpainting_unet": ["torch"], "modelscope.models.cv.skin_retouching.inpainting_model.gconv": ["torch"], "modelscope.models.cv.product_retrieval_embedding.item_detection": ["numpy", "cv2"], "modelscope.models.cv.product_retrieval_embedding.item_model": ["os", "numpy", "torch", "typing"], "modelscope.models.cv.product_retrieval_embedding.item_embedding": ["cv2", "numpy", "torch"], "modelscope.models.cv.animal_recognition.splat": ["torch"], "modelscope.models.cv.animal_recognition.resnet": ["math", "torch"], "modelscope.models.cv.image_to_image_translation.data.transforms": ["random", "math", "torchvision", "PIL"], "modelscope.models.cv.image_to_image_translation.models.autoencoder": ["math", "torch"], "modelscope.models.cv.image_to_image_translation.models.clip": ["math", "torch"], "modelscope.models.cv.image_to_image_translation.model_translation": ["math", "torch"], "modelscope.models.cv.image_to_image_translation.ops.apps": ["artist", "numpy", "torchvision", "os", "torch", "PIL"], "modelscope.models.cv.image_to_image_translation.ops.svd": ["torch"], "modelscope.models.cv.image_to_image_translation.ops.random_mask": ["numpy", "cv2"], "modelscope.models.cv.image_to_image_translation.ops.utils": ["io", "math", "multiprocessing", "numpy", "json", "hashlib", "base64", "cv2", "os", "zipfile", "binascii", "torch", "PIL"], "modelscope.models.cv.image_to_image_translation.ops.metrics": ["scipy", "numpy", "torch"], "modelscope.models.cv.image_to_image_translation.ops.diffusion": ["math", "torch"], "modelscope.models.cv.image_to_image_translation.ops.degradation": ["os", "random", "math", "torch", "scipy", "numpy", "cv2"], "modelscope.models.cv.image_to_image_translation.ops.losses": ["math", "torch"], "modelscope.models.cv.image_to_image_translation.ops.random_color": ["random", "colorsys"], "modelscope.models.cv.image_probing_model.backbone": ["math", "operator", "numpy", "torchvision", "torch", "functools", "sys", "collections", "PIL"], "modelscope.models.cv.image_probing_model.model": ["os", "torch", "json", "typing"], "modelscope.models.cv.image_probing_model.utils": ["re", "torch"], "modelscope.models.base.base_head": ["typing", "abc"], "modelscope.models.base.base_torch_head": ["torch", "typing"], "modelscope.models.base.base_torch_model": ["typing", "packaging", "os", "torch", "functools", "copy"], "modelscope.models.base.base_model": ["os", "typing", "abc"], "modelscope.models.multi_modal.video_synthesis.unet_sd": ["einops", "math", "torch"], "modelscope.models.multi_modal.video_synthesis.autoencoder": ["numpy", "torch"], "modelscope.models.multi_modal.video_synthesis.diffusion": ["torch"], "modelscope.models.multi_modal.video_synthesis.text_to_video_synthesis_model": ["einops", "open_clip", "typing", "os", "torch"], "modelscope.models.multi_modal.mplug_for_all_tasks": ["os", "typing"], "modelscope.models.multi_modal.vldoc.processing": ["numpy", "torchvision", "cv2", "typing", "torch", "timm", "collections", "PIL"], "modelscope.models.multi_modal.vldoc.model": ["math", "json", "re", "torchvision", "os", "copy", "torch", "sys", "logging"], "modelscope.models.multi_modal.vldoc.tokenization": ["os", "transformers"], "modelscope.models.multi_modal.vldoc.modeling_layout_roberta": ["os", "math", "torch", "transformers", "packaging"], "modelscope.models.multi_modal.vldoc.convnext": ["os", "timm", "torch"], "modelscope.models.multi_modal.vldoc.transformer_local": ["torch", "copy"], "modelscope.models.multi_modal.vldoc.conv_fpn_trans": ["apex", "random", "torch", "timm", "collections"], "modelscope.models.multi_modal.mplug.mvit": ["timm", "functools", "torch", "collections", "numpy", "fairscale"], "modelscope.models.multi_modal.mplug.modeling_mplug": ["math", "transformers", "typing", "os", "torch"], "modelscope.models.multi_modal.mplug.predictor": ["__future__", "torch"], "modelscope.models.multi_modal.mplug.configuration_mplug": ["os", "yaml", "typing", "transformers"], "modelscope.models.multi_modal.mplug.clip.clip": ["torch", "collections", "typing"], "modelscope.models.multi_modal.ofa_for_all_tasks": ["math", "json", "re", "typing", "string", "os", "torch", "functools"], "modelscope.models.multi_modal.guided_diffusion.unet": ["math", "torch", "abc", "numpy", "transformers"], "modelscope.models.multi_modal.guided_diffusion.script": [], "modelscope.models.multi_modal.guided_diffusion.respace": ["numpy", "torch"], "modelscope.models.multi_modal.guided_diffusion.gaussian_diffusion": ["numpy", "enum", "math", "torch"], "modelscope.models.multi_modal.gemm.gemm_model": ["numpy", "json", "torchvision", "typing", "os", "torch", "PIL"], "modelscope.models.multi_modal.gemm.tokenizer": ["os", "torch", "functools", "regex", "html", "gzip", "ftfy"], "modelscope.models.multi_modal.gemm.gemm_base": ["numpy", "json", "typing", "os", "torch", "collections"], "modelscope.models.multi_modal.ofa_for_text_to_image_synthesis_model": ["numpy", "json", "pkg_resources", "torchvision", "typing", "os", "torch", "taming", "PIL"], "modelscope.models.multi_modal.soonet.blocks": ["math", "torch"], "modelscope.models.multi_modal.soonet.model": ["os", "torch"], "modelscope.models.multi_modal.soonet.utils": ["numpy", "decord", "copy", "tqdm"], "modelscope.models.multi_modal.soonet.swin_transformer": ["numpy", "torch"], "modelscope.models.multi_modal.soonet.tokenizer": ["torch", "functools", "regex", "html", "gzip", "ftfy"], "modelscope.models.multi_modal.soonet.clip": ["torch", "warnings", "numpy", "collections", "typing"], "modelscope.models.multi_modal.team.team_model": ["numpy", "torchvision", "cv2", "typing", "tokenizers", "torch", "PIL"], "modelscope.models.multi_modal.team.utils": ["torch", "numpy", "collections", "transformers", "typing"], "modelscope.models.multi_modal.dpm_solver_pytorch": ["math", "torch"], "modelscope.models.multi_modal.mmr.models.until_module": ["numpy", "logging", "math", "torch"], "modelscope.models.multi_modal.mmr.models.modeling": ["platform", "os", "types", "torch", "collections"], "modelscope.models.multi_modal.mmr.models.module_cross": ["json", "torch", "__future__", "collections", "logging"], "modelscope.models.multi_modal.mmr.models.module_clip": ["hashlib", "typing", "os", "urllib", "torch", "warnings", "collections", "tqdm"], "modelscope.models.multi_modal.mmr.models.clip_for_mm_video_embedding": ["uuid", "numpy", "json", "typing", "os", "random", "urllib", "tempfile", "decord", "torch", "PIL"], "modelscope.models.multi_modal.mmr.models.tokenization_clip": ["os", "functools", "regex", "html", "gzip", "ftfy"], "modelscope.models.multi_modal.mmr.models.dynamic_inverted_softmax": ["numpy"], "modelscope.models.multi_modal.mmr.dataloaders.rawvideo_util": ["numpy", "torchvision", "cv2", "torch", "PIL"], "modelscope.models.multi_modal.diffusion.structbert": ["six", "math", "copy", "torch", "__future__", "numpy", "json"], "modelscope.models.multi_modal.diffusion.model": ["numpy", "json", "typing", "os", "torch"], "modelscope.models.multi_modal.diffusion.unet_upsampler_256": ["torch", "math", "functools"], "modelscope.models.multi_modal.diffusion.diffusion": ["math", "torch"], "modelscope.models.multi_modal.diffusion.unet_upsampler_1024": ["math", "torch"], "modelscope.models.multi_modal.diffusion.tokenizer": ["unicodedata", "collections", "six", "__future__"], "modelscope.models.multi_modal.diffusion.unet_generator": ["math", "torch"], "modelscope.models.multi_modal.mgeo.backbone": ["math", "transformers", "dataclasses", "typing", "os", "random", "torch", "warnings"], "modelscope.models.multi_modal.mgeo.text_classification": ["torch"], "modelscope.models.multi_modal.mgeo.token_classification": ["torch"], "modelscope.models.multi_modal.mgeo.text_ranking": ["torch"], "modelscope.models.multi_modal.clip.model": ["numpy", "json", "typing", "os", "torch", "collections"], "modelscope.models.multi_modal.clip.modeling_bert": ["io", "math", "json", "os", "torch", "sys", "__future__", "logging"], "modelscope.models.multi_modal.clip.bert_tokenizer": ["os", "unicodedata", "six", "__future__", "collections", "re"], "modelscope.models.multi_modal.clip.configuration_bert": ["__future__", "logging"], "modelscope.models.multi_modal.ofa.configuration_mmspeech": ["warnings", "transformers"], "modelscope.models.multi_modal.ofa.modeling_ofa": ["math", "transformers", "dataclasses", "typing", "packaging", "random", "apex", "torch"], "modelscope.models.multi_modal.ofa.tokenization_ofa_fast": ["tokenizers", "json", "transformers", "typing"], "modelscope.models.multi_modal.ofa.modeling_mmspeech": ["math", "numpy", "transformers", "dataclasses", "typing", "fairseq", "packaging", "apex", "torch"], "modelscope.models.multi_modal.ofa.generate.utils": ["torch_xla", "itertools", "torch", "amp_C", "collections"], "modelscope.models.multi_modal.ofa.generate.multihead_attention": ["fairseq", "typing", "math", "torch"], "modelscope.models.multi_modal.ofa.generate.token_generation_constraints": ["torch", "collections", "typing"], "modelscope.models.multi_modal.ofa.generate.ngram_repeat_block": ["math", "torch", "warnings", "fairseq", "typing"], "modelscope.models.multi_modal.ofa.generate.search": ["torch", "math", "typing"], "modelscope.models.multi_modal.ofa.generate.incremental_decoding_utils": ["uuid", "typing", "torch"], "modelscope.models.multi_modal.ofa.generate.sequence_generator": ["torch", "sys", "math", "typing"], "modelscope.models.multi_modal.ofa.tokenization_ofa": ["os", "collections", "transformers", "typing"], "modelscope.models.multi_modal.ofa.utils.utils": ["torch", "typing"], "modelscope.models.multi_modal.ofa.utils.constant": [], "modelscope.models.multi_modal.ofa.configuration_ofa": ["warnings", "transformers"], "modelscope.models.multi_modal.ofa.resnet": ["torch"], "modelscope.models.multi_modal.ofa.vit": ["collections", "fairseq", "torch"], "modelscope.models.multi_modal.multi_stage_diffusion.upsampler": ["math", "torch"], "modelscope.models.multi_modal.multi_stage_diffusion.model": ["math", "numpy", "json", "typing", "os", "torch", "PIL"], "modelscope.models.multi_modal.multi_stage_diffusion.gaussian_diffusion": ["math", "torch"], "modelscope.models.multi_modal.multi_stage_diffusion.tokenizer": ["torch", "functools", "regex", "html", "transformers", "gzip", "ftfy"], "modelscope.models.multi_modal.multi_stage_diffusion.decoder": ["math", "torch"], "modelscope.models.multi_modal.multi_stage_diffusion.prior": ["math", "torch"], "modelscope.models.multi_modal.multi_stage_diffusion.clip": ["math", "torch"], "modelscope.models.multi_modal.multi_stage_diffusion.xglm": ["math", "torch"], "modelscope.models.multi_modal.rleg.rleg": ["torch", "torchvision", "typing"], "modelscope.models.multi_modal.rleg.model": ["os", "json", "torch"], "modelscope.metrics.accuracy_metric": ["numpy", "typing"], "modelscope.metrics.image_colorization_metric": ["torch", "scipy", "numpy", "typing"], "modelscope.metrics.builder": ["typing"], "modelscope.metrics.base": ["typing", "abc"], "modelscope.metrics.text_ranking_metric": ["numpy", "typing"], "modelscope.metrics.map_metric": ["numpy", "typing"], "modelscope.metrics.video_super_resolution_metric.niqe": ["scipy", "numpy", "math", "cv2"], "modelscope.metrics.video_super_resolution_metric.video_super_resolution_metric": ["numpy", "typing"], "modelscope.metrics.video_super_resolution_metric.metric_util": ["numpy"], "modelscope.metrics.video_super_resolution_metric.matlab_functions": ["numpy", "math", "torch"], "modelscope.metrics.referring_video_object_segmentation_metric": ["pycocotools", "numpy", "typing", "torch", "tqdm"], "modelscope.metrics.image_quality_assessment_degradation_metric": ["scipy", "numpy", "cv2", "typing", "os", "tempfile", "torch", "sys", "collections", "tqdm"], "modelscope.metrics.inbatch_recall_metric": ["torch", "numpy", "typing"], "modelscope.metrics.movie_scene_segmentation_metric": ["numpy", "typing"], "modelscope.metrics.image_denoise_metric": ["torch", "cv2", "numpy", "typing"], "modelscope.metrics.image_quality_assessment_mos_metric": ["scipy", "numpy", "typing", "cv2", "os", "tempfile", "torch", "sys", "tqdm"], "modelscope.metrics.image_portrait_enhancement_metric": ["cv2", "numpy", "typing"], "modelscope.metrics.ocr_recognition_metric": ["edit_distance", "numpy", "torch", "typing"], "modelscope.metrics.image_inpainting_metric": ["torch", "scipy", "numpy", "typing"], "modelscope.metrics.prediction_saving_wrapper": ["numpy", "sklearn", "typing"], "modelscope.metrics.sequence_classification_metric": ["numpy", "sklearn", "typing"], "modelscope.metrics.ppl_metric": ["torch", "numpy", "math", "typing"], "modelscope.metrics.video_frame_interpolation_metric": ["math", "lpips", "numpy", "typing", "torch"], "modelscope.metrics.loss_metric": ["numpy", "sklearn", "typing"], "modelscope.metrics.video_stabilization_metric": ["numpy", "typing", "cv2", "os", "tempfile", "sys", "tqdm"], "modelscope.metrics.ciderD.ciderD_scorer": ["six", "math", "pdb", "numpy", "os", "copy", "__future__", "collections"], "modelscope.metrics.ciderD.ciderD": ["__future__"], "modelscope.metrics.ned_metric": ["numpy", "typing"], "modelscope.metrics.text_generation_metric": ["rouge", "nltk", "typing"], "modelscope.metrics.bleu_metric": ["sacrebleu", "typing", "itertools"], "modelscope.metrics.action_detection_evaluator": ["pandas", "scipy", "numpy", "os", "copy", "collections", "logging", "detectron2"], "modelscope.metrics.token_classification_metric": ["numpy", "importlib", "typing"], "modelscope.metrics.image_instance_segmentation_metric": ["pycocotools", "numpy", "typing", "os", "tempfile", "collections"], "modelscope.metrics.image_color_enhance_metric": ["cv2", "numpy", "typing"], "modelscope.metrics.audio_noise_metric": ["typing"], "modelscope.metrics.video_summarization_metric": ["numpy", "typing"], "modelscope.pipelines.builder": ["os", "typing"], "modelscope.pipelines.base": ["abc", "multiprocessing", "threading", "numpy", "typing", "packaging", "os", "random", "torch", "functools"], "modelscope.pipelines.audio.separation_pipeline": ["soundfile", "io", "numpy", "typing", "torch"], "modelscope.pipelines.audio.text_to_speech_pipeline": ["numpy", "typing"], "modelscope.pipelines.audio.kws_farfield_pipeline": ["soundfile", "io", "numpy", "typing", "wave"], "modelscope.pipelines.audio.punctuation_processing_pipeline": ["os", "shutil", "yaml", "typing"], "modelscope.pipelines.audio.inverse_text_processing_pipeline": ["os", "shutil", "yaml", "typing"], "modelscope.pipelines.audio.speaker_verification_pipeline": ["os", "shutil", "yaml", "typing"], "modelscope.pipelines.audio.timestamp_pipeline": ["json", "yaml", "typing", "os", "funasr"], "modelscope.pipelines.audio.linear_aec_pipeline": ["scipy", "numpy", "yaml", "typing", "os", "torch", "importlib"], "modelscope.pipelines.audio.speaker_verification_light_pipeline": ["soundfile", "torch", "io", "typing"], "modelscope.pipelines.audio.speaker_diarization_pipeline": ["shutil", "numpy", "json", "yaml", "typing", "os"], "modelscope.pipelines.audio.ans_dfsmn_pipeline": ["librosa", "soundfile", "io", "numpy", "typing", "os", "torch", "sys", "collections"], "modelscope.pipelines.audio.voice_activity_detection_pipeline": ["json", "yaml", "typing", "os", "funasr"], "modelscope.pipelines.audio.ans_pipeline": ["librosa", "soundfile", "io", "numpy", "typing", "torch"], "modelscope.pipelines.audio.asr_inference_pipeline": ["os", "json", "yaml", "typing"], "modelscope.pipelines.audio.asr_wenet_inference_pipeline": ["typing"], "modelscope.pipelines.audio.kws_kwsbp_pipeline": ["os", "json", "typing"], "modelscope.pipelines.audio.lm_infer_pipeline": ["os", "typing"], "modelscope.pipelines.util": ["os", "typing"], "modelscope.pipelines.science.protein_structure_pipeline": ["time", "numpy", "json", "typing", "unicore", "os", "torch"], "modelscope.pipelines.nlp.text_generation_pipeline": ["os", "torch", "typing"], "modelscope.pipelines.nlp.distributed_gpt_moe_pipeline": ["torch", "typing"], "modelscope.pipelines.nlp.document_segmentation_pipeline": ["numpy", "re", "transformers", "typing", "datasets", "torch"], "modelscope.pipelines.nlp.distributed_plug_pipeline": ["torch", "typing"], "modelscope.pipelines.nlp.document_grounded_dialog_generate_pipeline": ["typing"], "modelscope.pipelines.nlp.named_entity_recognition_pipeline": ["typing"], "modelscope.pipelines.nlp.feature_extraction_pipeline": ["os", "torch", "typing"], "modelscope.pipelines.nlp.extractive_summarization_pipeline": ["numpy", "re", "typing", "datasets", "torch"], "modelscope.pipelines.nlp.fasttext_text_classification_pipeline": ["numpy", "fasttext", "typing", "sentencepiece", "os"], "modelscope.pipelines.nlp.token_classification_pipeline": ["torch", "numpy", "typing"], "modelscope.pipelines.nlp.sentence_embedding_pipeline": ["torch", "numpy", "typing"], "modelscope.pipelines.nlp.translation_quality_estimation_pipeline": ["io", "transformers", "typing", "os", "torch"], "modelscope.pipelines.nlp.dialog_state_tracking_pipeline": ["typing"], "modelscope.pipelines.nlp.table_question_answering_pipeline": ["json", "transformers", "typing", "os", "torch"], "modelscope.pipelines.nlp.faq_question_answering_pipeline": ["typing"], "modelscope.pipelines.nlp.siamese_uie_pipeline": ["pathlib", "time", "math", "scipy", "json", "typing", "os", "torch", "copy", "logging", "tqdm"], "modelscope.pipelines.nlp.text_classification_pipeline": ["torch", "numpy", "typing"], "modelscope.pipelines.nlp.fill_mask_pipeline": ["numpy", "typing"], "modelscope.pipelines.nlp.zero_shot_classification_pipeline": ["torch", "scipy", "typing"], "modelscope.pipelines.nlp.dialog_intent_prediction_pipeline": ["typing"], "modelscope.pipelines.nlp.word_alignment_pipeline": ["numpy", "typing"], "modelscope.pipelines.nlp.mglm_text_summarization_pipeline": ["os", "typing"], "modelscope.pipelines.nlp.text_error_correction_pipeline": ["torch", "typing"], "modelscope.pipelines.nlp.word_segmentation_pipeline": ["torch", "typing"], "modelscope.pipelines.nlp.language_identification_pipline": ["numpy", "re", "typing", "os", "tensorflow"], "modelscope.pipelines.nlp.document_grounded_dialog_retrieval_pipeline": ["numpy", "json", "typing", "os", "faiss"], "modelscope.pipelines.nlp.translation_pipeline": ["jieba", "subword_nmt", "numpy", "sacremoses", "typing", "os", "tensorflow"], "modelscope.pipelines.nlp.dialog_modeling_pipeline": ["typing"], "modelscope.pipelines.nlp.interactive_translation_pipeline": ["jieba", "subword_nmt", "numpy", "sacremoses", "typing", "os", "tensorflow"], "modelscope.pipelines.nlp.fid_dialogue_pipeline": ["torch", "re", "typing"], "modelscope.pipelines.nlp.user_satisfaction_estimation_pipeline": ["torch", "numpy", "typing"], "modelscope.pipelines.nlp.text_ranking_pipeline": ["numpy", "typing"], "modelscope.pipelines.nlp.codegeex_code_translation_pipeline": ["typing"], "modelscope.pipelines.nlp.summarization_pipeline": ["torch", "typing"], "modelscope.pipelines.nlp.translation_evaluation_pipeline": ["numpy", "enum", "typing", "os", "torch"], "modelscope.pipelines.nlp.distributed_gpt3_pipeline": ["torch", "typing"], "modelscope.pipelines.nlp.document_grounded_dialog_rerank_pipeline": ["time", "numpy", "re", "transformers", "typing", "pprint", "os", "random", "torch", "sys", "collections", "ujson"], "modelscope.pipelines.nlp.conversational_text_to_sql_pipeline": ["torch", "text2sql_lgesql", "typing"], "modelscope.pipelines.nlp.information_extraction_pipeline": ["torch", "typing"], "modelscope.pipelines.nlp.automatic_post_editing_pipeline": ["jieba", "html", "numpy", "sacremoses", "typing", "sentencepiece", "os", "tensorflow"], "modelscope.pipelines.nlp.codegeex_code_generation_pipeline": ["typing"], "modelscope.pipelines.cv.body_3d_keypoints_pipeline": ["datetime", "matplotlib", "mpl_toolkits", "numpy", "cv2", "typing", "os", "tempfile", "torch"], "modelscope.pipelines.cv.image_color_enhance_pipeline": ["torch", "torchvision", "typing"], "modelscope.pipelines.cv.image_paintbyexample_pipeline": ["einops", "numpy", "torchvision", "cv2", "typing", "torch", "PIL"], "modelscope.pipelines.cv.animal_recognition_pipeline": ["numpy", "torchvision", "typing", "cv2", "os", "torch", "PIL"], "modelscope.pipelines.cv.image_bts_depth_estimation_pipeline": ["albumentations", "numpy", "cv2", "typing", "torch"], "modelscope.pipelines.cv.image_panoptic_segmentation_pipeline": ["numpy", "cv2", "typing", "torch", "PIL"], "modelscope.pipelines.cv.hand_2d_keypoints_pipeline": ["os"], "modelscope.pipelines.cv.image_skychange_pipeline": ["time", "pdb", "numpy", "typing", "cv2", "PIL"], "modelscope.pipelines.cv.video_inpainting_pipeline": ["typing"], "modelscope.pipelines.cv.image_inpainting_sdv2_pipeline": ["math", "diffusers", "numpy", "cv2", "typing", "os", "tempfile", "torch", "sys"], "modelscope.pipelines.cv.action_recognition_pipeline": ["os", "torch", "math", "typing"], "modelscope.pipelines.cv.vop_retrieval_pipeline": ["pickle", "math", "numpy", "gzip", "typing", "os", "random", "torch", "collections", "tqdm"], "modelscope.pipelines.cv.image_denoise_pipeline": ["torch", "torchvision", "typing"], "modelscope.pipelines.cv.vision_middleware_pipeline": ["math", "numpy", "torchvision", "typing", "os", "torch", "mmcv"], "modelscope.pipelines.cv.text_driven_segmentation_pipleline": ["typing"], "modelscope.pipelines.cv.video_single_object_tracking_pipeline": ["os", "cv2", "typing"], "modelscope.pipelines.cv.image_structured_model_probing_pipeline": ["math", "numpy", "torchvision", "typing", "os", "torch", "mmcv"], "modelscope.pipelines.cv.image_face_fusion_pipeline": ["numpy", "typing"], "modelscope.pipelines.cv.controllable_image_generation_pipeline": ["math", "subprocess", "numpy", "cv2", "typing", "os", "tempfile", "torch", "glob"], "modelscope.pipelines.cv.table_recognition_pipeline": ["math", "numpy", "cv2", "typing", "os", "torch", "PIL"], "modelscope.pipelines.cv.language_guided_video_summarization_pipeline": ["shutil", "numpy", "cv2", "typing", "os", "random", "tempfile", "torch", "clip", "PIL"], "modelscope.pipelines.cv.content_check_pipeline": ["numpy", "torchvision", "cv2", "typing", "os", "torch", "PIL"], "modelscope.pipelines.cv.image_quality_assessment_man_pipeline": ["math", "numpy", "torchvision", "cv2", "typing", "tempfile", "torch"], "modelscope.pipelines.cv.image_matting_pipeline": ["numpy", "cv2", "typing", "os", "tensorflow"], "modelscope.pipelines.cv.lineless_table_recognition_pipeline": ["math", "numpy", "cv2", "typing", "os", "torch", "PIL"], "modelscope.pipelines.cv.body_2d_keypoints_pipeline": ["numpy", "json", "torchvision", "cv2", "typing", "os", "torch", "PIL"], "modelscope.pipelines.cv.tbs_detection_utils.utils": ["colorsys", "pandas", "matplotlib", "numpy", "torchvision", "os", "torch", "__future__", "PIL"], "modelscope.pipelines.cv.image_depth_estimation_pipeline": ["numpy", "cv2", "typing", "torch", "PIL"], "modelscope.pipelines.cv.image_cartoon_pipeline": ["numpy", "cv2", "typing", "os", "tensorflow"], "modelscope.pipelines.cv.ocr_recognition_pipeline": [], "modelscope.pipelines.cv.image_to_image_translation_pipeline": ["io", "numpy", "torchvision", "cv2", "typing", "os", "torch", "sys", "PIL"], "modelscope.pipelines.cv.image_open_vocabulary_detection_pipeline": ["numpy", "typing", "cv2", "os", "torch", "PIL"], "modelscope.pipelines.cv.facial_landmark_confidence_pipeline": ["numpy", "typing", "cv2", "os", "torch", "PIL"], "modelscope.pipelines.cv.mobile_image_super_resolution_pipeline": ["numpy", "torchvision", "typing", "torch", "skimage"], "modelscope.pipelines.cv.image_semantic_segmentation_pipeline": ["numpy", "cv2", "typing", "torch", "PIL"], "modelscope.pipelines.cv.image_matching_pipeline": ["numpy", "cv2", "typing", "torch", "PIL"], "modelscope.pipelines.cv.face_reconstruction_pipeline": ["shutil", "scipy", "numpy", "cv2", "typing", "os", "torch", "tensorflow", "face_alignment", "PIL"], "modelscope.pipelines.cv.video_frame_interpolation_pipeline": ["math", "subprocess", "numpy", "torchvision", "cv2", "typing", "os", "tempfile", "torch", "glob"], "modelscope.pipelines.cv.tbs_detection_pipeline": ["colorsys", "numpy", "cv2", "typing", "os", "torch", "PIL"], "modelscope.pipelines.cv.indoor_layout_estimation_pipeline": ["cv2", "numpy", "typing"], "modelscope.pipelines.cv.video_human_matting_pipeline": ["moviepy", "numpy", "cv2", "typing", "os", "torch"], "modelscope.pipelines.cv.skin_retouching_pipeline": ["numpy", "torchvision", "cv2", "typing", "os", "torch", "tensorflow", "PIL"], "modelscope.pipelines.cv.video_deinterlace_pipeline": ["math", "subprocess", "numpy", "torchvision", "cv2", "typing", "os", "tempfile", "torch"], "modelscope.pipelines.cv.image_to_image_generate_pipeline": ["numpy", "torchvision", "cv2", "typing", "os", "torch", "PIL"], "modelscope.pipelines.cv.face_attribute_recognition_pipeline": ["numpy", "typing", "cv2", "os", "torch", "PIL"], "modelscope.pipelines.cv.image_human_parsing_pipeline": ["torch", "numpy", "torchvision", "typing"], "modelscope.pipelines.cv.video_depth_estimation_pipeline": ["typing"], "modelscope.pipelines.cv.image_body_reshaping_pipeline": ["typing"], "modelscope.pipelines.cv.mog_face_detection_pipeline": ["os", "numpy", "typing"], "modelscope.pipelines.cv.mask_face_recognition_pipeline": ["numpy", "typing", "cv2", "os", "torch", "collections", "PIL"], "modelscope.pipelines.cv.product_segmentation_pipeline": ["numpy", "typing"], "modelscope.pipelines.cv.ddpm_semantic_segmentation_pipeline": ["torch", "torchvision", "typing"], "modelscope.pipelines.cv.cmdssl_video_embedding_pipeline": ["numpy", "torchvision", "typing", "os", "decord", "torch", "PIL"], "modelscope.pipelines.cv.face_processing_base_pipeline": ["numpy", "typing", "cv2", "os", "torch", "PIL"], "modelscope.pipelines.cv.image_classification_pipeline": ["torch", "numpy", "typing"], "modelscope.pipelines.cv.referring_video_object_segmentation_pipeline": ["moviepy", "einops", "numpy", "torchvision", "typing", "tqdm", "tempfile", "torch", "PIL"], "modelscope.pipelines.cv.motion_generation_pipeline": ["numpy", "typing", "os", "tempfile", "torch"], "modelscope.pipelines.cv.image_inpainting_pipeline": ["numpy", "cv2", "typing", "torch", "PIL"], "modelscope.pipelines.cv.ocr_utils.table_process": ["random", "math", "copy", "torch", "numpy", "cv2"], "modelscope.pipelines.cv.ocr_utils.utils": ["numpy", "pyclipper", "cv2", "shapely"], "modelscope.pipelines.cv.ocr_utils.resnet18_v1": ["tensorflow", "tf_slim"], "modelscope.pipelines.cv.ocr_utils.model_resnet18_half": ["os", "torch"], "modelscope.pipelines.cv.ocr_utils.model_dla34": ["os", "numpy", "math", "torch"], "modelscope.pipelines.cv.ocr_utils.ocr_modules.timm_tinyc": ["math", "itertools", "torch", "copy", "functools", "collections", "logging"], "modelscope.pipelines.cv.ocr_utils.ocr_modules.vitstr": ["torch", "functools", "copy", "__future__", "logging"], "modelscope.pipelines.cv.ocr_utils.ocr_modules.convnext": ["torch"], "modelscope.pipelines.cv.ocr_utils.model_convnext_transformer": ["torch"], "modelscope.pipelines.cv.ocr_utils.ops": ["shutil", "math", "uuid", "numpy", "cv2", "tensorflow", "os", "sys", "absl"], "modelscope.pipelines.cv.ocr_utils.model_resnet_mutex_v4_linewithchar": ["tensorflow", "tf_slim"], "modelscope.pipelines.cv.ocr_utils.model_vlpt": ["os", "sys", "math", "torch"], "modelscope.pipelines.cv.ocr_utils.resnet_utils": ["tensorflow", "tf_slim", "collections"], "modelscope.pipelines.cv.image_quality_assessment_degradation_pipeline": ["math", "numpy", "torchvision", "cv2", "typing", "tempfile", "torch"], "modelscope.pipelines.cv.arc_face_recognition_pipeline": ["numpy", "typing", "cv2", "os", "torch", "PIL"], "modelscope.pipelines.cv.panorama_depth_estimation_pipeline": ["numpy", "cv2", "typing", "torch", "PIL"], "modelscope.pipelines.cv.video_colorization_pipeline": ["subprocess", "numpy", "torchvision", "cv2", "typing", "os", "tempfile", "torch", "PIL"], "modelscope.pipelines.cv.image_instance_segmentation_pipeline": ["numpy", "cv2", "typing", "os", "torch", "PIL"], "modelscope.pipelines.cv.movie_scene_segmentation_pipeline": ["torch", "typing"], "modelscope.pipelines.cv.action_detection_pipeline": ["os", "math", "typing"], "modelscope.pipelines.cv.product_retrieval_embedding_pipeline": ["numpy", "torchvision", "typing", "cv2", "os", "torch", "PIL"], "modelscope.pipelines.cv.image_restoration_pipeline": ["typing"], "modelscope.pipelines.cv.realtime_video_object_detection_pipeline": ["numpy", "json", "torchvision", "cv2", "typing", "os", "torch", "PIL"], "modelscope.pipelines.cv.live_category_pipeline": ["numpy", "torchvision", "typing", "os", "decord", "torch", "PIL"], "modelscope.pipelines.cv.video_stabilization_pipeline": ["math", "subprocess", "numpy", "cv2", "typing", "os", "tempfile", "torch", "glob"], "modelscope.pipelines.cv.video_category_pipeline": ["numpy", "json", "torchvision", "typing", "os", "decord", "torch", "PIL"], "modelscope.pipelines.cv.ulfd_face_detection_pipeline": ["numpy", "typing", "cv2", "os", "torch", "PIL"], "modelscope.pipelines.cv.video_multi_object_tracking_pipeline": ["os", "torch", "typing"], "modelscope.pipelines.cv.face_recognition_onnx_ir_pipeline": ["numpy", "typing", "cv2", "os", "onnxruntime", "torch", "PIL"], "modelscope.pipelines.cv.image_mvs_depth_estimation_pipeline": ["os", "shutil", "tempfile", "typing"], "modelscope.pipelines.cv.image_portrait_enhancement_pipeline": ["math", "scipy", "numpy", "cv2", "typing", "torch", "PIL"], "modelscope.pipelines.cv.crowd_counting_pipeline": ["math", "numpy", "torchvision", "typing", "torch", "PIL"], "modelscope.pipelines.cv.retina_face_detection_pipeline": ["numpy", "typing", "cv2", "os", "torch", "PIL"], "modelscope.pipelines.cv.human_reconstruction_pipeline": ["shutil", "numpy", "typing", "os", "torch", "trimesh"], "modelscope.pipelines.cv.general_recognition_pipeline": ["numpy", "torchvision", "typing", "cv2", "os", "torch", "PIL"], "modelscope.pipelines.cv.face_recognition_pipeline": ["numpy", "typing", "cv2", "os", "torch", "PIL"], "modelscope.pipelines.cv.hicossl_video_embedding_pipeline": ["os", "torch", "math", "typing"], "modelscope.pipelines.cv.mtcnn_face_detection_pipeline": ["os", "torch", "typing"], "modelscope.pipelines.cv.tinynas_detection_pipeline": ["typing"], "modelscope.pipelines.cv.image_salient_detection_pipeline": ["typing"], "modelscope.pipelines.cv.face_human_hand_detection_pipeline": ["numpy", "typing"], "modelscope.pipelines.cv.image_deblur_pipeline": ["torch", "torchvision", "typing"], "modelscope.pipelines.cv.virtual_try_on_pipeline": ["numpy", "typing", "cv2", "os", "torch", "PIL"], "modelscope.pipelines.cv.video_object_segmentation_pipeline": ["numpy", "torchvision", "typing", "os", "torch", "PIL"], "modelscope.pipelines.cv.face_recognition_onnx_fm_pipeline": ["numpy", "typing", "cv2", "os", "onnxruntime", "torch", "PIL"], "modelscope.pipelines.cv.object_detection_3d_pipeline": ["numpy", "cv2", "typing", "os", "tempfile", "torch", "PIL"], "modelscope.pipelines.cv.video_instance_segmentation_pipeline": ["numpy", "cv2", "typing", "os", "torch", "mmcv", "tqdm"], "modelscope.pipelines.cv.bad_image_detecting_pipeline": ["torch", "numpy", "typing"], "modelscope.pipelines.cv.pointcloud_sceneflow_estimation_pipeline": ["torch", "numpy", "plyfile", "typing"], "modelscope.pipelines.cv.maskdino_instance_segmentation_pipeline": ["torch", "torchvision", "typing"], "modelscope.pipelines.cv.image_quality_assessment_mos_pipeline": ["math", "numpy", "torchvision", "cv2", "typing", "tempfile", "torch"], "modelscope.pipelines.cv.face_liveness_ir_pipeline": ["numpy", "typing", "cv2", "os", "onnxruntime", "torch", "PIL"], "modelscope.pipelines.cv.face_liveness_xc_pipeline": ["numpy", "typing", "cv2", "os", "onnxruntime", "torch", "PIL"], "modelscope.pipelines.cv.image_driving_perception_pipeline": ["os", "cv2", "numpy", "typing"], "modelscope.pipelines.cv.facial_expression_recognition_pipeline": ["numpy", "typing", "cv2", "os", "torch", "PIL"], "modelscope.pipelines.cv.hand_static_pipeline": ["numpy", "typing"], "modelscope.pipelines.cv.video_panoptic_segmentation_pipeline": ["numpy", "cv2", "typing", "os", "torch", "mmcv", "tqdm"], "modelscope.pipelines.cv.video_super_resolution_pipeline": ["math", "subprocess", "numpy", "torchvision", "cv2", "typing", "os", "tempfile", "torch"], "modelscope.pipelines.cv.license_plate_detection_pipeline": ["math", "numpy", "cv2", "typing", "os", "torch", "PIL"], "modelscope.pipelines.cv.ddcolor_image_colorization_pipeline": ["numpy", "torchvision", "cv2", "typing", "torch"], "modelscope.pipelines.cv.vision_efficient_tuning_pipeline": ["torch", "numpy", "torchvision", "typing"], "modelscope.pipelines.cv.face_detection_pipeline": ["numpy", "typing", "cv2", "os", "torch", "PIL"], "modelscope.pipelines.cv.image_style_transfer_pipeline": ["os", "cv2", "numpy", "typing"], "modelscope.pipelines.cv.vop_retrieval_se_pipeline": ["numpy", "gzip", "typing", "os", "torch"], "modelscope.pipelines.cv.image_reid_person_pipeline": ["math", "torchvision", "typing", "os", "torch", "PIL"], "modelscope.pipelines.cv.image_detection_pipeline": ["numpy", "typing"], "modelscope.pipelines.cv.ocr_detection_pipeline": ["math", "numpy", "typing", "cv2", "os", "tf_slim", "torch", "tensorflow"], "modelscope.pipelines.cv.tinynas_classification_pipeline": ["math", "torchvision", "typing", "os", "torch"], "modelscope.pipelines.cv.image_colorization_pipeline": ["numpy", "torchvision", "cv2", "typing", "torch", "PIL"], "modelscope.pipelines.cv.card_detection_pipeline": ["typing"], "modelscope.pipelines.cv.video_summarization_pipeline": ["numpy", "cv2", "typing", "os", "torch", "tqdm"], "modelscope.pipelines.cv.image_defrcn_fewshot_pipeline": ["os", "numpy", "torch", "typing"], "modelscope.pipelines.cv.face_quality_assessment_pipeline": ["numpy", "typing", "cv2", "os", "onnxruntime", "torch", "PIL"], "modelscope.pipelines.cv.image_super_resolution_pipeline": ["numpy", "cv2", "typing", "torch", "PIL"], "modelscope.pipelines.cv.face_image_generation_pipeline": ["numpy", "typing", "cv2", "os", "torch", "PIL"], "modelscope.pipelines.cv.image_debanding_pipeline": ["torch", "torchvision", "typing"], "modelscope.pipelines.cv.vidt_pipeline": ["torch", "torchvision", "typing"], "modelscope.pipelines.cv.easycv_pipelines.base": ["numpy", "typing", "os", "easycv", "glob", "PIL"], "modelscope.pipelines.cv.easycv_pipelines.detection_pipeline": ["typing"], "modelscope.pipelines.cv.easycv_pipelines.segmentation_pipeline": ["numpy", "typing"], "modelscope.pipelines.cv.easycv_pipelines.face_2d_keypoints_pipeline": ["math", "numpy", "typing", "cv2", "copy"], "modelscope.pipelines.cv.easycv_pipelines.human_wholebody_keypoint_pipeline": ["os", "typing"], "modelscope.pipelines.cv.face_emotion_pipeline": ["numpy", "typing"], "modelscope.pipelines.cv.nerf_recon_acc_pipeline": ["typing"], "modelscope.pipelines.cv.shop_segmentation_pipleline": ["typing"], "modelscope.pipelines.cv.face_recognition_ood_pipeline": ["numpy", "typing", "cv2", "os", "torch", "PIL"], "modelscope.pipelines.multi_modal.document_vl_embedding_pipeline": ["torch", "typing"], "modelscope.pipelines.multi_modal.text2sql_pipeline": ["torch", "typing"], "modelscope.pipelines.multi_modal.visual_entailment_pipeline": ["torch", "typing"], "modelscope.pipelines.multi_modal.ocr_recognition_pipeline": ["torch", "typing"], "modelscope.pipelines.multi_modal.text_to_video_synthesis_pipeline": ["einops", "typing", "cv2", "tempfile", "torch"], "modelscope.pipelines.multi_modal.diffusers_wrapped.diffusers_pipeline": ["os", "typing"], "modelscope.pipelines.multi_modal.diffusers_wrapped.stable_diffusion.stable_diffusion_pipeline": ["diffusers", "numpy", "typing", "cv2", "torch", "PIL"], "modelscope.pipelines.multi_modal.diffusers_wrapped.stable_diffusion.chinese_stable_diffusion_pipeline": ["diffusers", "numpy", "transformers", "typing", "cv2", "torch", "PIL"], "modelscope.pipelines.multi_modal.sudoku_pipeline": ["torch", "typing"], "modelscope.pipelines.multi_modal.soonet_video_temporal_grounding_pipeline": ["numpy", "torchvision", "typing", "os", "torch"], "modelscope.pipelines.multi_modal.video_captioning_pipeline": ["torch", "typing"], "modelscope.pipelines.multi_modal.team_multi_modal_similarity_pipeline": ["typing"], "modelscope.pipelines.multi_modal.gridvlp_pipeline": ["time", "numpy", "json", "transformers", "typing", "os", "traceback", "torch", "PIL"], "modelscope.pipelines.multi_modal.mgeo_ranking_pipeline": ["torch", "numpy", "typing"], "modelscope.pipelines.multi_modal.generative_multi_modal_embedding_pipeline": ["typing"], "modelscope.pipelines.multi_modal.asr_pipeline": ["torch", "typing"], "modelscope.pipelines.multi_modal.visual_question_answering_pipeline": ["torch", "typing"], "modelscope.pipelines.multi_modal.text_to_image_synthesis_pipeline": ["torch", "typing"], "modelscope.pipelines.multi_modal.visual_grounding_pipeline": ["torch", "typing"], "modelscope.pipelines.multi_modal.image_captioning_pipeline": ["torch", "typing"], "modelscope.pipelines.multi_modal.multi_modal_embedding_pipeline": ["typing"], "modelscope.pipelines.multi_modal.video_multi_modal_embedding_pipeline": ["typing"], "modelscope.pipelines.multi_modal.disco_guided_diffusion_pipeline.disco_guided_diffusion": ["gc", "math", "numpy", "json", "torchvision", "cv2", "os", "importlib", "torch", "clip", "PIL"], "modelscope.pipelines.multi_modal.disco_guided_diffusion_pipeline.utils": ["fractions", "math", "torch", "warnings", "numpy"], "modelscope.pipelines.multi_modal.image_text_retrieval_pipeline": ["torch", "typing"], "modelscope.pipelines.multi_modal.video_question_answering_pipeline": ["torch", "typing"], "modelscope.preprocessors.builder": [], "modelscope.preprocessors.movie_scene_segmentation.transforms": ["numpy", "numbers", "torchvision", "typing", "os", "random", "torch", "PIL"], "modelscope.preprocessors.base": ["os", "typing", "abc"], "modelscope.preprocessors.image": ["io", "numpy", "cv2", "typing", "PIL"], "modelscope.preprocessors.audio": ["io", "scipy", "numpy", "typing", "os", "torch"], "modelscope.preprocessors.asr": ["os", "typing"], "modelscope.preprocessors.kws": ["os", "yaml", "typing"], "modelscope.preprocessors.multi_modal": ["io", "numpy", "json", "torchvision", "typing", "os", "decord", "torch", "timm", "PIL"], "modelscope.preprocessors.video": ["math", "uuid", "numpy", "torchvision", "os", "random", "urllib", "tempfile", "decord", "torch"], "modelscope.preprocessors.science.uni_fold": ["tarfile", "ipdb", "hashlib", "gzip", "re", "requests", "random", "time", "pathlib", "pickle", "numpy", "json", "unittest", "typing", "os", "torch", "logging", "tqdm"], "modelscope.preprocessors.nlp.transformers_tokenizer": ["os", "collections", "json", "transformers"], "modelscope.preprocessors.nlp.faq_question_answering_preprocessor": ["torch", "typing"], "modelscope.preprocessors.nlp.translation_evaluation_preprocessor": ["transformers", "typing"], "modelscope.preprocessors.nlp.document_grounded_dialog_retrieval_preprocessor": ["os", "torch", "transformers", "typing"], "modelscope.preprocessors.nlp.text_classification_preprocessor": ["numpy", "typing"], "modelscope.preprocessors.nlp.text_ranking_preprocessor": ["transformers", "typing"], "modelscope.preprocessors.nlp.token_classification_preprocessor": ["torch", "numpy", "typing"], "modelscope.preprocessors.nlp.document_segmentation_preprocessor": ["typing"], "modelscope.preprocessors.nlp.token_classification_thai_preprocessor": ["typing"], "modelscope.preprocessors.nlp.sentence_embedding_preprocessor": ["typing"], "modelscope.preprocessors.nlp.utils": ["numpy", "json", "transformers", "typing", "os", "collections"], "modelscope.preprocessors.nlp.feature_extraction_preprocessor": ["numpy", "typing"], "modelscope.preprocessors.nlp.bert_seq_cls_tokenizer": ["transformers", "typing"], "modelscope.preprocessors.nlp.document_grounded_dialog_generate_preprocessor": ["os", "torch", "transformers", "typing"], "modelscope.preprocessors.nlp.mgeo_ranking_preprocessor": ["torch", "transformers", "typing"], "modelscope.preprocessors.nlp.fill_mask_preprocessor": ["abc", "numpy", "re", "typing", "os", "torch"], "modelscope.preprocessors.nlp.space_T_en.fields.common_utils": ["itertools", "numpy", "text2sql_lgesql", "sqlite3", "os", "nltk"], "modelscope.preprocessors.nlp.space_T_en.fields.preprocess_dataset": ["text2sql_lgesql"], "modelscope.preprocessors.nlp.space_T_en.fields.process_dataset": ["os", "sys", "pickle", "text2sql_lgesql"], "modelscope.preprocessors.nlp.space_T_en.fields.parse": [], "modelscope.preprocessors.nlp.space_T_en.conversational_text_to_sql_preprocessor": ["json", "text2sql_lgesql", "typing", "os", "torch"], "modelscope.preprocessors.nlp.token_classification_viet_preprocessor": ["torch", "typing"], "modelscope.preprocessors.nlp.mglm_summarization_preprocessor": ["os", "re", "typing"], "modelscope.preprocessors.nlp.text_error_correction": ["os", "torch", "transformers", "typing"], "modelscope.preprocessors.nlp.zero_shot_classification_preprocessor": ["typing"], "modelscope.preprocessors.nlp.space_T_cn.fields.schema_link": ["re"], "modelscope.preprocessors.nlp.space_T_cn.fields.database": ["json", "sqlite3", "tqdm"], "modelscope.preprocessors.nlp.space_T_cn.fields.struct": [], "modelscope.preprocessors.nlp.space_T_cn.table_question_answering_preprocessor": ["os", "torch", "transformers", "typing"], "modelscope.preprocessors.nlp.text_generation_preprocessor": ["os", "numpy", "torch", "typing"], "modelscope.preprocessors.nlp.dialog_classification_use_preprocessor": ["torch", "transformers", "typing"], "modelscope.preprocessors.nlp.space.preprocess": ["os", "glob"], "modelscope.preprocessors.nlp.space.lazy_dataset": ["json"], "modelscope.preprocessors.nlp.space.data_loader": ["os", "numpy", "math"], "modelscope.preprocessors.nlp.space.fields.gen_field": ["itertools", "numpy", "json", "asyncio", "os", "random", "collections"], "modelscope.preprocessors.nlp.space.fields.intent_field": ["time", "itertools", "multiprocessing", "numpy", "json", "re", "random", "os", "collections", "glob", "tqdm"], "modelscope.preprocessors.nlp.space.sampler": ["numpy"], "modelscope.preprocessors.nlp.space.tensorlistdataset": ["torch"], "modelscope.preprocessors.nlp.space.batch": [], "modelscope.preprocessors.nlp.space.args": ["json", "argparse"], "modelscope.preprocessors.nlp.space.dst_processors": ["six", "logging", "numpy", "json", "re", "tqdm"], "modelscope.preprocessors.nlp.space.dialog_intent_prediction_preprocessor": ["os", "json", "typing"], "modelscope.preprocessors.nlp.space.dialog_modeling_preprocessor": ["os", "typing"], "modelscope.preprocessors.nlp.space.tokenizer": ["unicodedata", "regex", "json", "os", "functools", "sys", "__future__", "collections", "logging"], "modelscope.preprocessors.nlp.space.dialog_state_tracking_preprocessor": ["typing"], "modelscope.preprocessors.nlp.relation_extraction_preprocessor": ["transformers", "typing"], "modelscope.preprocessors.nlp.siamese_uie_preprocessor": ["transformers", "typing"], "modelscope.preprocessors.nlp.document_grounded_dialog_rerank_preprocessor": ["transformers", "typing", "os", "torch", "copy"], "modelscope.preprocessors.nlp.word_alignment_preprocessor": ["itertools", "numpy", "typing", "os", "torch"], "modelscope.preprocessors.cv.image_quality_assessment_mos": ["math", "numpy", "torchvision", "cv2", "typing"], "modelscope.preprocessors.cv.action_detection_mapper": ["random", "decord", "torch", "copy", "scipy", "numpy", "detectron2"], "modelscope.preprocessors.cv.controllable_image_generation": ["math", "numpy", "torchvision", "cv2", "typing", "os", "torch", "PIL"], "modelscope.preprocessors.cv.video_stabilization": ["cv2", "numpy", "torch"], "modelscope.preprocessors.cv.image_classification_preprocessor": ["numpy", "torchvision", "cv2", "typing", "os", "torch", "PIL"], "modelscope.preprocessors.cv.image_quality_assessment_man": ["math", "numpy", "torchvision", "typing", "torch", "PIL"], "modelscope.preprocessors.cv.bad_image_detecting_preprocessor": ["math", "numpy", "torchvision", "typing", "torch", "PIL"], "modelscope.preprocessors.cv.util": ["os", "sys", "shutil", "collections"], "modelscope.preprocessors.cv.video_super_resolution": ["os", "collections", "cv2"], "modelscope.preprocessors.cv.image_restoration_preprocessor": ["math", "numpy", "torchvision", "typing", "torch", "PIL"], "modelscope.preprocessors.cv.cv2_transforms": ["random", "math", "torch", "numpy", "collections", "numbers", "cv2"], "modelscope.preprocessors.cv.mmcls_preprocessor": ["os", "numpy", "typing"], "modelscope.preprocessors.cv.timer": ["time"], "modelscope.preprocessors.tts": ["os", "kantts", "typing"], "modelscope.preprocessors.ofa.image_classification": ["torchvision", "typing", "torch", "functools", "timm", "PIL"], "modelscope.preprocessors.ofa.text_classification": ["torch", "typing"], "modelscope.preprocessors.ofa.base": ["io", "numpy", "json", "re", "string", "os", "torch", "torchaudio", "PIL"], "modelscope.preprocessors.ofa.sudoku": ["torch", "numpy", "typing"], "modelscope.preprocessors.ofa.visual_question_answering": ["torch", "torchvision", "typing", "PIL"], "modelscope.preprocessors.ofa.asr": ["librosa", "pathlib", "soundfile", "fairseq", "typing", "os", "random", "torch"], "modelscope.preprocessors.ofa.text2sql": ["re", "typing", "os", "random", "torch"], "modelscope.preprocessors.ofa.summarization": ["torch", "typing"], "modelscope.preprocessors.ofa.ocr_recognition": ["torchvision", "typing", "zhconv", "unicodedata2", "torch"], "modelscope.preprocessors.ofa.visual_entailment": ["torch", "torchvision", "typing", "PIL"], "modelscope.preprocessors.ofa.utils.get_tables": ["sys", "traceback", "sqlite3"], "modelscope.preprocessors.ofa.utils.vision_helper": ["numpy", "cv2"], "modelscope.preprocessors.ofa.utils.text2phone": [], "modelscope.preprocessors.ofa.utils.bridge_content_encoder": ["functools", "rapidfuzz", "sqlite3", "typing", "difflib"], "modelscope.preprocessors.ofa.utils.constant": [], "modelscope.preprocessors.ofa.utils.audio_helper": ["torch", "numpy", "typing"], "modelscope.preprocessors.ofa.utils.random_help": ["torch_xla", "torch"], "modelscope.preprocessors.ofa.utils.collate": ["torch", "numpy", "typing"], "modelscope.preprocessors.ofa.utils.transforms": ["random", "torch", "numpy", "torchvision", "PIL"], "modelscope.preprocessors.ofa.visual_grounding": ["numpy", "torchvision", "typing", "torch", "PIL"], "modelscope.preprocessors.ofa.text_to_image_synthesis": ["torch", "typing"], "modelscope.preprocessors.ofa.image_captioning": ["torch", "torchvision", "typing"], "modelscope.preprocessors.common": ["time", "numpy", "typing", "torch", "collections"], "modelscope.trainers.lrscheduler.builder": ["inspect", "torch", "packaging"], "modelscope.trainers.lrscheduler.warmup.base": ["torch"], "modelscope.trainers.lrscheduler.warmup.warmup": [], "modelscope.trainers.optimizer.builder": ["torch", "inspect", "typing"], "modelscope.trainers.optimizer.child_tuning_adamw_optimizer": ["math", "numpy", "typing", "types", "torch"], "modelscope.trainers.builder": [], "modelscope.trainers.base": ["time", "os", "typing", "abc"], "modelscope.trainers.audio.kws_farfield_trainer": ["datetime", "pickle", "math", "numpy", "typing", "os", "torch", "glob"], "modelscope.trainers.audio.kws_utils.file_utils": ["re"], "modelscope.trainers.audio.kws_utils.runtime_utils": ["shutil", "json", "re", "os", "stat", "sys", "codecs", "collections"], "modelscope.trainers.audio.kws_utils.det_utils": ["matplotlib", "threading", "kaldiio", "json", "numpy", "os", "torch", "glob"], "modelscope.trainers.audio.kws_utils.model_utils": ["shutil", "numpy", "re", "yaml", "os", "torch", "glob"], "modelscope.trainers.audio.kws_utils.batch_utils": ["datetime", "math", "numpy", "typing", "os", "torch", "sys", "collections"], "modelscope.trainers.audio.ans_trainer": [], "modelscope.trainers.audio.separation_trainer": ["csv", "numpy", "typing", "os", "torch", "torchaudio", "speechbrain", "tqdm"], "modelscope.trainers.audio.asr_trainer": ["shutil", "json", "typing", "os", "tempfile", "funasr"], "modelscope.trainers.audio.tts_trainer": ["shutil", "json", "typing", "os", "zipfile", "tempfile"], "modelscope.trainers.audio.kws_nearfield_trainer": ["datetime", "tensorboardX", "re", "yaml", "typing", "os", "torch", "copy"], "modelscope.trainers.nlp_trainer": ["os", "numpy", "torch", "typing"], "modelscope.trainers.default_config": ["typing"], "modelscope.trainers.hooks.lr_scheduler_hook": [], "modelscope.trainers.hooks.iter_timer_hook": ["time"], "modelscope.trainers.hooks.optimizer.base": ["logging", "torch"], "modelscope.trainers.hooks.optimizer.torch_optimizer_hook": ["logging"], "modelscope.trainers.hooks.optimizer.apex_optimizer_hook": ["logging", "torch", "packaging"], "modelscope.trainers.hooks.hook": ["functools"], "modelscope.trainers.hooks.builder": [], "modelscope.trainers.hooks.megatron_hook": ["os", "torch", "megatron_util", "copy"], "modelscope.trainers.hooks.clip_clamp_logit_scale_hook": ["torch"], "modelscope.trainers.hooks.priority": ["enum", "typing"], "modelscope.trainers.hooks.ddp_hook": [], "modelscope.trainers.hooks.logger.base": ["numpy", "numbers", "torch", "abc"], "modelscope.trainers.hooks.logger.tensorboard_hook": ["os", "numpy", "torch"], "modelscope.trainers.hooks.logger.text_logger_hook": ["datetime", "json", "os", "torch", "collections"], "modelscope.trainers.hooks.evaluation_hook": ["collections"], "modelscope.trainers.hooks.checkpoint_hook": ["numpy", "re", "packaging", "os", "random", "torch"], "modelscope.trainers.hooks.early_stop_hook": ["numpy"], "modelscope.trainers.hooks.compression.utils": ["torch"], "modelscope.trainers.hooks.compression.sparsity_hook": ["os"], "modelscope.trainers.hooks.deepspeed_hook": ["shutil", "megatron_util", "deepspeed", "os", "torch"], "modelscope.trainers.parallel.builder": ["torch"], "modelscope.trainers.parallel.utils": [], "modelscope.trainers.nlp.document_grounded_dialog_generate_trainer": ["sacrebleu", "json", "re", "transformers", "string", "os", "torch", "rouge", "collections", "tqdm"], "modelscope.trainers.nlp.csanmt_translation_trainer": ["time", "tensorflow", "os", "typing"], "modelscope.trainers.nlp.document_grounded_dialog_rerank_trainer": ["time", "numpy", "transformers", "typing", "os", "random", "torch"], "modelscope.trainers.nlp.plug_trainer": ["megatron_util", "typing", "deepspeed", "os", "torch"], "modelscope.trainers.nlp.siamese_uie_trainer": ["time", "math", "numpy", "json", "typing", "os", "random", "torch", "collections"], "modelscope.trainers.nlp.document_grounded_dialog_retrieval_trainer": ["numpy", "json", "transformers", "os", "torch", "faiss", "tqdm"], "modelscope.trainers.nlp.gpt3_trainer": ["os", "torch", "typing", "copy"], "modelscope.trainers.nlp.table_question_answering_trainer": ["time", "numpy", "json", "typing", "os", "torch", "tqdm"], "modelscope.trainers.nlp.text_generation_trainer": ["collections", "torch"], "modelscope.trainers.nlp.gpt_moe_trainer": ["megatron_util", "typing", "os", "torch", "collections"], "modelscope.trainers.nlp.sequence_classification_trainer": ["time", "numpy", "typing"], "modelscope.trainers.nlp.sentence_embedding_trainer": ["time", "numpy", "transformers", "dataclasses", "typing", "torch", "tqdm"], "modelscope.trainers.nlp.space.eval": ["sklearn", "math", "numpy", "json", "nltk", "collections"], "modelscope.trainers.nlp.space.dialog_modeling_trainer": ["time", "os", "numpy", "typing"], "modelscope.trainers.nlp.space.trainer.intent_trainer": ["time", "numpy", "json", "transformers", "os", "torch", "collections", "tqdm"], "modelscope.trainers.nlp.space.trainer.gen_trainer": ["time", "numpy", "json", "transformers", "os", "torch", "collections", "tqdm"], "modelscope.trainers.nlp.space.metrics.metrics_tracker": ["collections", "math"], "modelscope.trainers.nlp.space.dialog_intent_trainer": ["os", "numpy", "typing"], "modelscope.trainers.nlp.text_ranking_trainer": ["time", "numpy", "dataclasses", "typing", "torch", "tqdm"], "modelscope.trainers.nlp.faq_question_answering_trainer": ["contextlib", "distutils", "numpy", "dataclasses", "typing", "torch", "functools", "collections"], "modelscope.trainers.utils.log_buffer": ["numpy", "collections"], "modelscope.trainers.utils.inference": ["shutil", "pickle", "os", "torch", "collections", "logging", "tqdm"], "modelscope.trainers.trainer": ["inspect", "distutils", "json", "typing", "os", "torch", "copy", "functools", "collections"], "modelscope.trainers.cv.image_defrcn_fewshot_detection_trainer": ["typing", "os", "torch", "collections", "detectron2"], "modelscope.trainers.cv.image_instance_segmentation_trainer": [], "modelscope.trainers.cv.ocr_recognition_trainer": ["time", "collections", "torch"], "modelscope.trainers.cv.card_detection_scrfd_trainer": [], "modelscope.trainers.cv.face_detection_scrfd_trainer": ["time", "os", "typing", "copy"], "modelscope.trainers.cv.image_inpainting_trainer": ["time", "collections", "torch"], "modelscope.trainers.cv.movie_scene_segmentation_trainer": [], "modelscope.trainers.cv.image_classifition_trainer": ["time", "numpy", "typing", "os", "torch", "copy"], "modelscope.trainers.cv.referring_video_object_segmentation_trainer": ["os", "torch"], "modelscope.trainers.cv.cartoon_translation_trainer": ["numpy", "typing", "packaging", "os", "tensorflow", "tqdm"], "modelscope.trainers.cv.action_detection_trainer": ["typing", "os", "fvcore", "torch", "detectron2"], "modelscope.trainers.cv.nerf_recon_acc_trainer": ["time", "datetime", "numpy", "cv2", "typing", "os", "random", "torch", "glob", "tqdm"], "modelscope.trainers.cv.ocr_detection_db_trainer": ["datetime", "time", "easydict", "math", "numpy", "typing", "os", "copy", "torch", "tqdm"], "modelscope.trainers.cv.vision_efficient_tuning_trainer": ["typing", "torch"], "modelscope.trainers.cv.image_portrait_enhancement_trainer": ["collections", "torch"], "modelscope.trainers.cv.image_detection_damoyolo_trainer": ["datetime", "time", "easydict", "math", "typing", "os", "torch"], "modelscope.trainers.training_args": ["re", "dataclasses", "typing", "functools", "argparse"], "modelscope.trainers.easycv.utils.hooks": [], "modelscope.trainers.easycv.utils.register_util": ["inspect", "logging"], "modelscope.trainers.easycv.utils.metric": ["torch", "numpy", "typing", "itertools"], "modelscope.trainers.easycv.trainer": ["typing", "easycv", "torch", "functools", "copy"], "modelscope.trainers.multi_modal.mplug.mplug_trainer": ["torch", "collections", "typing"], "modelscope.trainers.multi_modal.team.team_trainer": ["sklearn", "numpy", "typing", "os", "torch", "collections"], "modelscope.trainers.multi_modal.team.team_trainer_utils": ["torchvision", "torch", "PIL"], "modelscope.trainers.multi_modal.mgeo_ranking_trainer": ["torch", "dataclasses", "typing"], "modelscope.trainers.multi_modal.clip.clip_trainer": ["os", "torch", "math", "typing"], "modelscope.trainers.multi_modal.clip.clip_trainer_utils": ["inspect", "math", "os", "torch", "functools"], "modelscope.trainers.multi_modal.ofa.ofa_trainer_utils": ["os", "shutil", "math", "torch", "numpy", "transformers"], "modelscope.trainers.multi_modal.ofa.ofa_trainer": ["shutil", "math", "json", "typing", "os", "tempfile", "torch", "functools"], "modelscope.msdatasets.task_datasets.reds_image_deblurring_dataset": [], "modelscope.msdatasets.task_datasets.gopro_image_deblurring_dataset": [], "modelscope.msdatasets.task_datasets.sidd_image_denoising": [], "modelscope.msdatasets.task_datasets.video_summarization_dataset": [], "modelscope.msdatasets.task_datasets.torch_base_dataset": [], "modelscope.msdatasets.dataset_cls.custom_datasets.reds_image_deblurring_dataset": ["numpy", "cv2"], "modelscope.msdatasets.dataset_cls.custom_datasets.image_inpainting.aug": ["albumentations", "imgaug"], "modelscope.msdatasets.dataset_cls.custom_datasets.image_inpainting.image_inpainting_dataset": ["albumentations", "enum", "numpy", "cv2", "os", "glob"], "modelscope.msdatasets.dataset_cls.custom_datasets.builder": [], "modelscope.msdatasets.dataset_cls.custom_datasets.movie_scene_segmentation.sampler": ["random", "numpy"], "modelscope.msdatasets.dataset_cls.custom_datasets.movie_scene_segmentation.movie_scene_segmentation_dataset": ["json", "torchvision", "os", "random", "torch", "copy"], "modelscope.msdatasets.dataset_cls.custom_datasets.gopro_image_deblurring_dataset": ["numpy", "cv2"], "modelscope.msdatasets.dataset_cls.custom_datasets.image_portrait_enhancement.data_utils": ["cv2", "torch"], "modelscope.msdatasets.dataset_cls.custom_datasets.image_portrait_enhancement.image_portrait_enhancement_dataset": ["numpy", "cv2"], "modelscope.msdatasets.dataset_cls.custom_datasets.image_classification.classification_dataset": ["easycv"], "modelscope.msdatasets.dataset_cls.custom_datasets.language_guided_video_summarization_dataset": ["numpy", "json", "os", "torch", "h5py"], "modelscope.msdatasets.dataset_cls.custom_datasets.sidd_image_denoising.data_utils": ["cv2", "torch"], "modelscope.msdatasets.dataset_cls.custom_datasets.sidd_image_denoising.sidd_image_denoising_dataset": ["numpy", "cv2"], "modelscope.msdatasets.dataset_cls.custom_datasets.sidd_image_denoising.transforms": ["random"], "modelscope.msdatasets.dataset_cls.custom_datasets.referring_video_object_segmentation.referring_video_object_segmentation_dataset": ["pandas", "pycocotools", "numpy", "json", "torchvision", "os", "torch", "h5py", "glob", "tqdm"], "modelscope.msdatasets.dataset_cls.custom_datasets.referring_video_object_segmentation.transformers": ["random", "torchvision", "torch", "PIL"], "modelscope.msdatasets.dataset_cls.custom_datasets.image_quality_assmessment_mos.image_quality_assessment_mos_dataset": [], "modelscope.msdatasets.dataset_cls.custom_datasets.video_super_resolution.video_super_resolution_dataset": ["torch", "numpy", "collections", "cv2"], "modelscope.msdatasets.dataset_cls.custom_datasets.audio.kws_nearfield_processor": ["random", "torch", "torchaudio", "numpy", "kaldiio", "json", "logging"], "modelscope.msdatasets.dataset_cls.custom_datasets.audio.kws_farfield_dataset": ["math", "threading", "numpy", "os", "queue", "torch"], "modelscope.msdatasets.dataset_cls.custom_datasets.audio.asr_dataset": ["os"], "modelscope.msdatasets.dataset_cls.custom_datasets.audio.kws_nearfield_dataset": ["random", "torch"], "modelscope.msdatasets.dataset_cls.custom_datasets.object_detection.detection_dataset": ["easycv"], "modelscope.msdatasets.dataset_cls.custom_datasets.image_instance_segmentation_coco_dataset": ["os", "numpy", "pycocotools"], "modelscope.msdatasets.dataset_cls.custom_datasets.image_colorization.image_colorization_dataset": ["cv2", "numpy", "torch"], "modelscope.msdatasets.dataset_cls.custom_datasets.mgeo_ranking_dataset": ["random", "torch", "json", "typing"], "modelscope.msdatasets.dataset_cls.custom_datasets.image_quality_assessment_degradation.image_quality_assessment_degradation_dataset": ["torchvision"], "modelscope.msdatasets.dataset_cls.custom_datasets.damoyolo.evaluation.coco.coco_eval": ["os", "collections", "tempfile", "torch"], "modelscope.msdatasets.dataset_cls.custom_datasets.damoyolo.samplers.iteration_based_batch_sampler": ["torch"], "modelscope.msdatasets.dataset_cls.custom_datasets.damoyolo.samplers.grouped_batch_sampler": ["itertools", "torch"], "modelscope.msdatasets.dataset_cls.custom_datasets.damoyolo.samplers.distributed": ["math", "torch"], "modelscope.msdatasets.dataset_cls.custom_datasets.damoyolo.collate_batch": [], "modelscope.msdatasets.dataset_cls.custom_datasets.damoyolo.transforms.build": [], "modelscope.msdatasets.dataset_cls.custom_datasets.damoyolo.transforms.transforms": ["random", "torch", "numpy", "torchvision", "cv2"], "modelscope.msdatasets.dataset_cls.custom_datasets.damoyolo.datasets.mosaic_wrapper": ["math", "numpy", "cv2", "random", "torch"], "modelscope.msdatasets.dataset_cls.custom_datasets.damoyolo.datasets.coco": ["torch", "numpy", "torchvision", "cv2"], "modelscope.msdatasets.dataset_cls.custom_datasets.damoyolo.build": ["bisect", "math", "torch", "copy"], "modelscope.msdatasets.dataset_cls.custom_datasets.torch_custom_dataset": ["torch", "typing"], "modelscope.msdatasets.dataset_cls.custom_datasets.ocr_detection.processes.data_process": [], "modelscope.msdatasets.dataset_cls.custom_datasets.ocr_detection.processes.make_seg_detection_data": ["numpy", "pyclipper", "shapely", "cv2"], "modelscope.msdatasets.dataset_cls.custom_datasets.ocr_detection.processes.make_icdar_data": ["torch", "numpy", "collections", "cv2"], "modelscope.msdatasets.dataset_cls.custom_datasets.ocr_detection.processes.augment_data": ["imgaug", "numpy", "math", "cv2"], "modelscope.msdatasets.dataset_cls.custom_datasets.ocr_detection.processes.make_border_map": ["numpy", "pyclipper", "shapely", "cv2"], "modelscope.msdatasets.dataset_cls.custom_datasets.ocr_detection.processes.random_crop_data": ["numpy", "cv2"], "modelscope.msdatasets.dataset_cls.custom_datasets.ocr_detection.processes.normalize_image": ["numpy", "torch"], "modelscope.msdatasets.dataset_cls.custom_datasets.ocr_detection.data_loader": ["bisect", "math", "numpy", "imgaug", "torch"], "modelscope.msdatasets.dataset_cls.custom_datasets.ocr_detection.image_dataset": ["bisect", "math", "numpy", "cv2", "os", "functools", "torch", "logging", "glob"], "modelscope.msdatasets.dataset_cls.custom_datasets.ocr_detection.measures.quad_measurer": ["numpy"], "modelscope.msdatasets.dataset_cls.custom_datasets.ocr_detection.measures.iou_evaluator": ["numpy", "collections", "shapely"], "modelscope.msdatasets.dataset_cls.custom_datasets.ocr_detection.augmenter": ["imgaug"], "modelscope.msdatasets.dataset_cls.custom_datasets.face_2d_keypoins.face_2d_keypoints_dataset": ["easycv"], "modelscope.msdatasets.dataset_cls.custom_datasets.text_ranking_dataset": ["random", "torch", "typing"], "modelscope.msdatasets.dataset_cls.custom_datasets.bad_image_detecting.bad_image_detecting_dataset": [], "modelscope.msdatasets.dataset_cls.custom_datasets.video_frame_interpolation.video_frame_interpolation_dataset": ["cv2", "numpy", "torch"], "modelscope.msdatasets.dataset_cls.custom_datasets.video_frame_interpolation.data_utils": ["cv2", "torch"], "modelscope.msdatasets.dataset_cls.custom_datasets.video_summarization_dataset": ["numpy", "json", "os", "torch", "h5py"], "modelscope.msdatasets.dataset_cls.custom_datasets.veco_dataset": ["numpy", "datasets", "typing"], "modelscope.msdatasets.dataset_cls.custom_datasets.human_wholebody_keypoint.human_wholebody_keypoint_dataset": ["easycv"], "modelscope.msdatasets.dataset_cls.custom_datasets.image_semantic_segmentation.segmentation_dataset": ["easycv"], "modelscope.msdatasets.dataset_cls.custom_datasets.hand_2d_keypoints.hand_2d_keypoints_dataset": ["easycv"], "modelscope.msdatasets.dataset_cls.custom_datasets.easycv_base": ["os"], "modelscope.msdatasets.dataset_cls.custom_datasets.ocr_recognition_dataset": ["six", "numpy", "json", "cv2", "os", "torch", "lmdb", "PIL"], "modelscope.msdatasets.dataset_cls.custom_datasets.video_stabilization.video_stabilization_dataset": [], "modelscope.msdatasets.dataset_cls.dataset": ["os", "datasets", "copy", "PIL"], "modelscope.msdatasets.ms_dataset": ["numpy", "typing", "os", "datasets", "warnings", "tensorflow"], "modelscope.msdatasets.audio.asr_dataset": [], "modelscope.msdatasets.meta.data_meta_config": [], "modelscope.msdatasets.meta.data_meta_manager": ["shutil", "json", "os", "datasets", "collections"], "modelscope.msdatasets.data_loader.data_loader": ["datasets", "typing", "abc"], "modelscope.msdatasets.data_loader.data_loader_manager": ["os", "enum", "datasets", "abc"], "modelscope.msdatasets.utils.upload_utils": ["os", "multiprocessing", "tqdm"], "modelscope.msdatasets.utils.delete_utils": [], "modelscope.msdatasets.utils.oss_utils": ["oss2", "multiprocessing", "os", "datasets", "__future__"], "modelscope.msdatasets.utils.dataset_utils": ["os", "collections", "typing"], "modelscope.msdatasets.auth.auth_config": ["http", "typing"], "modelscope.msdatasets.download.download_config": ["datasets", "typing"], "modelscope.msdatasets.download.download_manager": ["datasets"], "modelscope.msdatasets.download.dataset_builder": ["pandas", "pyarrow", "typing", "os", "datasets"], "modelscope.msdatasets.context.dataset_context_config": ["typing"], "modelscope.msdatasets.data_files.data_files_manager": ["os", "datasets", "typing"], "modelscope.exporters.builder": [], "modelscope.exporters.base": ["os", "typing", "abc"], "modelscope.exporters.tf_model_exporter": ["os", "tensorflow", "typing"], "modelscope.exporters.nlp.csanmt_for_translation_exporter": ["os", "tensorflow", "typing"], "modelscope.exporters.nlp.sbert_for_sequence_classification_exporter": ["collections", "typing", "torch"], "modelscope.exporters.nlp.model_for_token_classification_exporter": ["torch", "collections", "typing"], "modelscope.exporters.nlp.sbert_for_zero_shot_classification_exporter": ["collections", "typing"], "modelscope.exporters.cv.object_detection_damoyolo_exporter": ["numpy", "typing", "os", "functools", "torch", "onnx"], "modelscope.exporters.cv.cartoon_translation_exporter": ["os", "tensorflow", "typing", "packaging"], "modelscope.exporters.cv.face_detection_scrfd_exporter": ["numpy", "typing", "os", "functools", "torch", "onnx"], "modelscope.exporters.torch_model_exporter": ["contextlib", "itertools", "typing", "os", "torch"]}, "version": "1.4.1", "md5": "ba0c639b3af75c469d4dce8b8c960f4c", "files_mtime": {"/usr/local/lib/python3.9/dist-packages/modelscope/models/audio/punc/generic_punctuation.py": 1679246024.068371, "/usr/local/lib/python3.9/dist-packages/modelscope/models/audio/sv/generic_speaker_verification.py": 1679246024.0703714, "/usr/local/lib/python3.9/dist-packages/modelscope/models/audio/sv/ecapa_tdnn.py": 1679246024.0693712, "/usr/local/lib/python3.9/dist-packages/modelscope/models/audio/itn/generic_inverse_text_processing.py": 1679246024.0653703, "/usr/local/lib/python3.9/dist-packages/modelscope/models/audio/aec/layers/affine_transform.py": 1679246024.0613694, "/usr/local/lib/python3.9/dist-packages/modelscope/models/audio/aec/layers/uni_deep_fsmn.py": 1679246024.0613694, "/usr/local/lib/python3.9/dist-packages/modelscope/models/audio/aec/layers/deep_fsmn.py": 1679246024.0613694, "/usr/local/lib/python3.9/dist-packages/modelscope/models/audio/aec/layers/activations.py": 1679246024.0613694, "/usr/local/lib/python3.9/dist-packages/modelscope/models/audio/aec/layers/layer_base.py": 1679246024.0613694, "/usr/local/lib/python3.9/dist-packages/modelscope/models/audio/aec/network/modulation_loss.py": 1679246024.0623698, "/usr/local/lib/python3.9/dist-packages/modelscope/models/audio/aec/network/loss.py": 1679246024.0623698, "/usr/local/lib/python3.9/dist-packages/modelscope/models/audio/aec/network/se_net.py": 1679246024.0623698, "/usr/local/lib/python3.9/dist-packages/modelscope/models/audio/separation/mossformer.py": 1679246024.068371, "/usr/local/lib/python3.9/dist-packages/modelscope/models/audio/separation/layer_norm.py": 1679246024.068371, "/usr/local/lib/python3.9/dist-packages/modelscope/models/audio/separation/mossformer_conv_module.py": 1679246024.0693712, "/usr/local/lib/python3.9/dist-packages/modelscope/models/audio/separation/mossformer_block.py": 1679246024.0693712, "/usr/local/lib/python3.9/dist-packages/modelscope/models/audio/kws/farfield/model.py": 1679246024.0673707, "/usr/local/lib/python3.9/dist-packages/modelscope/models/audio/kws/farfield/model_def.py": 1679246024.0673707, "/usr/local/lib/python3.9/dist-packages/modelscope/models/audio/kws/farfield/fsmn.py": 1679246024.0663705, "/usr/local/lib/python3.9/dist-packages/modelscope/models/audio/kws/farfield/fsmn_sele_v2.py": 1679246024.0663705, "/usr/local/lib/python3.9/dist-packages/modelscope/models/audio/kws/nearfield/model.py": 1679246024.068371, "/usr/local/lib/python3.9/dist-packages/modelscope/models/audio/kws/nearfield/cmvn.py": 1679246024.0673707, "/usr/local/lib/python3.9/dist-packages/modelscope/models/audio/kws/nearfield/fsmn.py": 1679246024.0673707, "/usr/local/lib/python3.9/dist-packages/modelscope/models/audio/kws/generic_key_word_spotting.py": 1679246024.0663705, "/usr/local/lib/python3.9/dist-packages/modelscope/models/audio/ans/se_module_complex.py": 1679246024.06337, "/usr/local/lib/python3.9/dist-packages/modelscope/models/audio/ans/conv_stft.py": 1679246024.06337, "/usr/local/lib/python3.9/dist-packages/modelscope/models/audio/ans/denoise_net.py": 1679246024.06337, "/usr/local/lib/python3.9/dist-packages/modelscope/models/audio/ans/frcrn.py": 1679246024.06337, "/usr/local/lib/python3.9/dist-packages/modelscope/models/audio/ans/complex_nn.py": 1679246024.06337, "/usr/local/lib/python3.9/dist-packages/modelscope/models/audio/ans/layers/affine_transform.py": 1679246024.0643702, "/usr/local/lib/python3.9/dist-packages/modelscope/models/audio/ans/layers/uni_deep_fsmn.py": 1679246024.0653703, "/usr/local/lib/python3.9/dist-packages/modelscope/models/audio/ans/layers/activations.py": 1679246024.0643702, "/usr/local/lib/python3.9/dist-packages/modelscope/models/audio/ans/layers/layer_base.py": 1679246024.0643702, "/usr/local/lib/python3.9/dist-packages/modelscope/models/audio/ans/unet.py": 1679246024.0643702, "/usr/local/lib/python3.9/dist-packages/modelscope/models/audio/tts/voice.py": 1679246024.0703714, "/usr/local/lib/python3.9/dist-packages/modelscope/models/audio/tts/sambert_hifi.py": 1679246024.0703714, "/usr/local/lib/python3.9/dist-packages/modelscope/models/audio/asr/wenet_automatic_speech_recognition.py": 1679246024.0653703, "/usr/local/lib/python3.9/dist-packages/modelscope/models/audio/asr/generic_automatic_speech_recognition.py": 1679246024.0653703, "/usr/local/lib/python3.9/dist-packages/modelscope/models/science/unifold/model.py": 1679246024.3844364, "/usr/local/lib/python3.9/dist-packages/modelscope/models/science/unifold/msa/parsers.py": 1679246024.3904376, "/usr/local/lib/python3.9/dist-packages/modelscope/models/science/unifold/msa/msa_identifiers.py": 1679246024.3904376, "/usr/local/lib/python3.9/dist-packages/modelscope/models/science/unifold/msa/tools/kalign.py": 1679246024.392438, "/usr/local/lib/python3.9/dist-packages/modelscope/models/science/unifold/msa/tools/hhblits.py": 1679246024.3914378, "/usr/local/lib/python3.9/dist-packages/modelscope/models/science/unifold/msa/tools/utils.py": 1679246024.393438, "/usr/local/lib/python3.9/dist-packages/modelscope/models/science/unifold/msa/tools/hmmsearch.py": 1679246024.392438, "/usr/local/lib/python3.9/dist-packages/modelscope/models/science/unifold/msa/tools/hhsearch.py": 1679246024.392438, "/usr/local/lib/python3.9/dist-packages/modelscope/models/science/unifold/msa/tools/jackhmmer.py": 1679246024.392438, "/usr/local/lib/python3.9/dist-packages/modelscope/models/science/unifold/msa/tools/hmmbuild.py": 1679246024.392438, "/usr/local/lib/python3.9/dist-packages/modelscope/models/science/unifold/msa/utils.py": 1679246024.3914378, "/usr/local/lib/python3.9/dist-packages/modelscope/models/science/unifold/msa/templates.py": 1679246024.3914378, "/usr/local/lib/python3.9/dist-packages/modelscope/models/science/unifold/msa/mmcif.py": 1679246024.3904376, "/usr/local/lib/python3.9/dist-packages/modelscope/models/science/unifold/msa/pipeline.py": 1679246024.3904376, "/usr/local/lib/python3.9/dist-packages/modelscope/models/science/unifold/dataset.py": 1679246024.3844364, "/usr/local/lib/python3.9/dist-packages/modelscope/models/science/unifold/config.py": 1679246024.3844364, "/usr/local/lib/python3.9/dist-packages/modelscope/models/science/unifold/modules/alphafold.py": 1679246024.3874369, "/usr/local/lib/python3.9/dist-packages/modelscope/models/science/unifold/modules/confidence.py": 1679246024.3874369, "/usr/local/lib/python3.9/dist-packages/modelscope/models/science/unifold/modules/frame.py": 1679246024.388437, "/usr/local/lib/python3.9/dist-packages/modelscope/models/science/unifold/modules/triangle_multiplication.py": 1679246024.3894374, "/usr/local/lib/python3.9/dist-packages/modelscope/models/science/unifold/modules/auxillary_heads.py": 1679246024.3874369, "/usr/local/lib/python3.9/dist-packages/modelscope/models/science/unifold/modules/attentions.py": 1679246024.3874369, "/usr/local/lib/python3.9/dist-packages/modelscope/models/science/unifold/modules/template.py": 1679246024.3894374, "/usr/local/lib/python3.9/dist-packages/modelscope/models/science/unifold/modules/structure_module.py": 1679246024.3894374, "/usr/local/lib/python3.9/dist-packages/modelscope/models/science/unifold/modules/common.py": 1679246024.3874369, "/usr/local/lib/python3.9/dist-packages/modelscope/models/science/unifold/modules/embedders.py": 1679246024.388437, "/usr/local/lib/python3.9/dist-packages/modelscope/models/science/unifold/modules/featurization.py": 1679246024.388437, "/usr/local/lib/python3.9/dist-packages/modelscope/models/science/unifold/modules/evoformer.py": 1679246024.388437, "/usr/local/lib/python3.9/dist-packages/modelscope/models/science/unifold/data/protein.py": 1679246024.3864367, "/usr/local/lib/python3.9/dist-packages/modelscope/models/science/unifold/data/process.py": 1679246024.3854365, "/usr/local/lib/python3.9/dist-packages/modelscope/models/science/unifold/data/utils.py": 1679246024.3864367, "/usr/local/lib/python3.9/dist-packages/modelscope/models/science/unifold/data/residue_constants.py": 1679246024.3864367, "/usr/local/lib/python3.9/dist-packages/modelscope/models/science/unifold/data/process_multimer.py": 1679246024.3854365, "/usr/local/lib/python3.9/dist-packages/modelscope/models/science/unifold/data/msa_pairing.py": 1679246024.3854365, "/usr/local/lib/python3.9/dist-packages/modelscope/models/science/unifold/data/data_ops.py": 1679246024.3854365, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/ofa/modeling_ofa.py": 1679246024.3284247, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/ofa/resnet.py": 1679246024.3284247, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/ofa/configuration_ofa.py": 1679246024.3274245, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/ofa/tokenization_ofa.py": 1679246024.3284247, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/ofa/vit.py": 1679246024.3294249, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/ofa/utils/constant.py": 1679246024.3314254, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/ofa/utils/utils.py": 1679246024.3324256, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/ofa/modeling_mmspeech.py": 1679246024.3274245, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/ofa/tokenization_ofa_fast.py": 1679246024.3284247, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/ofa/configuration_mmspeech.py": 1679246024.3264244, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/ofa/generate/search.py": 1679246024.330425, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/ofa/generate/incremental_decoding_utils.py": 1679246024.3294249, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/ofa/generate/sequence_generator.py": 1679246024.330425, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/ofa/generate/utils.py": 1679246024.3314254, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/ofa/generate/ngram_repeat_block.py": 1679246024.330425, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/ofa/generate/token_generation_constraints.py": 1679246024.3314254, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/ofa/generate/multihead_attention.py": 1679246024.3294249, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/vldoc/model.py": 1679246024.3374267, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/vldoc/modeling_layout_roberta.py": 1679246024.3374267, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/vldoc/transformer_local.py": 1679246024.3384268, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/vldoc/processing.py": 1679246024.3374267, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/vldoc/convnext.py": 1679246024.3364263, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/vldoc/tokenization.py": 1679246024.3374267, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/vldoc/conv_fpn_trans.py": 1679246024.3364263, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/dpm_solver_pytorch.py": 1679246024.310421, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/video_synthesis/text_to_video_synthesis_model.py": 1679246024.335426, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/video_synthesis/unet_sd.py": 1679246024.3364263, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/video_synthesis/diffusion.py": 1679246024.335426, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/video_synthesis/autoencoder.py": 1679246024.335426, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/diffusion/model.py": 1679246024.3134217, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/diffusion/unet_generator.py": 1679246024.314422, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/diffusion/unet_upsampler_1024.py": 1679246024.314422, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/diffusion/structbert.py": 1679246024.3134217, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/diffusion/tokenizer.py": 1679246024.314422, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/diffusion/unet_upsampler_256.py": 1679246024.314422, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/diffusion/diffusion.py": 1679246024.3134217, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/ofa_for_all_tasks.py": 1679246024.3114212, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/mgeo/text_classification.py": 1679246024.3184226, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/mgeo/text_ranking.py": 1679246024.3184226, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/mgeo/token_classification.py": 1679246024.3184226, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/mgeo/backbone.py": 1679246024.3174224, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/gemm/gemm_model.py": 1679246024.315422, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/gemm/gemm_base.py": 1679246024.315422, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/gemm/tokenizer.py": 1679246024.315422, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/ofa_for_text_to_image_synthesis_model.py": 1679246024.3114212, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/mmr/models/clip_for_mm_video_embedding.py": 1679246024.319423, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/mmr/models/module_clip.py": 1679246024.3204231, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/mmr/models/module_cross.py": 1679246024.3204231, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/mmr/models/dynamic_inverted_softmax.py": 1679246024.319423, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/mmr/models/tokenization_clip.py": 1679246024.3214233, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/mmr/models/until_module.py": 1679246024.3214233, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/mmr/models/modeling.py": 1679246024.319423, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/mmr/dataloaders/rawvideo_util.py": 1679246024.319423, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/rleg/model.py": 1679246024.3324256, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/rleg/rleg.py": 1679246024.3324256, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/guided_diffusion/gaussian_diffusion.py": 1679246024.3164222, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/guided_diffusion/respace.py": 1679246024.3164222, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/guided_diffusion/script.py": 1679246024.3164222, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/guided_diffusion/unet.py": 1679246024.3174224, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/soonet/model.py": 1679246024.3334258, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/soonet/utils.py": 1679246024.334426, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/soonet/clip.py": 1679246024.3334258, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/soonet/blocks.py": 1679246024.3324256, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/soonet/swin_transformer.py": 1679246024.3334258, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/soonet/tokenizer.py": 1679246024.3334258, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/mplug/modeling_mplug.py": 1679246024.3224235, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/mplug/predictor.py": 1679246024.3234236, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/mplug/configuration_mplug.py": 1679246024.3224235, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/mplug/mvit.py": 1679246024.3234236, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/mplug/clip/clip.py": 1679246024.3244238, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/team/team_model.py": 1679246024.334426, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/team/utils.py": 1679246024.334426, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/multi_stage_diffusion/decoder.py": 1679246024.3244238, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/multi_stage_diffusion/model.py": 1679246024.3254242, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/multi_stage_diffusion/upsampler.py": 1679246024.3264244, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/multi_stage_diffusion/gaussian_diffusion.py": 1679246024.3254242, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/multi_stage_diffusion/xglm.py": 1679246024.3264244, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/multi_stage_diffusion/clip.py": 1679246024.3244238, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/multi_stage_diffusion/tokenizer.py": 1679246024.3254242, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/multi_stage_diffusion/prior.py": 1679246024.3254242, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/mplug_for_all_tasks.py": 1679246024.310421, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/clip/model.py": 1679246024.3124213, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/clip/configuration_bert.py": 1679246024.3124213, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/clip/bert_tokenizer.py": 1679246024.3114212, "/usr/local/lib/python3.9/dist-packages/modelscope/models/multi_modal/clip/modeling_bert.py": 1679246024.3124213, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/veco/text_classification.py": 1679246024.3824358, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/veco/token_classification.py": 1679246024.3824358, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/veco/backbone.py": 1679246024.3814356, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/veco/fill_mask.py": 1679246024.3824358, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/veco/configuration.py": 1679246024.3824358, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/plug_mental/text_classification.py": 1679246024.3704333, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/plug_mental/adv_utils.py": 1679246024.3694332, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/plug_mental/backbone.py": 1679246024.3704333, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/plug_mental/configuration.py": 1679246024.3704333, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/gpt_neo/backbone.py": 1679246024.3534298, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/lstm/token_classification.py": 1679246024.3564305, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/lstm/backbone.py": 1679246024.3564305, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/codegeex/codegeex_for_code_generation.py": 1679246024.343428, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/codegeex/codegeex.py": 1679246024.343428, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/codegeex/codegeex_for_code_translation.py": 1679246024.343428, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/codegeex/tokenizer.py": 1679246024.344428, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/codegeex/inference.py": 1679246024.343428, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/csanmt/translation.py": 1679246024.344428, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/T5/backbone.py": 1679246024.339427, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/T5/text2text_generation.py": 1679246024.339427, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/T5/configuration.py": 1679246024.339427, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/bart/text_error_correction.py": 1679246024.3404272, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/gpt3/distributed_gpt3.py": 1679246024.3494291, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/gpt3/text_generation.py": 1679246024.3504293, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/gpt3/tokenizer.py": 1679246024.3504293, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/gpt3/backbone.py": 1679246024.3494291, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/gpt3/configuration.py": 1679246024.3494291, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/dgds/document_grounded_dialog_rerank.py": 1679246024.3474286, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/dgds/document_grounded_dialog_generate.py": 1679246024.3464284, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/dgds/document_grounded_dialog_retrieval.py": 1679246024.3474286, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/dgds/backbone.py": 1679246024.3464284, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/task_models/text_classification.py": 1679246024.3794353, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/task_models/text_ranking.py": 1679246024.3804355, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/task_models/information_extraction.py": 1679246024.3794353, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/task_models/text_generation.py": 1679246024.3804355, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/task_models/task_model.py": 1679246024.3794353, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/task_models/token_classification.py": 1679246024.3804355, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/task_models/feature_extraction.py": 1679246024.3794353, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/task_models/fill_mask.py": 1679246024.3794353, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/space/dialog_intent_prediction.py": 1679246024.372434, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/space/dialog_modeling.py": 1679246024.372434, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/space/model/tokenization_space.py": 1679246024.3744342, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/space/model/unified_transformer.py": 1679246024.3744342, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/space/model/gen_unified_transformer.py": 1679246024.373434, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/space/model/generator.py": 1679246024.373434, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/space/model/model_base.py": 1679246024.373434, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/space/model/intent_unified_transformer.py": 1679246024.373434, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/space/dialog_state_tracking.py": 1679246024.372434, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/space/modules/functions.py": 1679246024.3744342, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/space/modules/feedforward.py": 1679246024.3744342, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/space/modules/transformer_block.py": 1679246024.3754344, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/space/modules/embedder.py": 1679246024.3744342, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/space/modules/multihead_attention.py": 1679246024.3754344, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/space/configuration.py": 1679246024.372434, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/fid_plug/text_generation.py": 1679246024.3484287, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/fid_plug/backbone.py": 1679246024.3474286, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/fid_plug/configuration.py": 1679246024.3484287, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/space_T_en/text_to_sql.py": 1679246024.3764346, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/bloom/backbone.py": 1679246024.3424275, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/structbert/text_classification.py": 1679246024.3784351, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/structbert/adv_utils.py": 1679246024.3774347, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/structbert/token_classification.py": 1679246024.3784351, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/structbert/faq_question_answering.py": 1679246024.3774347, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/structbert/backbone.py": 1679246024.3774347, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/structbert/fill_mask.py": 1679246024.3784351, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/structbert/configuration.py": 1679246024.3774347, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/heads/text_classification_head.py": 1679246024.3544302, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/heads/token_classification_head.py": 1679246024.3554304, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/heads/torch_pretrain_head.py": 1679246024.3554304, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/heads/fill_mask_head.py": 1679246024.3544302, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/heads/text_ranking_head.py": 1679246024.3544302, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/heads/infromation_extraction_head.py": 1679246024.3544302, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/heads/crf_head.py": 1679246024.3544302, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/heads/text_generation_head.py": 1679246024.3544302, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/mglm/process_grid.py": 1679246024.359431, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/mglm/train_utils.py": 1679246024.359431, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/mglm/configure_data.py": 1679246024.3584309, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/mglm/model/modeling_glm.py": 1679246024.364432, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/mglm/model/downstream.py": 1679246024.363432, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/mglm/model/transformer.py": 1679246024.364432, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/mglm/model/modeling_bert.py": 1679246024.363432, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/mglm/model/prompt.py": 1679246024.364432, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/mglm/model/distributed.py": 1679246024.363432, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/mglm/utils.py": 1679246024.359431, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/mglm/data_utils/samplers.py": 1679246024.3614316, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/mglm/data_utils/tokenization_gpt2.py": 1679246024.3624318, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/mglm/data_utils/lazy_loader.py": 1679246024.3614316, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/mglm/data_utils/sp_tokenizer.py": 1679246024.3614316, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/mglm/data_utils/wordpiece.py": 1679246024.3624318, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/mglm/data_utils/tokenization.py": 1679246024.3624318, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/mglm/data_utils/extraction.py": 1679246024.3614316, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/mglm/data_utils/corpora.py": 1679246024.3604314, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/mglm/data_utils/file_utils.py": 1679246024.3614316, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/mglm/data_utils/datasets.py": 1679246024.3604314, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/mglm/arguments.py": 1679246024.3574307, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/mglm/blocklm_utils.py": 1679246024.3584309, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/mglm/generation_utils.py": 1679246024.3584309, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/mglm/run_test.py": 1679246024.359431, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/mglm/test/test_rel_shift.py": 1679246024.3654323, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/mglm/test/test_block.py": 1679246024.3654323, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/mglm/mglm_for_text_summarization.py": 1679246024.359431, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/hf_transformers/backbone.py": 1679246024.3554304, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/megatron_bert/backbone.py": 1679246024.3564305, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/megatron_bert/fill_mask.py": 1679246024.3574307, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/megatron_bert/configuration.py": 1679246024.3574307, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/ponet/tokenization.py": 1679246024.3714335, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/ponet/document_segmentation.py": 1679246024.3714335, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/ponet/backbone.py": 1679246024.3704333, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/ponet/fill_mask.py": 1679246024.3714335, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/ponet/configuration.py": 1679246024.3714335, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/unite/configuration_unite.py": 1679246024.3804355, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/unite/modeling_unite.py": 1679246024.3814356, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/plug/distributed_plug.py": 1679246024.3694332, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/plug/generator.py": 1679246024.3694332, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/plug/backbone.py": 1679246024.368433, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/plug/AnnealingLR.py": 1679246024.368433, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/plug/configuration.py": 1679246024.368433, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/gpt2/backbone.py": 1679246024.3484287, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/deberta_v2/tokenization_fast.py": 1679246024.3464284, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/deberta_v2/tokenization.py": 1679246024.3464284, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/deberta_v2/backbone.py": 1679246024.3454282, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/deberta_v2/fill_mask.py": 1679246024.3454282, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/deberta_v2/configuration.py": 1679246024.3454282, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/gpt_moe/distributed_gpt_moe.py": 1679246024.3514295, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/gpt_moe/moe/sharded_moe.py": 1679246024.3524296, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/gpt_moe/moe/layer.py": 1679246024.3524296, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/gpt_moe/moe/mappings.py": 1679246024.3524296, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/gpt_moe/moe/utils.py": 1679246024.3534298, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/gpt_moe/moe/experts.py": 1679246024.3524296, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/gpt_moe/text_generation.py": 1679246024.3514295, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/gpt_moe/checkpointing.py": 1679246024.3504293, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/gpt_moe/tokenizer.py": 1679246024.3514295, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/gpt_moe/backbone.py": 1679246024.3504293, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/gpt_moe/configuration.py": 1679246024.3514295, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/palm_v2/text_generation.py": 1679246024.3664327, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/palm_v2/dureader_eval.py": 1679246024.3654323, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/palm_v2/configuration.py": 1679246024.3654323, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/peer/text_classification.py": 1679246024.3674328, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/peer/sas_utils.py": 1679246024.3674328, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/peer/backbone.py": 1679246024.3674328, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/peer/configuration.py": 1679246024.3674328, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/bert/text_classification.py": 1679246024.3414273, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/bert/sentence_embedding.py": 1679246024.3414273, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/bert/word_alignment.py": 1679246024.3424275, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/bert/siamese_uie.py": 1679246024.3414273, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/bert/text_ranking.py": 1679246024.3414273, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/bert/token_classification.py": 1679246024.3424275, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/bert/document_segmentation.py": 1679246024.3414273, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/bert/backbone.py": 1679246024.3404272, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/bert/fill_mask.py": 1679246024.3414273, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/bert/configuration.py": 1679246024.3404272, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/use/user_satisfaction_estimation.py": 1679246024.3814356, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/use/transformer.py": 1679246024.3814356, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/space_T_cn/table_question_answering.py": 1679246024.3764346, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/space_T_cn/backbone.py": 1679246024.3754344, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/space_T_cn/configuration.py": 1679246024.3764346, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/xlm_roberta/backbone.py": 1679246024.383436, "/usr/local/lib/python3.9/dist-packages/modelscope/models/nlp/xlm_roberta/configuration.py": 1679246024.383436, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/vidt/model.py": 1679246024.30542, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/vidt/fpn_fusion.py": 1679246024.3044198, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/vidt/head.py": 1679246024.30542, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/vidt/deformable_transformer.py": 1679246024.3044198, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/vidt/backbone.py": 1679246024.3044198, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_emotion/emotion_model.py": 1679246024.1043785, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_emotion/emotion_infer.py": 1679246024.1043785, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_emotion/efficient/model.py": 1679246024.1043785, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_emotion/efficient/utils.py": 1679246024.1053786, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_emotion/face_alignment/face_align.py": 1679246024.1053786, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_emotion/face_alignment/face.py": 1679246024.1053786, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/table_recognition/modules/lore_detector.py": 1679246024.248408, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/table_recognition/modules/lore_processor.py": 1679246024.248408, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/table_recognition/lineless_table_process.py": 1679246024.247408, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/table_recognition/model_lore.py": 1679246024.247408, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/super_resolution/arch_util.py": 1679246024.2464077, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/super_resolution/ecbsr_model.py": 1679246024.2464077, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/super_resolution/rrdbnet_arch.py": 1679246024.2464077, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/super_resolution/ecb.py": 1679246024.2464077, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/human_reconstruction/Reconstruction.py": 1679246024.116381, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/human_reconstruction/models/PixToMesh.py": 1679246024.1183813, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/human_reconstruction/models/geometry.py": 1679246024.1183813, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/human_reconstruction/models/Embedding.py": 1679246024.1183813, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/human_reconstruction/models/human_segmenter.py": 1679246024.1183813, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/human_reconstruction/models/Res_backbone.py": 1679246024.1183813, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/human_reconstruction/models/networks.py": 1679246024.1183813, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/human_reconstruction/models/Surface_head.py": 1679246024.1183813, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/human_reconstruction/models/detectors.py": 1679246024.1183813, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/human_reconstruction/utils.py": 1679246024.1183813, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_matching/utils/misc.py": 1679246024.175393, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_matching/loftr_quadtree/backbone/resnet_fpn.py": 1679246024.1723924, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_matching/loftr_quadtree/loftr_module/transformer.py": 1679246024.1733928, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_matching/loftr_quadtree/loftr_module/quadtree_attention.py": 1679246024.1733928, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_matching/loftr_quadtree/loftr_module/fine_preprocess.py": 1679246024.1723924, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_matching/loftr_quadtree/loftr_module/linear_attention.py": 1679246024.1733928, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_matching/loftr_quadtree/utils/coarse_matching.py": 1679246024.174393, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_matching/loftr_quadtree/utils/fine_matching.py": 1679246024.174393, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_matching/loftr_quadtree/utils/position_encoding.py": 1679246024.174393, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_matching/loftr_quadtree/loftr.py": 1679246024.1713922, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_matching/config/default.py": 1679246024.1713922, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_matching/quadtree_attention_model.py": 1679246024.170392, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/robust_image_classification/easyrobust_model.py": 1679246024.2354054, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/body_3d_keypoints/hdformer/hdformer_detector.py": 1679246024.0793731, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/body_3d_keypoints/hdformer/block.py": 1679246024.0793731, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/body_3d_keypoints/hdformer/hdformer.py": 1679246024.0793731, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/body_3d_keypoints/hdformer/backbone.py": 1679246024.078373, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/body_3d_keypoints/hdformer/directed_graph.py": 1679246024.0793731, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/body_3d_keypoints/hdformer/skeleton.py": 1679246024.0793731, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/body_3d_keypoints/cannonical_pose/body_3d_pose.py": 1679246024.078373, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/body_3d_keypoints/cannonical_pose/canonical_pose_modules.py": 1679246024.078373, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_deinterlace/UNet_for_video_deinterlace.py": 1679246024.262411, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_deinterlace/models/archs.py": 1679246024.2634113, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_deinterlace/models/deep_fourier_upsampling.py": 1679246024.2634113, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_deinterlace/models/enh.py": 1679246024.2634113, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_deinterlace/models/utils.py": 1679246024.2644114, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_deinterlace/models/fre.py": 1679246024.2634113, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_deinterlace/deinterlace_arch.py": 1679246024.2634113, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/bad_image_detecting/bad_image_detecting.py": 1679246024.0763726, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/cmdssl_video_embedding/c3d.py": 1679246024.083374, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/cmdssl_video_embedding/resnet3d.py": 1679246024.083374, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/cmdssl_video_embedding/resnet2p1d.py": 1679246024.083374, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_to_image_translation/models/clip.py": 1679246024.2003982, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_to_image_translation/models/autoencoder.py": 1679246024.2003982, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_to_image_translation/model_translation.py": 1679246024.199398, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_to_image_translation/data/transforms.py": 1679246024.2003982, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_to_image_translation/ops/random_color.py": 1679246024.2023988, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_to_image_translation/ops/utils.py": 1679246024.203399, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_to_image_translation/ops/random_mask.py": 1679246024.2023988, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_to_image_translation/ops/degradation.py": 1679246024.2013984, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_to_image_translation/ops/losses.py": 1679246024.2023988, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_to_image_translation/ops/svd.py": 1679246024.2023988, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_to_image_translation/ops/apps.py": 1679246024.2013984, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_to_image_translation/ops/metrics.py": 1679246024.2023988, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_to_image_translation/ops/diffusion.py": 1679246024.2013984, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/object_detection/yolox_pai.py": 1679246024.2134008, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/object_detection/mmdet_ms/backbones/vit.py": 1679246024.2144012, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/object_detection/mmdet_ms/utils/checkpoint.py": 1679246024.2174017, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/object_detection/mmdet_ms/utils/convModule_norm.py": 1679246024.2174017, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/object_detection/mmdet_ms/roi_heads/bbox_heads/convfc_bbox_head.py": 1679246024.2164016, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/object_detection/mmdet_ms/roi_heads/mask_heads/fcn_mask_head.py": 1679246024.2164016, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/object_detection/mmdet_ms/necks/fpn.py": 1679246024.2154014, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/object_detection/mmdet_ms/dense_heads/anchor_head.py": 1679246024.2144012, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/object_detection/mmdet_ms/dense_heads/rpn_head.py": 1679246024.2144012, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/object_detection/mmdet_model.py": 1679246024.2134008, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/object_detection/dino.py": 1679246024.2134008, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/hand_static/hand_model.py": 1679246024.116381, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/hand_static/networks.py": 1679246024.116381, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/nerf_recon_acc/nerf_recon_acc.py": 1679246024.2114005, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/nerf_recon_acc/dataloader/read_write_model.py": 1679246024.2124007, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/nerf_recon_acc/dataloader/nerf_dataset.py": 1679246024.2114005, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/nerf_recon_acc/network/nerf.py": 1679246024.2124007, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/nerf_recon_acc/network/segmenter.py": 1679246024.2124007, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/nerf_recon_acc/network/utils.py": 1679246024.2124007, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/nerf_recon_acc/nerf_preprocess.py": 1679246024.2114005, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/motion_generation/model.py": 1679246024.2073996, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/motion_generation/modules/cfg_sampler.py": 1679246024.2073996, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/motion_generation/modules/gaussian_diffusion.py": 1679246024.2084, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/motion_generation/modules/mdm.py": 1679246024.2084, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/motion_generation/modules/respace.py": 1679246024.2084, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/motion_generation/modules/rotation2xyz.py": 1679246024.2084, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/motion_generation/modules/smpl.py": 1679246024.2094002, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/virual_tryon/sdafnet.py": 1679246024.30542, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/animal_recognition/resnet.py": 1679246024.0763726, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/animal_recognition/splat.py": 1679246024.0763726, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/product_retrieval_embedding/item_embedding.py": 1679246024.2314045, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/product_retrieval_embedding/item_model.py": 1679246024.2314045, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/product_retrieval_embedding/item_detection.py": 1679246024.2314045, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_quality_assessment_man/maniqa.py": 1679246024.1853952, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_quality_assessment_man/image_quality_assessment_man.py": 1679246024.1853952, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_quality_assessment_man/swin.py": 1679246024.1853952, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_paintbyexample/model.py": 1679246024.179394, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/body_2d_keypoints/w48.py": 1679246024.0773728, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/body_2d_keypoints/hrnet_v2.py": 1679246024.0773728, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/body_2d_keypoints/hrnet_basic_modules.py": 1679246024.0773728, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_human_parsing/m2fp_net.py": 1679246024.1523883, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_human_parsing/backbone/deeplab_resnet.py": 1679246024.1533885, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_human_parsing/parsing_utils.py": 1679246024.1523883, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_human_parsing/m2fp/m2fp_encoder.py": 1679246024.1543887, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_human_parsing/m2fp/m2fp_decoder.py": 1679246024.1533885, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_body_reshaping/model.py": 1679246024.1203818, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_body_reshaping/pose_estimator/model.py": 1679246024.121382, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_body_reshaping/pose_estimator/body.py": 1679246024.121382, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_body_reshaping/pose_estimator/util.py": 1679246024.121382, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_body_reshaping/image_body_reshaping.py": 1679246024.1203818, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_body_reshaping/slim_utils.py": 1679246024.1203818, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_body_reshaping/person_info.py": 1679246024.1203818, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_detection/mtcnn/models/get_nets.py": 1679246024.0933762, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_detection/mtcnn/models/box_utils.py": 1679246024.092376, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_detection/mtcnn/models/detector.py": 1679246024.092376, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_detection/mtcnn/models/first_stage.py": 1679246024.092376, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_detection/retinaface/models/net.py": 1679246024.0953765, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_detection/retinaface/models/retinaface.py": 1679246024.0953765, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_detection/retinaface/utils.py": 1679246024.0943763, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_detection/retinaface/detection.py": 1679246024.0943763, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_detection/ulfd_slim/detection.py": 1679246024.102378, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_detection/ulfd_slim/vision/box_utils.py": 1679246024.102378, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_detection/ulfd_slim/vision/ssd/mb_tiny_fd.py": 1679246024.1033783, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_detection/ulfd_slim/vision/ssd/fd_config.py": 1679246024.1033783, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_detection/ulfd_slim/vision/ssd/ssd.py": 1679246024.1033783, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_detection/ulfd_slim/vision/ssd/predictor.py": 1679246024.1033783, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_detection/ulfd_slim/vision/ssd/data_preprocessing.py": 1679246024.1033783, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_detection/ulfd_slim/vision/transforms.py": 1679246024.102378, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_detection/ulfd_slim/vision/mb_tiny.py": 1679246024.102378, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_detection/mogface/models/resnet.py": 1679246024.0913758, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_detection/mogface/models/utils.py": 1679246024.0913758, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_detection/mogface/models/mogprednet.py": 1679246024.0913758, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_detection/mogface/models/detectors.py": 1679246024.0913758, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_detection/mogface/models/mogface.py": 1679246024.0913758, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_detection/scrfd/preprocessor.py": 1679246024.0953765, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_detection/scrfd/scrfd_detect.py": 1679246024.0953765, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_detection/scrfd/mmdet_patch/models/backbones/resnet.py": 1679246024.0993774, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_detection/scrfd/mmdet_patch/models/backbones/mobilenet.py": 1679246024.0993774, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_detection/scrfd/mmdet_patch/models/detectors/scrfd.py": 1679246024.1013777, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_detection/scrfd/mmdet_patch/models/detectors/tinymog.py": 1679246024.1013777, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_detection/scrfd/mmdet_patch/models/detectors/single_stage.py": 1679246024.1013777, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_detection/scrfd/mmdet_patch/models/detectors/base.py": 1679246024.1003776, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_detection/scrfd/mmdet_patch/models/dense_heads/scrfd_head.py": 1679246024.1003776, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_detection/scrfd/mmdet_patch/datasets/retinaface.py": 1679246024.097377, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_detection/scrfd/mmdet_patch/datasets/pipelines/formating.py": 1679246024.0983772, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_detection/scrfd/mmdet_patch/datasets/pipelines/loading.py": 1679246024.0983772, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_detection/scrfd/mmdet_patch/datasets/pipelines/transforms.py": 1679246024.0983772, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_detection/scrfd/mmdet_patch/datasets/pipelines/auto_augment.py": 1679246024.0983772, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_detection/scrfd/mmdet_patch/core/bbox/transforms.py": 1679246024.0963767, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_detection/scrfd/mmdet_patch/core/post_processing/bbox_nms.py": 1679246024.097377, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_detection/scrfd/tinymog_detect.py": 1679246024.0963767, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_detection/peppa_pig_face/face_landmark.py": 1679246024.0933762, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_detection/peppa_pig_face/LK/lk.py": 1679246024.0943763, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_detection/peppa_pig_face/facer.py": 1679246024.0933762, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_detection/peppa_pig_face/face_detector.py": 1679246024.0933762, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/referring_video_object_segmentation/model.py": 1679246024.232405, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/referring_video_object_segmentation/utils/position_encoding_2d.py": 1679246024.2344053, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/referring_video_object_segmentation/utils/matcher.py": 1679246024.233405, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/referring_video_object_segmentation/utils/criterion.py": 1679246024.233405, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/referring_video_object_segmentation/utils/postprocessing.py": 1679246024.2344053, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/referring_video_object_segmentation/utils/swin_transformer.py": 1679246024.2354054, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/referring_video_object_segmentation/utils/mttr.py": 1679246024.233405, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/referring_video_object_segmentation/utils/misc.py": 1679246024.233405, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/referring_video_object_segmentation/utils/backbone.py": 1679246024.233405, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/referring_video_object_segmentation/utils/segmentation.py": 1679246024.2344053, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/referring_video_object_segmentation/utils/multimodal_transformer.py": 1679246024.2344053, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/abnormal_object_detection/mmdet_ms/roi_head/mask_scoring_roi_head.py": 1679246024.073372, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/abnormal_object_detection/mmdet_ms/roi_head/roi_extractors/single_level_roi_extractor.py": 1679246024.073372, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/abnormal_object_detection/mmdet_model.py": 1679246024.0723717, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_panoptic_segmentation/r50_panseg_model.py": 1679246024.179394, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_panoptic_segmentation/panseg_model.py": 1679246024.179394, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_super_resolution/real_basicvsr_for_video_super_resolution.py": 1679246024.3034196, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_super_resolution/common.py": 1679246024.3034196, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_super_resolution/basicvsr_net.py": 1679246024.3024194, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_super_resolution/msrresnet_lite_model.py": 1679246024.3034196, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_super_resolution/real_basicvsr_net.py": 1679246024.3034196, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/stream_yolo/models/darknet.py": 1679246024.2444074, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/stream_yolo/models/tal_head.py": 1679246024.2454076, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/stream_yolo/models/dfp_pafpn.py": 1679246024.2444074, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/stream_yolo/models/streamyolo.py": 1679246024.2454076, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/stream_yolo/models/network_blocks.py": 1679246024.2454076, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/stream_yolo/utils/boxes.py": 1679246024.2454076, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/stream_yolo/utils/format.py": 1679246024.2454076, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/stream_yolo/realtime_video_detector.py": 1679246024.2424068, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/stream_yolo/data/data_augment.py": 1679246024.2434072, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/stream_yolo/exp/yolox_base.py": 1679246024.2434072, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/stream_yolo/exp/base_exp.py": 1679246024.2434072, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/stream_yolo/exp/build.py": 1679246024.2434072, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/stream_yolo/exp/default/streamyolo.py": 1679246024.2444074, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/product_segmentation/seg_infer.py": 1679246024.232405, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/product_segmentation/net.py": 1679246024.232405, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_reid_person/pass_model.py": 1679246024.1873956, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_reid_person/transreid_model.py": 1679246024.1873956, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_depth_estimation/geometry/camera.py": 1679246024.2654116, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_depth_estimation/geometry/pose_utils.py": 1679246024.2654116, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_depth_estimation/geometry/camera_utils.py": 1679246024.2654116, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_depth_estimation/geometry/pose.py": 1679246024.2654116, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_depth_estimation/configs/default_config.py": 1679246024.2644114, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_depth_estimation/models/model_wrapper.py": 1679246024.2664118, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_depth_estimation/models/sfm_model_mf.py": 1679246024.2664118, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_depth_estimation/models/model_utils.py": 1679246024.2664118, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_depth_estimation/models/model_checkpoint.py": 1679246024.2664118, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_depth_estimation/models/sup_model_mf.py": 1679246024.2674122, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_depth_estimation/utils/config.py": 1679246024.2704127, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_depth_estimation/utils/image.py": 1679246024.2704127, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_depth_estimation/utils/depth.py": 1679246024.2704127, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_depth_estimation/utils/horovod.py": 1679246024.2704127, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_depth_estimation/utils/augmentations.py": 1679246024.2694125, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_depth_estimation/utils/misc.py": 1679246024.2714128, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_depth_estimation/utils/types.py": 1679246024.2714128, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_depth_estimation/utils/load.py": 1679246024.2714128, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_depth_estimation/utils/image_gt.py": 1679246024.2704127, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_depth_estimation/networks/depth_pose/depth_pose_net.py": 1679246024.2674122, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_depth_estimation/networks/optim/update.py": 1679246024.2694125, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_depth_estimation/networks/optim/extractor.py": 1679246024.2694125, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_depth_estimation/networks/layers/resnet/resnet_encoder.py": 1679246024.2684124, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_depth_estimation/networks/layers/resnet/layers.py": 1679246024.2684124, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_depth_estimation/networks/layers/resnet/pose_decoder.py": 1679246024.2684124, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_depth_estimation/networks/layers/resnet/depth_decoder.py": 1679246024.2684124, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_depth_estimation/dro_model.py": 1679246024.2644114, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_colorization/unet/utils.py": 1679246024.1293836, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_colorization/unet/unet.py": 1679246024.1293836, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_colorization/ddcolor/loss.py": 1679246024.126383, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_colorization/ddcolor/ddcolor.py": 1679246024.1253827, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_colorization/ddcolor/utils/vgg.py": 1679246024.1283834, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_colorization/ddcolor/utils/transformer_utils.py": 1679246024.1273832, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_colorization/ddcolor/utils/convnext.py": 1679246024.1273832, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_colorization/ddcolor/utils/position_encoding.py": 1679246024.1273832, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_colorization/ddcolor/utils/unet.py": 1679246024.1283834, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_colorization/ddcolor/ddcolor_for_image_colorization.py": 1679246024.126383, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_inpainting/inpainting.py": 1679246024.277414, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_inpainting/inpainting_model.py": 1679246024.277414, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/movie_scene_segmentation/model.py": 1679246024.2094002, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/movie_scene_segmentation/get_model.py": 1679246024.2094002, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/movie_scene_segmentation/utils/trn.py": 1679246024.2104003, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/movie_scene_segmentation/utils/save_op.py": 1679246024.2104003, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/movie_scene_segmentation/utils/head.py": 1679246024.2104003, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/movie_scene_segmentation/utils/shot_encoder.py": 1679246024.2104003, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/skin_retouching/retinaface/box_utils.py": 1679246024.2414067, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/skin_retouching/retinaface/utils.py": 1679246024.2424068, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/skin_retouching/retinaface/net.py": 1679246024.2414067, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/skin_retouching/retinaface/network.py": 1679246024.2414067, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/skin_retouching/retinaface/prior_box.py": 1679246024.2424068, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/skin_retouching/retinaface/predict_single.py": 1679246024.2414067, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/skin_retouching/detection_model/detection_module.py": 1679246024.2404065, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/skin_retouching/detection_model/detection_unet_in.py": 1679246024.2404065, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/skin_retouching/utils.py": 1679246024.2394063, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/skin_retouching/inpainting_model/gconv.py": 1679246024.2404065, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/skin_retouching/inpainting_model/inpainting_unet.py": 1679246024.2404065, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/skin_retouching/unet_deploy.py": 1679246024.2394063, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/skin_retouching/weights_init.py": 1679246024.2394063, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/vision_efficient_tuning/model.py": 1679246024.30642, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/vision_efficient_tuning/timm_weight_init.py": 1679246024.3074203, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/vision_efficient_tuning/petl.py": 1679246024.30642, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/vision_efficient_tuning/head.py": 1679246024.30642, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/vision_efficient_tuning/timm_helpers.py": 1679246024.30642, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/vision_efficient_tuning/vision_efficient_tuning.py": 1679246024.3074203, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/vision_efficient_tuning/timm_vision_transformer.py": 1679246024.3074203, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/vision_efficient_tuning/backbone.py": 1679246024.30642, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/controllable_image_generation/controlnet.py": 1679246024.0843742, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/controllable_image_generation/annotator/midas/utils.py": 1679246024.0853746, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/controllable_image_generation/annotator/midas/midas/base_model.py": 1679246024.0853746, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/controllable_image_generation/annotator/midas/midas/midas_net.py": 1679246024.0863748, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/controllable_image_generation/annotator/midas/midas/midas_net_custom.py": 1679246024.0863748, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/controllable_image_generation/annotator/midas/midas/blocks.py": 1679246024.0853746, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/controllable_image_generation/annotator/midas/midas/vit.py": 1679246024.0863748, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/controllable_image_generation/annotator/midas/midas/transforms.py": 1679246024.0863748, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/controllable_image_generation/annotator/midas/midas/dpt_depth.py": 1679246024.0863748, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/controllable_image_generation/annotator/midas/api.py": 1679246024.0853746, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/controllable_image_generation/annotator/annotator.py": 1679246024.0843742, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/controllable_image_generation/annotator/mlsd/utils.py": 1679246024.087375, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/controllable_image_generation/annotator/mlsd/mbv2_mlsd_large.py": 1679246024.087375, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/controllable_image_generation/annotator/openpose/model.py": 1679246024.088375, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/controllable_image_generation/annotator/openpose/body.py": 1679246024.087375, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/controllable_image_generation/annotator/openpose/util.py": 1679246024.088375, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/controllable_image_generation/annotator/openpose/hand.py": 1679246024.088375, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_portrait_enhancement/gpen.py": 1679246024.1803942, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_portrait_enhancement/retinaface/models/net.py": 1679246024.1833947, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_portrait_enhancement/retinaface/models/retinaface.py": 1679246024.1833947, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_portrait_enhancement/retinaface/utils.py": 1679246024.1833947, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_portrait_enhancement/retinaface/detection.py": 1679246024.1823945, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_portrait_enhancement/align_faces.py": 1679246024.1803942, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_portrait_enhancement/losses/model_irse.py": 1679246024.1823945, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_portrait_enhancement/losses/helpers.py": 1679246024.1823945, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_portrait_enhancement/losses/losses.py": 1679246024.1823945, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_portrait_enhancement/eqface/fqa.py": 1679246024.1813943, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_portrait_enhancement/eqface/model_resnet.py": 1679246024.1813943, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_portrait_enhancement/image_portrait_enhancement.py": 1679246024.1813943, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/ocr_recognition/model.py": 1679246024.2264037, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/ocr_recognition/modules/convnext.py": 1679246024.2264037, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/ocr_recognition/modules/convnextvit.py": 1679246024.2264037, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/ocr_recognition/modules/crnn.py": 1679246024.2264037, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/ocr_recognition/modules/vitstr.py": 1679246024.2274039, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/ocr_recognition/modules/timm_tinyc.py": 1679246024.2274039, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/ocr_recognition/preprocessor.py": 1679246024.2264037, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_semantic_segmentation/segformer.py": 1679246024.189396, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_semantic_segmentation/semantic_seg_model.py": 1679246024.189396, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_semantic_segmentation/vit_adapter/models/decode_heads/base_decode_head.py": 1679246024.1933968, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_semantic_segmentation/vit_adapter/models/decode_heads/mask2former_head_from_mmseg.py": 1679246024.1933968, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_semantic_segmentation/vit_adapter/models/backbone/adapter_modules.py": 1679246024.1913965, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_semantic_segmentation/vit_adapter/models/backbone/base/beit.py": 1679246024.1923966, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_semantic_segmentation/vit_adapter/models/backbone/beit_adapter.py": 1679246024.1923966, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_semantic_segmentation/vit_adapter/models/segmentors/encoder_decoder_mask2former.py": 1679246024.194397, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_semantic_segmentation/vit_adapter/models/segmentors/base_segmentor.py": 1679246024.1933968, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_semantic_segmentation/vit_adapter/utils/seg_func.py": 1679246024.194397, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_semantic_segmentation/vit_adapter/utils/data_process_func.py": 1679246024.194397, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_semantic_segmentation/vit_adapter/utils/builder.py": 1679246024.194397, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_semantic_segmentation/ddpm_seg/pixel_classifier.py": 1679246024.190396, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_semantic_segmentation/ddpm_seg/utils.py": 1679246024.190396, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_semantic_segmentation/ddpm_seg/data_util.py": 1679246024.189396, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_semantic_segmentation/ddpm_seg/feature_extractors.py": 1679246024.190396, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_semantic_segmentation/pan_merge/base_panoptic_fusion_head.py": 1679246024.190396, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_semantic_segmentation/pan_merge/maskformer_semantic_head.py": 1679246024.190396, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_semantic_segmentation/ddpm_segmentation_model.py": 1679246024.189396, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_reconstruction/models/facerecon_model.py": 1679246024.1113799, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_reconstruction/models/opt.py": 1679246024.11238, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_reconstruction/models/nv_diffrast.py": 1679246024.11238, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_reconstruction/models/networks.py": 1679246024.11238, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_reconstruction/models/losses.py": 1679246024.1113799, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_reconstruction/models/bfm.py": 1679246024.1113799, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_reconstruction/models/facelandmark/large_model_infer.py": 1679246024.1133802, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_reconstruction/models/facelandmark/nets/large_base_lmks_net.py": 1679246024.1133802, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_reconstruction/models/facelandmark/nets/large_eyeball_net.py": 1679246024.1133802, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_reconstruction/models/facelandmark/large_base_lmks_infer.py": 1679246024.11238, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_reconstruction/utils.py": 1679246024.1103797, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_defrcn_fewshot/models/resnet.py": 1679246024.1353848, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_defrcn_fewshot/models/defrcn.py": 1679246024.1343846, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_defrcn_fewshot/models/calibration_layer.py": 1679246024.1333845, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_defrcn_fewshot/models/gdl.py": 1679246024.1343846, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_defrcn_fewshot/models/roi_heads.py": 1679246024.1353848, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_defrcn_fewshot/models/fast_rcnn.py": 1679246024.1343846, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_defrcn_fewshot/evaluation/evaluator.py": 1679246024.1323843, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_defrcn_fewshot/evaluation/coco_evaluation.py": 1679246024.1323843, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_defrcn_fewshot/evaluation/pascal_voc_evaluation.py": 1679246024.1333845, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_defrcn_fewshot/defrcn_for_fewshot.py": 1679246024.131384, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_defrcn_fewshot/utils/model_surgery_op.py": 1679246024.136385, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_defrcn_fewshot/utils/register_data.py": 1679246024.136385, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_defrcn_fewshot/utils/configuration_mapper.py": 1679246024.136385, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_defrcn_fewshot/utils/requirements_check.py": 1679246024.1373851, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_defrcn_fewshot/utils/coco_register.py": 1679246024.136385, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_defrcn_fewshot/utils/voc_register.py": 1679246024.1373851, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_object_segmentation/model.py": 1679246024.285416, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_object_segmentation/mod_resnet.py": 1679246024.285416, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_object_segmentation/inference_memory_bank.py": 1679246024.285416, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_object_segmentation/inference_core.py": 1679246024.285416, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_object_segmentation/eval_network.py": 1679246024.285416, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_object_segmentation/aggregate.py": 1679246024.2844157, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_object_segmentation/cbam.py": 1679246024.2844157, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_object_segmentation/network.py": 1679246024.286416, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_object_segmentation/modules.py": 1679246024.286416, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_human_matting/model.py": 1679246024.2754138, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_human_matting/models/decoder.py": 1679246024.2754138, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_human_matting/models/lraspp.py": 1679246024.276414, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_human_matting/models/deep_guided_filter.py": 1679246024.276414, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_human_matting/models/effv2.py": 1679246024.276414, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_human_matting/models/matting.py": 1679246024.276414, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/pointcloud_sceneflow_estimation/rcp_model.py": 1679246024.2304044, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/pointcloud_sceneflow_estimation/pointnet2_utils.py": 1679246024.2304044, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/pointcloud_sceneflow_estimation/sf_rcp.py": 1679246024.2304044, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/pointcloud_sceneflow_estimation/common.py": 1679246024.2304044, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_inpainting/model.py": 1679246024.1563892, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_inpainting/default.py": 1679246024.1553888, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_inpainting/refinement.py": 1679246024.1563892, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_inpainting/modules/perceptual.py": 1679246024.1593897, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_inpainting/modules/adversarial.py": 1679246024.1573894, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_inpainting/modules/feature_matching.py": 1679246024.1573894, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_inpainting/modules/ffc.py": 1679246024.1583896, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_inpainting/modules/ade20k/resnet.py": 1679246024.16039, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_inpainting/modules/ade20k/base.py": 1679246024.16039, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_inpainting/modules/inception.py": 1679246024.1583896, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_inpainting/modules/pix2pixhd.py": 1679246024.1593897, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_inpainting/base.py": 1679246024.1553888, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_mvs_depth_estimation/casmvs_model.py": 1679246024.1763933, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_mvs_depth_estimation/colmap2mvsnet.py": 1679246024.1773934, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_mvs_depth_estimation/cas_mvsnet.py": 1679246024.175393, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_mvs_depth_estimation/general_eval_dataset.py": 1679246024.1773934, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_mvs_depth_estimation/utils.py": 1679246024.1783936, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_mvs_depth_estimation/depth_filter.py": 1679246024.1773934, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_mvs_depth_estimation/module.py": 1679246024.1783936, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/action_detection/modules/resnet.py": 1679246024.0743723, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/action_detection/modules/action_detection_pytorch.py": 1679246024.0743723, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/action_detection/action_detection_onnx.py": 1679246024.073372, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_summarization/base_model.py": 1679246024.301419, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_summarization/summarizer.py": 1679246024.301419, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_summarization/kts/cpd_nonlin.py": 1679246024.3024194, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_summarization/kts/cpd_auto.py": 1679246024.3024194, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_summarization/pgl_sum.py": 1679246024.301419, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/action_recognition/tada_convnext.py": 1679246024.0753725, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/action_recognition/s3dg.py": 1679246024.0753725, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/action_recognition/temporal_patch_shift_transformer.py": 1679246024.0753725, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/action_recognition/models.py": 1679246024.0743723, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/facial_expression_recognition/fer/vgg.py": 1679246024.1143804, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/facial_expression_recognition/fer/transforms.py": 1679246024.1143804, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/facial_expression_recognition/fer/facial_expression_recognition.py": 1679246024.1143804, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_debanding/rrdb/rrdb_image_debanding.py": 1679246024.1303837, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/vision_middleware/model.py": 1679246024.3084207, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/vision_middleware/vim.py": 1679246024.3084207, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/vision_middleware/head.py": 1679246024.3084207, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/vision_middleware/backbone.py": 1679246024.3084207, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_quality_assessment_mos/image_quality_assessment_mos.py": 1679246024.1863954, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_quality_assessment_mos/heads/simple_head.py": 1679246024.1873956, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_quality_assessment_mos/backbones/resnet.py": 1679246024.1863954, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_quality_assessment_mos/censeo_ivqa_model.py": 1679246024.1863954, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/object_detection_3d/depe/mmdet3d_plugin/models/backbones/vovnet.py": 1679246024.2214026, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/object_detection_3d/depe/mmdet3d_plugin/models/utils/petr_transformer.py": 1679246024.223403, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/object_detection_3d/depe/mmdet3d_plugin/models/utils/positional_encoding.py": 1679246024.2244031, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/object_detection_3d/depe/mmdet3d_plugin/models/detectors/petr3d.py": 1679246024.223403, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/object_detection_3d/depe/mmdet3d_plugin/models/necks/cp_fpn.py": 1679246024.223403, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/object_detection_3d/depe/mmdet3d_plugin/models/dense_heads/depth_net.py": 1679246024.2224028, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/object_detection_3d/depe/mmdet3d_plugin/models/dense_heads/petrv2_dednhead.py": 1679246024.2224028, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/object_detection_3d/depe/mmdet3d_plugin/datasets/pipelines/loading.py": 1679246024.2204025, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/object_detection_3d/depe/mmdet3d_plugin/datasets/pipelines/transform_3d.py": 1679246024.2214026, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/object_detection_3d/depe/mmdet3d_plugin/datasets/nuscenes_dataset.py": 1679246024.2204025, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/object_detection_3d/depe/mmdet3d_plugin/core/bbox/match_costs/match_cost.py": 1679246024.2204025, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/object_detection_3d/depe/mmdet3d_plugin/core/bbox/util.py": 1679246024.218402, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/object_detection_3d/depe/mmdet3d_plugin/core/bbox/coders/nms_free_coder.py": 1679246024.219402, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/object_detection_3d/depe/mmdet3d_plugin/core/bbox/assigners/hungarian_assigner_3d.py": 1679246024.219402, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/object_detection_3d/depe/result_vis.py": 1679246024.218402, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/object_detection_3d/depe/depe_detect.py": 1679246024.2174017, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_attribute_recognition/fair_face/face_attribute_recognition.py": 1679246024.0903754, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_generation/stylegan2.py": 1679246024.1063788, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_generation/op/fused_act.py": 1679246024.1063788, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_generation/op/conv2d_gradfix.py": 1679246024.1063788, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_generation/op/upfirdn2d.py": 1679246024.1063788, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_skychange/ptsemseg/hrnet_backnone.py": 1679246024.1963975, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_skychange/ptsemseg/BlockModules.py": 1679246024.1963975, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_skychange/ptsemseg/hrnet_super_and_ocr.py": 1679246024.1963975, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_skychange/ptsemseg/unet.py": 1679246024.1973977, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_skychange/skychange.py": 1679246024.1953971, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_skychange/preprocessor.py": 1679246024.1953971, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_skychange/skychange_model.py": 1679246024.1953971, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_color_enhance/image_color_enhance.py": 1679246024.1233823, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_color_enhance/adaint/adaint.py": 1679246024.1233823, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_color_enhance/csrnet.py": 1679246024.1233823, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_color_enhance/deeplpf/deeplpf_image_color_enhance.py": 1679246024.1243825, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_color_enhance/deeplpf/deeplpfnet.py": 1679246024.1243825, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/facial_landmark_confidence/flc/manual_landmark_net.py": 1679246024.1153808, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/facial_landmark_confidence/flc/facial_landmark_confidence.py": 1679246024.1153808, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_instance_segmentation/model.py": 1679246024.1613903, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_instance_segmentation/maskdino_swin.py": 1679246024.1613903, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_instance_segmentation/postprocess_utils.py": 1679246024.1623905, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_instance_segmentation/maskdino/maskdino_encoder.py": 1679246024.1693919, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_instance_segmentation/maskdino/dino_decoder.py": 1679246024.1683917, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_instance_segmentation/maskdino/ms_deform_attn.py": 1679246024.1693919, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_instance_segmentation/maskdino/utils.py": 1679246024.170392, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_instance_segmentation/maskdino/maskdino_decoder.py": 1679246024.1683917, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_instance_segmentation/maskdino/position_encoding.py": 1679246024.1693919, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_instance_segmentation/cascade_mask_rcnn_swin.py": 1679246024.16039, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_instance_segmentation/backbones/swin_transformer.py": 1679246024.1633906, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_instance_segmentation/maskdino_model.py": 1679246024.1613903, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_instance_segmentation/datasets/transforms.py": 1679246024.1633906, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_instance_segmentation/utils.py": 1679246024.277414, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_instance_segmentation/video_knet.py": 1679246024.2784145, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_instance_segmentation/track/mask_hungarian_assigner.py": 1679246024.281415, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_instance_segmentation/track/kernel_update_head.py": 1679246024.281415, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_instance_segmentation/head/kernel_updator.py": 1679246024.2804148, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_instance_segmentation/head/kernel_update_head.py": 1679246024.2804148, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_instance_segmentation/head/kernel_frame_iter_head.py": 1679246024.2784145, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_instance_segmentation/head/kernel_iter_head.py": 1679246024.2794147, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_instance_segmentation/head/kernel_head.py": 1679246024.2794147, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_instance_segmentation/neck/msdeformattn_decoder.py": 1679246024.2804148, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/vop_retrieval/model.py": 1679246024.3094208, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/vop_retrieval/basic_utils.py": 1679246024.3094208, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/vop_retrieval/model_se.py": 1679246024.3094208, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/vop_retrieval/tokenization_clip.py": 1679246024.310421, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/vop_retrieval/backbone.py": 1679246024.3094208, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/indoor_layout_estimation/panovit.py": 1679246024.203399, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/indoor_layout_estimation/networks/modality/layout.py": 1679246024.2053993, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/indoor_layout_estimation/networks/backbone/resnet_DA.py": 1679246024.204399, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/indoor_layout_estimation/networks/backbone/vit_horizon_pry_image.py": 1679246024.204399, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/indoor_layout_estimation/networks/utils.py": 1679246024.204399, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/indoor_layout_estimation/networks/panovit.py": 1679246024.203399, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/indoor_layout_estimation/networks/misc/post_proc.py": 1679246024.2053993, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/indoor_layout_estimation/networks/misc/fourier.py": 1679246024.2053993, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/indoor_layout_estimation/networks/misc/panostretch.py": 1679246024.2053993, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_denoise/nafnet_for_image_denoise.py": 1679246024.1383855, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_denoise/nafnet/arch_util.py": 1679246024.1383855, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_denoise/nafnet/NAFNet_arch.py": 1679246024.1383855, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_classification/resnet50_cc.py": 1679246024.1223822, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_classification/utils.py": 1679246024.1223822, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_classification/backbones/nextvit.py": 1679246024.1223822, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_classification/backbones/beit_v2.py": 1679246024.1223822, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_classification/mmcls_model.py": 1679246024.121382, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_to_image_generation/model.py": 1679246024.1973977, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_to_image_generation/models/clip.py": 1679246024.1983979, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_to_image_generation/models/autoencoder.py": 1679246024.1983979, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_to_image_generation/data/transforms.py": 1679246024.1973977, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_to_image_generation/ops/losses.py": 1679246024.199398, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_to_image_generation/ops/diffusion.py": 1679246024.199398, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_face_fusion/image_face_fusion.py": 1679246024.145387, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_face_fusion/facegan/model.py": 1679246024.146387, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_face_fusion/facegan/gan_wrap.py": 1679246024.145387, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_face_fusion/facegan/op/fused_act.py": 1679246024.1473873, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_face_fusion/facegan/op/conv2d_gradfix.py": 1679246024.146387, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_face_fusion/facegan/op/upfirdn2d.py": 1679246024.1473873, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_face_fusion/facelib/align_trans.py": 1679246024.1483874, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_face_fusion/facelib/matlab_cp2tform.py": 1679246024.1483874, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_face_fusion/network/model_irse.py": 1679246024.1513882, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_face_fusion/network/facerecon_model.py": 1679246024.150388, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_face_fusion/network/ops.py": 1679246024.1513882, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_face_fusion/network/aei_flow_net.py": 1679246024.1493876, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_face_fusion/network/dense_motion.py": 1679246024.150388, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_face_fusion/network/bfm.py": 1679246024.1493876, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_face_fusion/network/aad_layer.py": 1679246024.1493876, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_driving_perception/image_driving_percetion_model.py": 1679246024.1443868, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_driving_perception/utils.py": 1679246024.1443868, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_driving_perception/preprocessor.py": 1679246024.1443868, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_depth_estimation_bts/depth_estimation_bts_model.py": 1679246024.141386, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_depth_estimation_bts/networks/decoder.py": 1679246024.1423862, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_depth_estimation_bts/networks/utils.py": 1679246024.1433864, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_depth_estimation_bts/networks/encoder.py": 1679246024.1433864, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_depth_estimation_bts/networks/bts_model.py": 1679246024.1423862, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_human_hand_detection/one_stage_detector.py": 1679246024.107379, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_human_hand_detection/nanodet_plus_head.py": 1679246024.107379, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_human_hand_detection/utils.py": 1679246024.1083791, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_human_hand_detection/ghost_pan.py": 1679246024.107379, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_human_hand_detection/det_infer.py": 1679246024.107379, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_human_hand_detection/shufflenetv2.py": 1679246024.1083791, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_multi_object_tracking/models/model.py": 1679246024.2824152, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_multi_object_tracking/models/yolo.py": 1679246024.2824152, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_multi_object_tracking/models/decode.py": 1679246024.2824152, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_multi_object_tracking/models/common.py": 1679246024.2824152, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_multi_object_tracking/tracker/multitracker.py": 1679246024.2834153, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_multi_object_tracking/tracker/basetrack.py": 1679246024.2824152, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_multi_object_tracking/tracker/matching.py": 1679246024.2834153, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_multi_object_tracking/utils/kalman_filter.py": 1679246024.2834153, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_multi_object_tracking/utils/utils.py": 1679246024.2844157, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_multi_object_tracking/utils/image.py": 1679246024.2834153, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_multi_object_tracking/utils/visualization.py": 1679246024.2844157, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/tinynas_classfication/super_res_kxkx.py": 1679246024.252409, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/tinynas_classfication/global_utils.py": 1679246024.2504086, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/tinynas_classfication/master_net.py": 1679246024.2514088, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/tinynas_classfication/basic_blocks.py": 1679246024.2504086, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/tinynas_classfication/super_blocks.py": 1679246024.2514088, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/tinynas_classfication/plain_net_utils.py": 1679246024.2514088, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/tinynas_classfication/super_res_k1kxk1.py": 1679246024.252409, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/tinynas_classfication/model_zoo.py": 1679246024.2514088, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/tinynas_classfication/super_res_idwexkx.py": 1679246024.2514088, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/crowd_counting/hrnet_aspp_relu.py": 1679246024.0893753, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/crowd_counting/cc_model.py": 1679246024.0893753, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/panorama_depth_estimation/unifuse_model.py": 1679246024.228404, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/panorama_depth_estimation/networks/equi.py": 1679246024.228404, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/panorama_depth_estimation/networks/resnet.py": 1679246024.2294042, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/panorama_depth_estimation/networks/layers.py": 1679246024.228404, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/panorama_depth_estimation/networks/mobilenet.py": 1679246024.2294042, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/panorama_depth_estimation/networks/util.py": 1679246024.2294042, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/panorama_depth_estimation/networks/unifuse.py": 1679246024.2294042, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/language_guided_video_summarization/summarizer.py": 1679246024.2063994, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/language_guided_video_summarization/transformer/sub_layers.py": 1679246024.2073996, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/language_guided_video_summarization/transformer/layers.py": 1679246024.2063994, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/language_guided_video_summarization/transformer/models.py": 1679246024.2063994, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/language_guided_video_summarization/transformer/modules.py": 1679246024.2073996, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/cartoon/model_tf.py": 1679246024.0803735, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/cartoon/loss.py": 1679246024.0803735, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/cartoon/utils.py": 1679246024.0803735, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/cartoon/facelib/face_landmark.py": 1679246024.0813737, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/cartoon/facelib/LK/lk.py": 1679246024.0823739, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/cartoon/facelib/config.py": 1679246024.0813737, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/cartoon/facelib/facer.py": 1679246024.0813737, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/cartoon/facelib/face_detector.py": 1679246024.0813737, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/cartoon/mtcnn_pytorch/src/align_trans.py": 1679246024.0823739, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/cartoon/mtcnn_pytorch/src/matlab_cp2tform.py": 1679246024.0823739, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/cartoon/network.py": 1679246024.0803735, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_deblur/nafnet_for_image_deblur.py": 1679246024.131384, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/tinynas_detection/damo/augmentations/box_level_augs/geometric_augs.py": 1679246024.2554097, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/tinynas_detection/damo/augmentations/box_level_augs/box_level_augs.py": 1679246024.2544093, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/tinynas_detection/damo/augmentations/box_level_augs/gaussian_maps.py": 1679246024.2554097, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/tinynas_detection/damo/augmentations/box_level_augs/color_augs.py": 1679246024.2544093, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/tinynas_detection/damo/augmentations/scale_aware_aug.py": 1679246024.2544093, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/tinynas_detection/damo/base_models/losses/distill_loss.py": 1679246024.2594104, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/tinynas_detection/damo/base_models/losses/gfocal_loss.py": 1679246024.2594104, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/tinynas_detection/damo/base_models/heads/gfocal_v2_tiny.py": 1679246024.2584102, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/tinynas_detection/damo/base_models/heads/zero_head.py": 1679246024.2584102, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/tinynas_detection/damo/base_models/backbones/darknet.py": 1679246024.2554097, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/tinynas_detection/damo/base_models/backbones/tinynas_res.py": 1679246024.25641, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/tinynas_detection/damo/base_models/backbones/tinynas_csp.py": 1679246024.25641, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/tinynas_detection/damo/base_models/necks/giraffe_fpn_btn.py": 1679246024.2604105, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/tinynas_detection/damo/base_models/necks/giraffe_fpn.py": 1679246024.2604105, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/tinynas_detection/damo/base_models/necks/giraffe_config.py": 1679246024.2594104, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/tinynas_detection/damo/base_models/core/ota_assigner.py": 1679246024.25741, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/tinynas_detection/damo/base_models/core/utils.py": 1679246024.25741, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/tinynas_detection/damo/base_models/core/ops.py": 1679246024.25741, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/tinynas_detection/damo/base_models/core/weight_init.py": 1679246024.2584102, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/tinynas_detection/damo/base_models/core/base_ops.py": 1679246024.25641, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/tinynas_detection/damo/base_models/core/repvgg_block.py": 1679246024.25741, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/tinynas_detection/damo/base_models/core/neck_ops.py": 1679246024.25741, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/tinynas_detection/damo/apis/detector_inference.py": 1679246024.2534091, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/tinynas_detection/damo/apis/detector_evaluater.py": 1679246024.2534091, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/tinynas_detection/damo/utils/boxes.py": 1679246024.262411, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/tinynas_detection/damo/utils/model_utils.py": 1679246024.262411, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/tinynas_detection/damo/utils/scheduler.py": 1679246024.262411, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/tinynas_detection/damo/detectors/detector.py": 1679246024.2604105, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/tinynas_detection/damo/structures/bounding_box.py": 1679246024.261411, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/tinynas_detection/damo/structures/boxlist_ops.py": 1679246024.261411, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/tinynas_detection/damo/structures/image_list.py": 1679246024.261411, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/tinynas_detection/utils.py": 1679246024.2534091, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/tinynas_detection/detector.py": 1679246024.252409, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/tinynas_detection/tinynas_damoyolo.py": 1679246024.252409, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/tinynas_detection/tinynas_detector.py": 1679246024.2534091, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_depth_estimation/networks/newcrf_layers.py": 1679246024.1403859, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_depth_estimation/networks/uper_crf_head.py": 1679246024.141386, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_depth_estimation/networks/newcrf_utils.py": 1679246024.1403859, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_depth_estimation/networks/newcrf_depth.py": 1679246024.1393857, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_depth_estimation/networks/swin_transformer.py": 1679246024.1403859, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_depth_estimation/newcrfs_model.py": 1679246024.1393857, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_recognition/align_face.py": 1679246024.1083791, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_recognition/torchkit/backbone/model_irse.py": 1679246024.1103797, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_recognition/torchkit/backbone/facemask_backbone.py": 1679246024.1093795, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_recognition/torchkit/backbone/arcface_backbone.py": 1679246024.1093795, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_recognition/torchkit/backbone/model_resnet.py": 1679246024.1103797, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_recognition/torchkit/backbone/common.py": 1679246024.1093795, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_recognition/torchkit/rts_backbone.py": 1679246024.1093795, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/salient_detection/salient_model.py": 1679246024.2354054, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/salient_detection/models/u2net.py": 1679246024.2364056, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/salient_detection/models/backbone/Res2Net_v1b.py": 1679246024.237406, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/salient_detection/models/utils.py": 1679246024.2364056, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/salient_detection/models/senet.py": 1679246024.2364056, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/salient_detection/models/modules.py": 1679246024.2364056, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_probing_model/model.py": 1679246024.1843948, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_probing_model/utils.py": 1679246024.1843948, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_probing_model/backbone.py": 1679246024.1843948, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_quality_assessment_degradation/degradation_model.py": 1679246024.1843948, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_quality_assessment_degradation/image_quality_assessment_degradation.py": 1679246024.1853952, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/open_vocabulary_detection_vild/vild.py": 1679246024.2274039, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/text_driven_segmentation/model.py": 1679246024.2494085, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/text_driven_segmentation/lseg_vit.py": 1679246024.2494085, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/text_driven_segmentation/clip.py": 1679246024.248408, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/text_driven_segmentation/lseg_blocks.py": 1679246024.2494085, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/text_driven_segmentation/lseg_model.py": 1679246024.2494085, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/text_driven_segmentation/simple_tokenizer.py": 1679246024.2504086, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/text_driven_segmentation/lseg_net.py": 1679246024.2494085, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/text_driven_segmentation/lseg_base.py": 1679246024.248408, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_restoration/image_restoration_model.py": 1679246024.1883957, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_restoration/demoire_models/nets.py": 1679246024.1883957, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_panoptic_segmentation/backbone/swin_transformer.py": 1679246024.2874162, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_panoptic_segmentation/backbone/swin_checkpoint.py": 1679246024.2874162, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_panoptic_segmentation/visualizer.py": 1679246024.286416, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_panoptic_segmentation/video_k_net.py": 1679246024.286416, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_panoptic_segmentation/track/quasi_dense_embed_tracker.py": 1679246024.290417, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_panoptic_segmentation/head/kernel_updator.py": 1679246024.2884164, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_panoptic_segmentation/head/kernel_update_head.py": 1679246024.2884164, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_panoptic_segmentation/head/mask.py": 1679246024.2894166, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_panoptic_segmentation/head/kernel_iter_head.py": 1679246024.2884164, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_panoptic_segmentation/head/semantic_fpn_wrapper.py": 1679246024.2894166, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_panoptic_segmentation/head/track_heads.py": 1679246024.2894166, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_panoptic_segmentation/head/kernel_head.py": 1679246024.2884164, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_panoptic_segmentation/neck/fpn.py": 1679246024.2894166, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_frame_interpolation/VFINet_for_video_frame_interpolation.py": 1679246024.2714128, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_frame_interpolation/interp_model/IFNet_swin.py": 1679246024.2734134, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_frame_interpolation/interp_model/refinenet_arch.py": 1679246024.2744136, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_frame_interpolation/interp_model/transformer_layers.py": 1679246024.2744136, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_frame_interpolation/interp_model/UNet.py": 1679246024.2734134, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_frame_interpolation/interp_model/flow_reversal.py": 1679246024.2734134, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_frame_interpolation/utils/utils.py": 1679246024.2754138, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_frame_interpolation/utils/scene_change_detection.py": 1679246024.2744136, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_frame_interpolation/flow_model/update.py": 1679246024.2734134, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_frame_interpolation/flow_model/corr.py": 1679246024.272413, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_frame_interpolation/flow_model/extractor.py": 1679246024.272413, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_frame_interpolation/flow_model/raft.py": 1679246024.272413, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_frame_interpolation/VFINet_arch.py": 1679246024.2714128, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/hand_2d_keypoints/hand_2d_keypoints.py": 1679246024.1153808, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_stabilization/utils/RAFTUtils.py": 1679246024.2984185, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_stabilization/utils/ProjectionUtils.py": 1679246024.2984185, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_stabilization/utils/math_utils.py": 1679246024.2994187, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_stabilization/utils/IterativeSmooth.py": 1679246024.2974184, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_stabilization/utils/image_utils.py": 1679246024.2984185, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_stabilization/utils/MedianFilter.py": 1679246024.2974184, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_stabilization/utils/WarpUtils.py": 1679246024.2984185, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_stabilization/DUT/MotionPro.py": 1679246024.2954178, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_stabilization/DUT/RAFT/update.py": 1679246024.2974184, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_stabilization/DUT/RAFT/corr.py": 1679246024.2964182, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_stabilization/DUT/RAFT/extractor.py": 1679246024.2974184, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_stabilization/DUT/RAFT/raft.py": 1679246024.2974184, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_stabilization/DUT/config.py": 1679246024.2964182, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_stabilization/DUT/rf_det_module.py": 1679246024.2964182, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_stabilization/DUT/DUT_raft.py": 1679246024.2954178, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_stabilization/DUT/rf_det_so.py": 1679246024.2964182, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_stabilization/DUT/Smoother.py": 1679246024.2954178, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_stabilization/DUTRAFTStabilizer.py": 1679246024.2944176, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/easycv_base.py": 1679246024.0723717, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_single_object_tracking/models/ostrack/base_backbone.py": 1679246024.2924173, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_single_object_tracking/models/ostrack/utils.py": 1679246024.2924173, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_single_object_tracking/models/ostrack/vit_ce.py": 1679246024.2924173, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_single_object_tracking/models/ostrack/ostrack.py": 1679246024.2924173, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_single_object_tracking/models/layers/patch_embed.py": 1679246024.2914171, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_single_object_tracking/models/layers/attn.py": 1679246024.2914171, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_single_object_tracking/models/layers/attn_blocks.py": 1679246024.2914171, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_single_object_tracking/models/layers/head.py": 1679246024.2914171, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_single_object_tracking/models/procontext/utils.py": 1679246024.2934175, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_single_object_tracking/models/procontext/vit_ce.py": 1679246024.2934175, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_single_object_tracking/models/procontext/procontext.py": 1679246024.2934175, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_single_object_tracking/tracker/procontext.py": 1679246024.2944176, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_single_object_tracking/tracker/ostrack.py": 1679246024.2934175, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_single_object_tracking/utils/utils.py": 1679246024.2944176, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_single_object_tracking/config/ostrack.py": 1679246024.290417, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/face_2d_keypoints/face_2d_keypoints_align.py": 1679246024.0893753, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_streaming_perception/longshortnet/models/dfp_pafpn_long.py": 1679246024.3004189, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_streaming_perception/longshortnet/models/longshort.py": 1679246024.3004189, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_streaming_perception/longshortnet/models/longshort_backbone_neck.py": 1679246024.301419, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_streaming_perception/longshortnet/models/dfp_pafpn_short.py": 1679246024.3004189, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_streaming_perception/longshortnet/longshortnet.py": 1679246024.2994187, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/video_streaming_perception/longshortnet/exp/longshortnet_base.py": 1679246024.3004189, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/ocr_detection/model.py": 1679246024.2244031, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/ocr_detection/utils.py": 1679246024.2254033, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/ocr_detection/modules/seg_detector_loss.py": 1679246024.2254033, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/ocr_detection/modules/dbnet.py": 1679246024.2254033, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/ocr_detection/preprocessor.py": 1679246024.2244031, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/human_wholebody_keypoint/human_wholebody_keypoint.py": 1679246024.1193814, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_binary_quant_classification/bnext.py": 1679246024.1193814, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/image_binary_quant_classification/binary_quant_model.py": 1679246024.1193814, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/shop_segmentation/shop_seg_base.py": 1679246024.2384062, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/shop_segmentation/head_fpn.py": 1679246024.237406, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/shop_segmentation/utils.py": 1679246024.2394063, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/shop_segmentation/shop_seg_model.py": 1679246024.2384062, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/shop_segmentation/models.py": 1679246024.2384062, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/shop_segmentation/common.py": 1679246024.237406, "/usr/local/lib/python3.9/dist-packages/modelscope/models/cv/shop_segmentation/neck_fpn.py": 1679246024.2384062, "/usr/local/lib/python3.9/dist-packages/modelscope/models/base/base_model.py": 1679246024.0713716, "/usr/local/lib/python3.9/dist-packages/modelscope/models/base/base_torch_head.py": 1679246024.0713716, "/usr/local/lib/python3.9/dist-packages/modelscope/models/base/base_head.py": 1679246024.0713716, "/usr/local/lib/python3.9/dist-packages/modelscope/models/base/base_torch_model.py": 1679246024.0713716, "/usr/local/lib/python3.9/dist-packages/modelscope/models/builder.py": 1679246024.0603693, "/usr/local/lib/python3.9/dist-packages/modelscope/metrics/accuracy_metric.py": 1679246024.0533679, "/usr/local/lib/python3.9/dist-packages/modelscope/metrics/image_quality_assessment_degradation_metric.py": 1679246024.0553682, "/usr/local/lib/python3.9/dist-packages/modelscope/metrics/ned_metric.py": 1679246024.0563686, "/usr/local/lib/python3.9/dist-packages/modelscope/metrics/text_ranking_metric.py": 1679246024.0573688, "/usr/local/lib/python3.9/dist-packages/modelscope/metrics/video_summarization_metric.py": 1679246024.058369, "/usr/local/lib/python3.9/dist-packages/modelscope/metrics/ocr_recognition_metric.py": 1679246024.0563686, "/usr/local/lib/python3.9/dist-packages/modelscope/metrics/image_instance_segmentation_metric.py": 1679246024.054368, "/usr/local/lib/python3.9/dist-packages/modelscope/metrics/token_classification_metric.py": 1679246024.0573688, "/usr/local/lib/python3.9/dist-packages/modelscope/metrics/map_metric.py": 1679246024.0553682, "/usr/local/lib/python3.9/dist-packages/modelscope/metrics/video_stabilization_metric.py": 1679246024.058369, "/usr/local/lib/python3.9/dist-packages/modelscope/metrics/bleu_metric.py": 1679246024.0533679, "/usr/local/lib/python3.9/dist-packages/modelscope/metrics/loss_metric.py": 1679246024.0553682, "/usr/local/lib/python3.9/dist-packages/modelscope/metrics/prediction_saving_wrapper.py": 1679246024.0563686, "/usr/local/lib/python3.9/dist-packages/modelscope/metrics/ciderD/ciderD_scorer.py": 1679246024.058369, "/usr/local/lib/python3.9/dist-packages/modelscope/metrics/ciderD/ciderD.py": 1679246024.058369, "/usr/local/lib/python3.9/dist-packages/modelscope/metrics/image_inpainting_metric.py": 1679246024.054368, "/usr/local/lib/python3.9/dist-packages/modelscope/metrics/image_color_enhance_metric.py": 1679246024.054368, "/usr/local/lib/python3.9/dist-packages/modelscope/metrics/sequence_classification_metric.py": 1679246024.0573688, "/usr/local/lib/python3.9/dist-packages/modelscope/metrics/video_super_resolution_metric/matlab_functions.py": 1679246024.059369, "/usr/local/lib/python3.9/dist-packages/modelscope/metrics/video_super_resolution_metric/niqe.py": 1679246024.059369, "/usr/local/lib/python3.9/dist-packages/modelscope/metrics/video_super_resolution_metric/metric_util.py": 1679246024.059369, "/usr/local/lib/python3.9/dist-packages/modelscope/metrics/video_super_resolution_metric/video_super_resolution_metric.py": 1679246024.059369, "/usr/local/lib/python3.9/dist-packages/modelscope/metrics/referring_video_object_segmentation_metric.py": 1679246024.0563686, "/usr/local/lib/python3.9/dist-packages/modelscope/metrics/image_denoise_metric.py": 1679246024.054368, "/usr/local/lib/python3.9/dist-packages/modelscope/metrics/image_quality_assessment_mos_metric.py": 1679246024.0553682, "/usr/local/lib/python3.9/dist-packages/modelscope/metrics/movie_scene_segmentation_metric.py": 1679246024.0563686, "/usr/local/lib/python3.9/dist-packages/modelscope/metrics/ppl_metric.py": 1679246024.0563686, "/usr/local/lib/python3.9/dist-packages/modelscope/metrics/video_frame_interpolation_metric.py": 1679246024.0573688, "/usr/local/lib/python3.9/dist-packages/modelscope/metrics/audio_noise_metric.py": 1679246024.0533679, "/usr/local/lib/python3.9/dist-packages/modelscope/metrics/base.py": 1679246024.0533679, "/usr/local/lib/python3.9/dist-packages/modelscope/metrics/inbatch_recall_metric.py": 1679246024.0553682, "/usr/local/lib/python3.9/dist-packages/modelscope/metrics/image_portrait_enhancement_metric.py": 1679246024.0553682, "/usr/local/lib/python3.9/dist-packages/modelscope/metrics/text_generation_metric.py": 1679246024.0573688, "/usr/local/lib/python3.9/dist-packages/modelscope/metrics/action_detection_evaluator.py": 1679246024.0533679, "/usr/local/lib/python3.9/dist-packages/modelscope/metrics/builder.py": 1679246024.0533679, "/usr/local/lib/python3.9/dist-packages/modelscope/metrics/image_colorization_metric.py": 1679246024.054368, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/audio/asr_wenet_inference_pipeline.py": 1679246024.4234443, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/audio/ans_pipeline.py": 1679246024.422444, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/audio/voice_activity_detection_pipeline.py": 1679246024.4254448, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/audio/ans_dfsmn_pipeline.py": 1679246024.422444, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/audio/kws_farfield_pipeline.py": 1679246024.4234443, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/audio/punctuation_processing_pipeline.py": 1679246024.4244444, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/audio/speaker_diarization_pipeline.py": 1679246024.4244444, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/audio/inverse_text_processing_pipeline.py": 1679246024.4234443, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/audio/linear_aec_pipeline.py": 1679246024.4244444, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/audio/lm_infer_pipeline.py": 1679246024.4244444, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/audio/speaker_verification_pipeline.py": 1679246024.4254448, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/audio/separation_pipeline.py": 1679246024.4244444, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/audio/kws_kwsbp_pipeline.py": 1679246024.4234443, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/audio/asr_inference_pipeline.py": 1679246024.4234443, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/audio/text_to_speech_pipeline.py": 1679246024.4254448, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/audio/speaker_verification_light_pipeline.py": 1679246024.4254448, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/audio/timestamp_pipeline.py": 1679246024.4254448, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/science/protein_structure_pipeline.py": 1679246024.4684536, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/multi_modal/diffusers_wrapped/diffusers_pipeline.py": 1679246024.4584515, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/multi_modal/diffusers_wrapped/stable_diffusion/stable_diffusion_pipeline.py": 1679246024.4584515, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/multi_modal/diffusers_wrapped/stable_diffusion/chinese_stable_diffusion_pipeline.py": 1679246024.4584515, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/multi_modal/asr_pipeline.py": 1679246024.4544508, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/multi_modal/multi_modal_embedding_pipeline.py": 1679246024.455451, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/multi_modal/image_text_retrieval_pipeline.py": 1679246024.455451, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/multi_modal/video_captioning_pipeline.py": 1679246024.4564512, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/multi_modal/visual_entailment_pipeline.py": 1679246024.4574513, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/multi_modal/video_multi_modal_embedding_pipeline.py": 1679246024.4574513, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/multi_modal/gridvlp_pipeline.py": 1679246024.4544508, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/multi_modal/text2sql_pipeline.py": 1679246024.4564512, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/multi_modal/image_captioning_pipeline.py": 1679246024.455451, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/multi_modal/team_multi_modal_similarity_pipeline.py": 1679246024.4564512, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/multi_modal/video_question_answering_pipeline.py": 1679246024.4574513, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/multi_modal/soonet_video_temporal_grounding_pipeline.py": 1679246024.455451, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/multi_modal/text_to_image_synthesis_pipeline.py": 1679246024.4564512, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/multi_modal/generative_multi_modal_embedding_pipeline.py": 1679246024.4544508, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/multi_modal/sudoku_pipeline.py": 1679246024.4564512, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/multi_modal/document_vl_embedding_pipeline.py": 1679246024.4544508, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/multi_modal/visual_question_answering_pipeline.py": 1679246024.4574513, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/multi_modal/text_to_video_synthesis_pipeline.py": 1679246024.4564512, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/multi_modal/mgeo_ranking_pipeline.py": 1679246024.455451, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/multi_modal/ocr_recognition_pipeline.py": 1679246024.455451, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/multi_modal/disco_guided_diffusion_pipeline/utils.py": 1679246024.4594517, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/multi_modal/disco_guided_diffusion_pipeline/disco_guided_diffusion.py": 1679246024.4594517, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/multi_modal/visual_grounding_pipeline.py": 1679246024.4574513, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/util.py": 1679246024.422444, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/base.py": 1679246024.421444, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/fid_dialogue_pipeline.py": 1679246024.4634526, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/interactive_translation_pipeline.py": 1679246024.4644527, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/document_segmentation_pipeline.py": 1679246024.4624524, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/codegeex_code_translation_pipeline.py": 1679246024.460452, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/dialog_modeling_pipeline.py": 1679246024.4614522, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/translation_quality_estimation_pipeline.py": 1679246024.4674535, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/extractive_summarization_pipeline.py": 1679246024.4624524, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/zero_shot_classification_pipeline.py": 1679246024.4674535, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/mglm_text_summarization_pipeline.py": 1679246024.4644527, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/word_alignment_pipeline.py": 1679246024.4674535, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/token_classification_pipeline.py": 1679246024.4664533, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/fill_mask_pipeline.py": 1679246024.4634526, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/codegeex_code_generation_pipeline.py": 1679246024.460452, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/text_error_correction_pipeline.py": 1679246024.4664533, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/text_ranking_pipeline.py": 1679246024.4664533, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/language_identification_pipline.py": 1679246024.4644527, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/sentence_embedding_pipeline.py": 1679246024.465453, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/document_grounded_dialog_rerank_pipeline.py": 1679246024.4624524, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/siamese_uie_pipeline.py": 1679246024.465453, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/fasttext_text_classification_pipeline.py": 1679246024.4634526, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/named_entity_recognition_pipeline.py": 1679246024.4644527, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/dialog_intent_prediction_pipeline.py": 1679246024.460452, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/translation_pipeline.py": 1679246024.4674535, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/distributed_gpt3_pipeline.py": 1679246024.4614522, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/faq_question_answering_pipeline.py": 1679246024.4634526, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/feature_extraction_pipeline.py": 1679246024.4634526, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/distributed_plug_pipeline.py": 1679246024.4614522, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/distributed_gpt_moe_pipeline.py": 1679246024.4614522, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/table_question_answering_pipeline.py": 1679246024.465453, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/text_classification_pipeline.py": 1679246024.465453, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/text_generation_pipeline.py": 1679246024.4664533, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/summarization_pipeline.py": 1679246024.465453, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/information_extraction_pipeline.py": 1679246024.4644527, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/dialog_state_tracking_pipeline.py": 1679246024.4614522, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/user_satisfaction_estimation_pipeline.py": 1679246024.4674535, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/conversational_text_to_sql_pipeline.py": 1679246024.460452, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/word_segmentation_pipeline.py": 1679246024.4674535, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/automatic_post_editing_pipeline.py": 1679246024.460452, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/document_grounded_dialog_retrieval_pipeline.py": 1679246024.4624524, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/translation_evaluation_pipeline.py": 1679246024.4664533, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/nlp/document_grounded_dialog_generate_pipeline.py": 1679246024.4614522, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_human_parsing_pipeline.py": 1679246024.4354467, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_restoration_pipeline.py": 1679246024.4374473, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/bad_image_detecting_pipeline.py": 1679246024.4274452, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/hand_2d_keypoints_pipeline.py": 1679246024.431446, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/table_recognition_pipeline.py": 1679246024.4434485, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/ulfd_face_detection_pipeline.py": 1679246024.4444487, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/video_frame_interpolation_pipeline.py": 1679246024.4454489, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/tinynas_detection_pipeline.py": 1679246024.4444487, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/movie_scene_segmentation_pipeline.py": 1679246024.441448, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/arc_face_recognition_pipeline.py": 1679246024.4274452, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_color_enhance_pipeline.py": 1679246024.4334464, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_quality_assessment_mos_pipeline.py": 1679246024.4374473, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_face_fusion_pipeline.py": 1679246024.4344466, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/vidt_pipeline.py": 1679246024.4484496, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/vision_efficient_tuning_pipeline.py": 1679246024.4484496, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/panorama_depth_estimation_pipeline.py": 1679246024.4424484, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_quality_assessment_man_pipeline.py": 1679246024.4374473, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/body_2d_keypoints_pipeline.py": 1679246024.4274452, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/face_liveness_ir_pipeline.py": 1679246024.4294455, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_debanding_pipeline.py": 1679246024.4334464, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_body_reshaping_pipeline.py": 1679246024.4324462, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/face_image_generation_pipeline.py": 1679246024.4294455, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_quality_assessment_degradation_pipeline.py": 1679246024.4364471, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/animal_recognition_pipeline.py": 1679246024.426445, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/video_panoptic_segmentation_pipeline.py": 1679246024.4474492, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_detection_pipeline.py": 1679246024.4344466, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/video_category_pipeline.py": 1679246024.4444487, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/face_recognition_onnx_ir_pipeline.py": 1679246024.430446, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/mobile_image_super_resolution_pipeline.py": 1679246024.4404478, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/face_recognition_ood_pipeline.py": 1679246024.430446, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_salient_detection_pipeline.py": 1679246024.4374473, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/face_quality_assessment_pipeline.py": 1679246024.430446, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_matching_pipeline.py": 1679246024.4354467, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/face_recognition_onnx_fm_pipeline.py": 1679246024.430446, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_cartoon_pipeline.py": 1679246024.4324462, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_classification_pipeline.py": 1679246024.4334464, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/face_liveness_xc_pipeline.py": 1679246024.4294455, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_to_image_generate_pipeline.py": 1679246024.4384475, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/video_deinterlace_pipeline.py": 1679246024.4454489, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/vision_middleware_pipeline.py": 1679246024.4484496, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/referring_video_object_segmentation_pipeline.py": 1679246024.4434485, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/facial_landmark_confidence_pipeline.py": 1679246024.431446, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_colorization_pipeline.py": 1679246024.4334464, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/ocr_detection_pipeline.py": 1679246024.441448, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/product_retrieval_embedding_pipeline.py": 1679246024.4424484, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/vop_retrieval_se_pipeline.py": 1679246024.4494498, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/video_instance_segmentation_pipeline.py": 1679246024.446449, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/live_category_pipeline.py": 1679246024.4404478, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/text_driven_segmentation_pipleline.py": 1679246024.4444487, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/vop_retrieval_pipeline.py": 1679246024.4484496, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/card_detection_pipeline.py": 1679246024.4274452, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/mask_face_recognition_pipeline.py": 1679246024.4404478, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/video_colorization_pipeline.py": 1679246024.4454489, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/crowd_counting_pipeline.py": 1679246024.4284453, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/tbs_detection_pipeline.py": 1679246024.4444487, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/face_recognition_pipeline.py": 1679246024.431446, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/retina_face_detection_pipeline.py": 1679246024.4434485, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/video_multi_object_tracking_pipeline.py": 1679246024.446449, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/tinynas_classification_pipeline.py": 1679246024.4444487, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/virtual_try_on_pipeline.py": 1679246024.4484496, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_open_vocabulary_detection_pipeline.py": 1679246024.4364471, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/video_stabilization_pipeline.py": 1679246024.4474492, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/content_check_pipeline.py": 1679246024.4284453, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_semantic_segmentation_pipeline.py": 1679246024.4374473, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/realtime_video_object_detection_pipeline.py": 1679246024.4424484, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/video_summarization_pipeline.py": 1679246024.4474492, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/ocr_utils/model_resnet_mutex_v4_linewithchar.py": 1679246024.45145, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/ocr_utils/resnet_utils.py": 1679246024.4524503, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/ocr_utils/utils.py": 1679246024.4524503, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/ocr_utils/ops.py": 1679246024.45145, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/ocr_utils/model_dla34.py": 1679246024.45045, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/ocr_utils/table_process.py": 1679246024.4524503, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/ocr_utils/ocr_modules/convnext.py": 1679246024.4534504, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/ocr_utils/ocr_modules/vitstr.py": 1679246024.4534504, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/ocr_utils/ocr_modules/timm_tinyc.py": 1679246024.4534504, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/ocr_utils/model_convnext_transformer.py": 1679246024.45045, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/ocr_utils/model_resnet18_half.py": 1679246024.45145, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/ocr_utils/resnet18_v1.py": 1679246024.4524503, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/ocr_utils/model_vlpt.py": 1679246024.45145, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_bts_depth_estimation_pipeline.py": 1679246024.4324462, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/ddcolor_image_colorization_pipeline.py": 1679246024.4284453, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/video_depth_estimation_pipeline.py": 1679246024.4454489, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/face_attribute_recognition_pipeline.py": 1679246024.4294455, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/action_recognition_pipeline.py": 1679246024.426445, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_reid_person_pipeline.py": 1679246024.4374473, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/license_plate_detection_pipeline.py": 1679246024.4394476, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_instance_segmentation_pipeline.py": 1679246024.4354467, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_skychange_pipeline.py": 1679246024.4384475, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/face_reconstruction_pipeline.py": 1679246024.431446, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_matting_pipeline.py": 1679246024.4354467, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/object_detection_3d_pipeline.py": 1679246024.441448, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/controllable_image_generation_pipeline.py": 1679246024.4284453, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/body_3d_keypoints_pipeline.py": 1679246024.4274452, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/pointcloud_sceneflow_estimation_pipeline.py": 1679246024.4424484, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_panoptic_segmentation_pipeline.py": 1679246024.4364471, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/face_processing_base_pipeline.py": 1679246024.430446, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/face_detection_pipeline.py": 1679246024.4294455, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_inpainting_sdv2_pipeline.py": 1679246024.4354467, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_depth_estimation_pipeline.py": 1679246024.4344466, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_driving_perception_pipeline.py": 1679246024.4344466, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/maskdino_instance_segmentation_pipeline.py": 1679246024.4404478, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/easycv_pipelines/face_2d_keypoints_pipeline.py": 1679246024.4494498, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/easycv_pipelines/segmentation_pipeline.py": 1679246024.45045, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/easycv_pipelines/detection_pipeline.py": 1679246024.4494498, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/easycv_pipelines/base.py": 1679246024.4494498, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/easycv_pipelines/human_wholebody_keypoint_pipeline.py": 1679246024.45045, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_super_resolution_pipeline.py": 1679246024.4384475, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/cmdssl_video_embedding_pipeline.py": 1679246024.4274452, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_portrait_enhancement_pipeline.py": 1679246024.4364471, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/mtcnn_face_detection_pipeline.py": 1679246024.441448, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/ddpm_semantic_segmentation_pipeline.py": 1679246024.4284453, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/mog_face_detection_pipeline.py": 1679246024.4404478, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/hand_static_pipeline.py": 1679246024.4324462, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/video_single_object_tracking_pipeline.py": 1679246024.4474492, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/hicossl_video_embedding_pipeline.py": 1679246024.4324462, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/product_segmentation_pipeline.py": 1679246024.4424484, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_defrcn_fewshot_pipeline.py": 1679246024.4334464, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_to_image_translation_pipeline.py": 1679246024.4394476, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/ocr_recognition_pipeline.py": 1679246024.441448, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_structured_model_probing_pipeline.py": 1679246024.4384475, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/video_object_segmentation_pipeline.py": 1679246024.446449, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/video_super_resolution_pipeline.py": 1679246024.4474492, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/shop_segmentation_pipleline.py": 1679246024.4434485, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/human_reconstruction_pipeline.py": 1679246024.4324462, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_deblur_pipeline.py": 1679246024.4334464, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/general_recognition_pipeline.py": 1679246024.431446, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_paintbyexample_pipeline.py": 1679246024.4364471, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/nerf_recon_acc_pipeline.py": 1679246024.441448, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/video_human_matting_pipeline.py": 1679246024.446449, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/face_human_hand_detection_pipeline.py": 1679246024.4294455, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/video_inpainting_pipeline.py": 1679246024.446449, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_inpainting_pipeline.py": 1679246024.4354467, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/language_guided_video_summarization_pipeline.py": 1679246024.4394476, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/facial_expression_recognition_pipeline.py": 1679246024.431446, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/lineless_table_recognition_pipeline.py": 1679246024.4394476, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/action_detection_pipeline.py": 1679246024.426445, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_mvs_depth_estimation_pipeline.py": 1679246024.4364471, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/motion_generation_pipeline.py": 1679246024.4404478, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/skin_retouching_pipeline.py": 1679246024.4434485, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_style_transfer_pipeline.py": 1679246024.4384475, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/image_denoise_pipeline.py": 1679246024.4344466, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/tbs_detection_utils/utils.py": 1679246024.4534504, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/face_emotion_pipeline.py": 1679246024.4294455, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/cv/indoor_layout_estimation_pipeline.py": 1679246024.4394476, "/usr/local/lib/python3.9/dist-packages/modelscope/pipelines/builder.py": 1679246024.421444, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/ofa/sudoku.py": 1679246024.4864573, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/ofa/text_classification.py": 1679246024.4864573, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/ofa/visual_entailment.py": 1679246024.4874575, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/ofa/summarization.py": 1679246024.4864573, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/ofa/visual_question_answering.py": 1679246024.4874575, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/ofa/visual_grounding.py": 1679246024.4874575, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/ofa/text2sql.py": 1679246024.4864573, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/ofa/utils/get_tables.py": 1679246024.4884577, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/ofa/utils/random_help.py": 1679246024.4884577, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/ofa/utils/constant.py": 1679246024.4884577, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/ofa/utils/text2phone.py": 1679246024.4884577, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/ofa/utils/vision_helper.py": 1679246024.489458, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/ofa/utils/transforms.py": 1679246024.489458, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/ofa/utils/bridge_content_encoder.py": 1679246024.4884577, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/ofa/utils/collate.py": 1679246024.4884577, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/ofa/utils/audio_helper.py": 1679246024.4874575, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/ofa/ocr_recognition.py": 1679246024.4854572, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/ofa/image_captioning.py": 1679246024.4854572, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/ofa/base.py": 1679246024.4854572, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/ofa/asr.py": 1679246024.4854572, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/ofa/text_to_image_synthesis.py": 1679246024.4864573, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/ofa/image_classification.py": 1679246024.4854572, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/kws.py": 1679246024.4694538, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/tts.py": 1679246024.470454, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/science/uni_fold.py": 1679246024.489458, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/audio.py": 1679246024.4684536, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/video.py": 1679246024.470454, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/movie_scene_segmentation/transforms.py": 1679246024.4734547, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/image.py": 1679246024.4694538, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/base.py": 1679246024.4694538, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/asr.py": 1679246024.4684536, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/relation_extraction_preprocessor.py": 1679246024.475455, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/document_segmentation_preprocessor.py": 1679246024.4744549, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/document_grounded_dialog_rerank_preprocessor.py": 1679246024.4744549, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/document_grounded_dialog_retrieval_preprocessor.py": 1679246024.4744549, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/text_error_correction.py": 1679246024.4764552, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/bert_seq_cls_tokenizer.py": 1679246024.4734547, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/space/preprocess.py": 1679246024.480456, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/space/dialog_modeling_preprocessor.py": 1679246024.479456, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/space/dialog_intent_prediction_preprocessor.py": 1679246024.479456, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/space/tensorlistdataset.py": 1679246024.480456, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/space/data_loader.py": 1679246024.479456, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/space/dst_processors.py": 1679246024.480456, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/space/lazy_dataset.py": 1679246024.480456, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/space/dialog_state_tracking_preprocessor.py": 1679246024.479456, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/space/fields/gen_field.py": 1679246024.4814563, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/space/fields/intent_field.py": 1679246024.4824564, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/space/batch.py": 1679246024.4784558, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/space/tokenizer.py": 1679246024.480456, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/space/args.py": 1679246024.4784558, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/space/sampler.py": 1679246024.480456, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/zero_shot_classification_preprocessor.py": 1679246024.4784558, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/utils.py": 1679246024.4774556, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/space_T_en/conversational_text_to_sql_preprocessor.py": 1679246024.4834569, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/space_T_en/fields/parse.py": 1679246024.484457, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/space_T_en/fields/process_dataset.py": 1679246024.484457, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/space_T_en/fields/preprocess_dataset.py": 1679246024.484457, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/space_T_en/fields/common_utils.py": 1679246024.484457, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/dialog_classification_use_preprocessor.py": 1679246024.4734547, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/faq_question_answering_preprocessor.py": 1679246024.4744549, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/token_classification_viet_preprocessor.py": 1679246024.4774556, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/text_generation_preprocessor.py": 1679246024.4764552, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/token_classification_thai_preprocessor.py": 1679246024.4774556, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/sentence_embedding_preprocessor.py": 1679246024.4764552, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/transformers_tokenizer.py": 1679246024.4774556, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/word_alignment_preprocessor.py": 1679246024.4784558, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/mglm_summarization_preprocessor.py": 1679246024.475455, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/siamese_uie_preprocessor.py": 1679246024.4764552, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/mgeo_ranking_preprocessor.py": 1679246024.475455, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/translation_evaluation_preprocessor.py": 1679246024.4774556, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/fill_mask_preprocessor.py": 1679246024.475455, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/document_grounded_dialog_generate_preprocessor.py": 1679246024.4744549, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/text_ranking_preprocessor.py": 1679246024.4764552, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/feature_extraction_preprocessor.py": 1679246024.475455, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/space_T_cn/table_question_answering_preprocessor.py": 1679246024.4824564, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/space_T_cn/fields/database.py": 1679246024.4824564, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/space_T_cn/fields/schema_link.py": 1679246024.4834569, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/space_T_cn/fields/struct.py": 1679246024.4834569, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/token_classification_preprocessor.py": 1679246024.4774556, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/nlp/text_classification_preprocessor.py": 1679246024.4764552, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/common.py": 1679246024.4694538, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/cv/video_stabilization.py": 1679246024.4724545, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/cv/image_restoration_preprocessor.py": 1679246024.4724545, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/cv/action_detection_mapper.py": 1679246024.470454, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/cv/image_quality_assessment_mos.py": 1679246024.4724545, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/cv/cv2_transforms.py": 1679246024.4714544, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/cv/mmcls_preprocessor.py": 1679246024.4724545, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/cv/timer.py": 1679246024.4724545, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/cv/controllable_image_generation.py": 1679246024.4714544, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/cv/image_classification_preprocessor.py": 1679246024.4714544, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/cv/util.py": 1679246024.4724545, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/cv/video_super_resolution.py": 1679246024.4734547, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/cv/image_quality_assessment_man.py": 1679246024.4714544, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/cv/bad_image_detecting_preprocessor.py": 1679246024.4714544, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/multi_modal.py": 1679246024.470454, "/usr/local/lib/python3.9/dist-packages/modelscope/preprocessors/builder.py": 1679246024.4694538, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/hooks/ddp_hook.py": 1679246024.5004601, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/hooks/evaluation_hook.py": 1679246024.5014606, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/hooks/checkpoint_hook.py": 1679246024.5004601, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/hooks/optimizer/apex_optimizer_hook.py": 1679246024.503461, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/hooks/optimizer/torch_optimizer_hook.py": 1679246024.504461, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/hooks/optimizer/base.py": 1679246024.504461, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/hooks/hook.py": 1679246024.5014606, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/hooks/megatron_hook.py": 1679246024.5024607, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/hooks/logger/text_logger_hook.py": 1679246024.503461, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/hooks/logger/tensorboard_hook.py": 1679246024.503461, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/hooks/logger/base.py": 1679246024.503461, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/hooks/lr_scheduler_hook.py": 1679246024.5014606, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/hooks/clip_clamp_logit_scale_hook.py": 1679246024.5004601, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/hooks/priority.py": 1679246024.5024607, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/hooks/early_stop_hook.py": 1679246024.5014606, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/hooks/iter_timer_hook.py": 1679246024.5014606, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/hooks/deepspeed_hook.py": 1679246024.5014606, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/hooks/compression/sparsity_hook.py": 1679246024.5024607, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/hooks/compression/utils.py": 1679246024.5024607, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/hooks/builder.py": 1679246024.5004601, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/audio/kws_utils/batch_utils.py": 1679246024.494459, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/audio/kws_utils/det_utils.py": 1679246024.494459, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/audio/kws_utils/model_utils.py": 1679246024.494459, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/audio/kws_utils/runtime_utils.py": 1679246024.4954593, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/audio/kws_utils/file_utils.py": 1679246024.494459, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/audio/kws_farfield_trainer.py": 1679246024.4934587, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/audio/separation_trainer.py": 1679246024.4934587, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/audio/asr_trainer.py": 1679246024.4924586, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/audio/kws_nearfield_trainer.py": 1679246024.4934587, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/audio/tts_trainer.py": 1679246024.4934587, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/audio/ans_trainer.py": 1679246024.4924586, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/parallel/utils.py": 1679246024.5144632, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/parallel/builder.py": 1679246024.513463, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/optimizer/child_tuning_adamw_optimizer.py": 1679246024.513463, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/optimizer/builder.py": 1679246024.513463, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/nlp_trainer.py": 1679246024.4914584, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/multi_modal/ofa/ofa_trainer_utils.py": 1679246024.5074618, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/multi_modal/ofa/ofa_trainer.py": 1679246024.5074618, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/multi_modal/mgeo_ranking_trainer.py": 1679246024.5054612, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/multi_modal/mplug/mplug_trainer.py": 1679246024.5064616, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/multi_modal/team/team_trainer.py": 1679246024.5074618, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/multi_modal/team/team_trainer_utils.py": 1679246024.5074618, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/multi_modal/clip/clip_trainer_utils.py": 1679246024.5064616, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/multi_modal/clip/clip_trainer.py": 1679246024.5064616, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/easycv/utils/register_util.py": 1679246024.49946, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/easycv/utils/metric.py": 1679246024.49946, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/easycv/utils/hooks.py": 1679246024.49946, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/easycv/trainer.py": 1679246024.49946, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/default_config.py": 1679246024.4914584, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/utils/log_buffer.py": 1679246024.5144632, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/utils/inference.py": 1679246024.5144632, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/base.py": 1679246024.4904583, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/trainer.py": 1679246024.4914584, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/nlp/sentence_embedding_trainer.py": 1679246024.509462, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/nlp/document_grounded_dialog_generate_trainer.py": 1679246024.508462, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/nlp/table_question_answering_trainer.py": 1679246024.5104623, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/nlp/text_ranking_trainer.py": 1679246024.5104623, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/nlp/sequence_classification_trainer.py": 1679246024.5104623, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/nlp/csanmt_translation_trainer.py": 1679246024.508462, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/nlp/space/metrics/metrics_tracker.py": 1679246024.5124629, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/nlp/space/dialog_modeling_trainer.py": 1679246024.5114625, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/nlp/space/trainer/gen_trainer.py": 1679246024.5124629, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/nlp/space/trainer/intent_trainer.py": 1679246024.5124629, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/nlp/space/dialog_intent_trainer.py": 1679246024.5114625, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/nlp/space/eval.py": 1679246024.5114625, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/nlp/siamese_uie_trainer.py": 1679246024.5104623, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/nlp/plug_trainer.py": 1679246024.509462, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/nlp/document_grounded_dialog_rerank_trainer.py": 1679246024.508462, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/nlp/gpt3_trainer.py": 1679246024.509462, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/nlp/text_generation_trainer.py": 1679246024.5104623, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/nlp/faq_question_answering_trainer.py": 1679246024.509462, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/nlp/document_grounded_dialog_retrieval_trainer.py": 1679246024.508462, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/nlp/gpt_moe_trainer.py": 1679246024.509462, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/cv/movie_scene_segmentation_trainer.py": 1679246024.4974597, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/cv/action_detection_trainer.py": 1679246024.4954593, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/cv/nerf_recon_acc_trainer.py": 1679246024.4984598, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/cv/card_detection_scrfd_trainer.py": 1679246024.4954593, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/cv/image_portrait_enhancement_trainer.py": 1679246024.4974597, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/cv/image_classifition_trainer.py": 1679246024.4964595, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/cv/referring_video_object_segmentation_trainer.py": 1679246024.4984598, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/cv/vision_efficient_tuning_trainer.py": 1679246024.4984598, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/cv/image_detection_damoyolo_trainer.py": 1679246024.4964595, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/cv/image_instance_segmentation_trainer.py": 1679246024.4974597, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/cv/face_detection_scrfd_trainer.py": 1679246024.4964595, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/cv/ocr_detection_db_trainer.py": 1679246024.4984598, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/cv/image_defrcn_fewshot_detection_trainer.py": 1679246024.4964595, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/cv/ocr_recognition_trainer.py": 1679246024.4984598, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/cv/cartoon_translation_trainer.py": 1679246024.4954593, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/cv/image_inpainting_trainer.py": 1679246024.4974597, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/training_args.py": 1679246024.4924586, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/builder.py": 1679246024.4914584, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/lrscheduler/warmup/warmup.py": 1679246024.5054612, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/lrscheduler/warmup/base.py": 1679246024.5054612, "/usr/local/lib/python3.9/dist-packages/modelscope/trainers/lrscheduler/builder.py": 1679246024.504461, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/audio/asr_dataset.py": 1679246024.3944383, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/data_loader/data_loader_manager.py": 1679246024.3954387, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/data_loader/data_loader.py": 1679246024.3954387, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/meta/data_meta_manager.py": 1679246024.4144425, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/meta/data_meta_config.py": 1679246024.4144425, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/download/download_config.py": 1679246024.4134424, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/download/download_manager.py": 1679246024.4144425, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/download/dataset_builder.py": 1679246024.4134424, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/data_files/data_files_manager.py": 1679246024.3954387, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/task_datasets/video_summarization_dataset.py": 1679246024.4154427, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/task_datasets/torch_base_dataset.py": 1679246024.4154427, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/task_datasets/sidd_image_denoising.py": 1679246024.4154427, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/task_datasets/reds_image_deblurring_dataset.py": 1679246024.4154427, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/task_datasets/gopro_image_deblurring_dataset.py": 1679246024.4144425, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/utils/oss_utils.py": 1679246024.4164429, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/utils/upload_utils.py": 1679246024.4164429, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/utils/dataset_utils.py": 1679246024.4154427, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/utils/delete_utils.py": 1679246024.4164429, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/ms_dataset.py": 1679246024.393438, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/face_2d_keypoins/face_2d_keypoints_dataset.py": 1679246024.4034402, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/audio/kws_farfield_dataset.py": 1679246024.3994393, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/audio/kws_nearfield_processor.py": 1679246024.3994393, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/audio/asr_dataset.py": 1679246024.3984392, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/audio/kws_nearfield_dataset.py": 1679246024.3994393, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/torch_custom_dataset.py": 1679246024.3984392, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/image_instance_segmentation_coco_dataset.py": 1679246024.397439, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/bad_image_detecting/bad_image_detecting_dataset.py": 1679246024.3994393, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/object_detection/detection_dataset.py": 1679246024.4074411, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/language_guided_video_summarization_dataset.py": 1679246024.397439, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/mgeo_ranking_dataset.py": 1679246024.397439, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/referring_video_object_segmentation/transformers.py": 1679246024.4114418, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/referring_video_object_segmentation/referring_video_object_segmentation_dataset.py": 1679246024.4104416, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/video_super_resolution/video_super_resolution_dataset.py": 1679246024.4134424, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/image_colorization/image_colorization_dataset.py": 1679246024.4044404, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/movie_scene_segmentation/movie_scene_segmentation_dataset.py": 1679246024.4074411, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/movie_scene_segmentation/sampler.py": 1679246024.4074411, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/image_portrait_enhancement/image_portrait_enhancement_dataset.py": 1679246024.4054406, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/image_portrait_enhancement/data_utils.py": 1679246024.4054406, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/image_semantic_segmentation/segmentation_dataset.py": 1679246024.4064407, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/video_summarization_dataset.py": 1679246024.3984392, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/image_inpainting/aug.py": 1679246024.4054406, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/image_inpainting/image_inpainting_dataset.py": 1679246024.4054406, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/text_ranking_dataset.py": 1679246024.3984392, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/veco_dataset.py": 1679246024.3984392, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/image_classification/classification_dataset.py": 1679246024.4044404, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/image_quality_assmessment_mos/image_quality_assessment_mos_dataset.py": 1679246024.4064407, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/reds_image_deblurring_dataset.py": 1679246024.397439, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/ocr_recognition_dataset.py": 1679246024.397439, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/gopro_image_deblurring_dataset.py": 1679246024.3964388, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/image_quality_assessment_degradation/image_quality_assessment_degradation_dataset.py": 1679246024.4064407, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/video_frame_interpolation/video_frame_interpolation_dataset.py": 1679246024.412442, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/video_frame_interpolation/data_utils.py": 1679246024.412442, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/hand_2d_keypoints/hand_2d_keypoints_dataset.py": 1679246024.4034402, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/sidd_image_denoising/data_utils.py": 1679246024.4114418, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/sidd_image_denoising/transforms.py": 1679246024.412442, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/sidd_image_denoising/sidd_image_denoising_dataset.py": 1679246024.4114418, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/video_stabilization/video_stabilization_dataset.py": 1679246024.412442, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/easycv_base.py": 1679246024.3964388, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/damoyolo/collate_batch.py": 1679246024.4004395, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/damoyolo/transforms/build.py": 1679246024.40244, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/damoyolo/transforms/transforms.py": 1679246024.40244, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/damoyolo/samplers/iteration_based_batch_sampler.py": 1679246024.40244, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/damoyolo/samplers/grouped_batch_sampler.py": 1679246024.40244, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/damoyolo/samplers/distributed.py": 1679246024.40244, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/damoyolo/evaluation/coco/coco_eval.py": 1679246024.40144, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/damoyolo/build.py": 1679246024.4004395, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/damoyolo/datasets/mosaic_wrapper.py": 1679246024.40144, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/damoyolo/datasets/coco.py": 1679246024.4004395, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/builder.py": 1679246024.3964388, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/ocr_detection/data_loader.py": 1679246024.4084413, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/ocr_detection/processes/random_crop_data.py": 1679246024.4104416, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/ocr_detection/processes/make_border_map.py": 1679246024.4094415, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/ocr_detection/processes/augment_data.py": 1679246024.4094415, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/ocr_detection/processes/make_icdar_data.py": 1679246024.4094415, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/ocr_detection/processes/make_seg_detection_data.py": 1679246024.4104416, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/ocr_detection/processes/data_process.py": 1679246024.4094415, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/ocr_detection/processes/normalize_image.py": 1679246024.4104416, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/ocr_detection/augmenter.py": 1679246024.4084413, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/ocr_detection/image_dataset.py": 1679246024.4084413, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/ocr_detection/measures/quad_measurer.py": 1679246024.4094415, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/ocr_detection/measures/iou_evaluator.py": 1679246024.4084413, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/custom_datasets/human_wholebody_keypoint/human_wholebody_keypoint_dataset.py": 1679246024.4034402, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/dataset_cls/dataset.py": 1679246024.3964388, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/auth/auth_config.py": 1679246024.3944383, "/usr/local/lib/python3.9/dist-packages/modelscope/msdatasets/context/dataset_context_config.py": 1679246024.3944383, "/usr/local/lib/python3.9/dist-packages/modelscope/exporters/torch_model_exporter.py": 1679246024.0473666, "/usr/local/lib/python3.9/dist-packages/modelscope/exporters/base.py": 1679246024.0463665, "/usr/local/lib/python3.9/dist-packages/modelscope/exporters/nlp/csanmt_for_translation_exporter.py": 1679246024.0483668, "/usr/local/lib/python3.9/dist-packages/modelscope/exporters/nlp/sbert_for_sequence_classification_exporter.py": 1679246024.0483668, "/usr/local/lib/python3.9/dist-packages/modelscope/exporters/nlp/model_for_token_classification_exporter.py": 1679246024.0483668, "/usr/local/lib/python3.9/dist-packages/modelscope/exporters/nlp/sbert_for_zero_shot_classification_exporter.py": 1679246024.0483668, "/usr/local/lib/python3.9/dist-packages/modelscope/exporters/cv/cartoon_translation_exporter.py": 1679246024.0473666, "/usr/local/lib/python3.9/dist-packages/modelscope/exporters/cv/face_detection_scrfd_exporter.py": 1679246024.0473666, "/usr/local/lib/python3.9/dist-packages/modelscope/exporters/cv/object_detection_damoyolo_exporter.py": 1679246024.0473666, "/usr/local/lib/python3.9/dist-packages/modelscope/exporters/builder.py": 1679246024.0463665, "/usr/local/lib/python3.9/dist-packages/modelscope/exporters/tf_model_exporter.py": 1679246024.0463665}, "modelscope_path": "/usr/local/lib/python3.9/dist-packages/modelscope"}